AI-Ranked Reddit Feed

5000 posts

r/ChatGPT jdw1977

Has anyone successfully monetized a custom GPT?

I recently built a custom GPT that got more traction than I expected, including a few hundred people actually trying it out. That made me wonder if something like this could work as a small product.

To do that, I tried a few white-label platforms (Pickaxe, CalStudio, LaunchLemonade) that let you package and sell a custom GPT.

On the surface, they all work. You can gate access, set pricing, and technically launch something.

But once I tried to treat it like a real product, the tradeoffs showed up pretty quickly:

  • Better UX came with worse economics
  • Better pricing models introduced more friction
  • More flexibility came with less reliability

I couldn’t find a setup where everything lined up. It felt less like choosing the “best” platform and more like choosing which tradeoff you’re willing to live with.

Curious what others have run into with white-label tools.

If so, what ended up working for you? Did you stick with a white-label solution, or take a different approach?

r/ChatGPT Top-Requirement-2102

ADHD Day Coach GPT

I created a GPT to help people with ADHD start the day. It's called ADHD Day Coach. It's built around the idea of collecting emotional state first, then working in cooperation with that to create a healthy plan for the day that doesn't override me or lead me to burnout. This is not a "life hack" or a way to improve productivity. This is about being psychologically healthy with what to choose to work on and why.

I encourage people here to try it out and tell me how it goes.

Link: https://chatgpt.com/g/g-69bd93f956908191849eec8a354c2160-adhd-day-coach

r/ClaudeCode Proud_Slide2403

Using your Anthropic quota for repetitive file restructuring is a massive waste of resources.

I absolutely love the reasoning capabilities of Claude for deep algorithmic debugging, but if you are feeding it thousands of lines of code just to update class names across a repository, you are doing it wrong.

The rate limit will lock you out exactly when you actually need it. I built a hybrid script that offloads all the low level bulk code generation to the Minimax M2.7 endpoint. It handles standard syntax execution flawlessly at a fraction of the cost, preserving my premium tier access for the architectural heavy lifting. I am curious how everyone else is managing their API limits during massive codebase overhauls. Do you just stop working when you hit the cap?

r/ClaudeCode Mastiff37

Can claude code take images from MCP (tool calls)?

I'm having trouble getting this to work. I'm packaging my image as an "Image" object but it blows context. If I work around by passing the file name and having claude read it from the file system it works just fine. I'm wondering if images from tool calls is not properly supported.

r/ClaudeAI double_mmaybach

Claude Skills for Stock Analysis and Trading?

Do you guys use any skills or agents for Claude to improve its market knowledge and ability such as analyzing stocks and crypto coins whether if they are trending or not. If there are these types of skills, agents, or roles where can I look for them?

r/LocalLLaMA Some-Ice-4455

Simplifying local LLM setup (llama.cpp + fallback handling)

I kept running into issues with local setups: CUDA instability dependency conflicts GPU fallback not behaving consistently So I started wrapping my setup to make it more predictable. Current setup: Model: Qwen (GGUF) Runtime: llama.cpp GPU/CPU fallback enabled Still working through: response consistency handling edge-case failures Curious how others here are managing stable local setups.

r/ChatGPT buttflapper444

Free AI is dead. What does that mean for the vibe coding community?

When chat gpt was first becoming popular I remember I tried it and could use it for like, days, and no limit was ever hit. I tried again recently just to see how low the limit was.... 2 messages with a c# program file, hit the limit in 15 mins. Now, you can't do anything with an attachment for free anymore. Claude is even worse, go look at their sub r/ClaudeAI . People hitting the limit in 1-5 messages for the rest of an entire day.

I'm really curious, what's that mean for the tens of millions who used it religiously for everything?

r/singularity Complete_Bee4911

using ai to detect ai might be one of the most cyberpunk feeeback loops we’ve normalised

r/LocalLLaMA vishnoo

What would you use for local coding assist on a "weak" machine (6GB VRAM 32 GB RAM) - light FE coding, no architecture. is QWEN3 good enough?

so as it says, I am not a FE eng, but want to do some light FE work
I don't need the smartest model but need to get some work done.
I ran out of tokens (20$ a month) for the week on day 2, so thinking of running something local
I tried serving QWEN3 with ollama and connecting codex to it, but it was clunky at best.

I figured I'd ask the experts

so local windows machine, I ran it on WSL, but codex then had issues accessing the local directories. is it better to run it in PowerShell (shudder)

gemma4:26 (quantized) also sort of fits but provided worse results.

to sum up
1. WSL vs windows native
2. codex? (claude-code blocked local models) opencode?
3. qwen? gemma?

r/ClaudeCode yucek

Built a Word MCP server so Claude Code can actually edit Word Documents Live

A few weeks ago I posted about word-mcp-live, a Word MCP server that lets Claude edit documents while they're open in Word — with real tracked changes, comments, and per-action undo. The main complaint was fair: it only worked on Windows.

It now works on macOS too.

Same tool names, same parameters, same behavior. The server detects your platform and uses the right backend — COM on Windows, JavaScript for Automation (JXA) on Mac. You don't configure anything differently.

https://reddit.com/link/1sg19yx/video/g812bxv8f0ug1/player

Setup is the same on both platforms:

{ "mcpServers": { "word": { "command": "uvx", "args": ["--from", "word-mcp-live", "word_mcp_server"], "env": { "MCP_AUTHOR": "Your Name" } } } } 

That's it. uvx handles Python, dependencies, everything. Works in Claude Code, Claude Desktop, Cursor, VS Code — any MCP client.

What works on Mac (40 of 44 live tools):

  • Read/write/find/replace text (full Turkish/Unicode support)
  • Bold, italic, font changes, paragraph formatting
  • Track changes — real Word revisions with your name
  • Comments — add, delete, list (threaded replies not available yet)
  • Tables — read cells, write cells, add rows
  • Page layout, headers/footers, bookmarks, section breaks
  • Equations, cross-references
  • Paragraph diagnostics (keep_with_next chains, style issues)
  • Per-operation undo

What doesn't work on Mac (4 tools):

  • Threaded comment replies (not in Word for Mac's scripting dictionary)
  • Comment resolve/unresolve (same reason)
  • Undo history inspection
  • Watermarks (needs VBA, which Apple killed via sandboxing)

These return a clear error message instead of crashing.

How it works under the hood:

On Windows, pywin32 talks to Word through COM — direct access to the running Word.Application object.

On Mac, there's no COM. Instead, the server builds JavaScript for Automation (JXA) scripts and executes them via osascript. Word for Mac exposes an AppleScript dictionary with ~9,600 lines of classes and commands. Most things map cleanly — doc.Paragraphs(i).Range.Text in COM becomes d.paragraphs[i].textObject.content() in JXA.

Some gotchas we found:

  • Find/Replace only works through the Selection object on Mac (range-based find is broken)
  • Font access is fontObject not font in JXA
  • Table cells need getCellFromTable(table, {row, column}) instead of table.Cell(row, col)
  • Word for Mac's do script VBA bridge is completely dead (Apple sandboxing killed it in Word 365)
  • Adding comments only works via AppleScript, not JXA (different syntax for make new Word comment)

124 tools total now: 80 cross-platform (python-docx), 44 live editing (Windows COM + macOS JXA).

GitHub: github.com/ykarapazar/word-mcp-live PyPI: pip install word-mcp-live

r/LocalLLM Outrageous_Mark9761

Vox — Local AI that actually controls your Mac (Mail, Messages, files)

Hi everyone, built Vox.

Problem:
Most AI tools on Mac stop at answering. You still have to switch apps and actually do the work yourself. If not then its going to some cloud server run by open ai or anthropic.

Comparison:
Tools like ChatGPT, Claude, or Raycast mostly give responses or shortcuts. Vox is built to directly act through macOS apps (Mail, Messages, Finder, screen control) instead of just suggesting what to do. Plus it gives convenience, you don't have to be tech savvy to use it, install it and already connected to everything. Indexes your files too, and all locally.

Pricing:
Free and open source
https://www.vox-ai.chat
https://github.com/vox-ai-app/vox

Runs fully locally on your machine (model + voice + memory). No accounts, no telemetry, works offline.

Right now it can:

  • read and draft replies in Mail.app
  • send messages through Messages
  • search, move, and organize files
  • read the screen and click / scroll
  • create docs, PDFs, presentations
  • run multi-step tasks like research + summaries
  • schedule recurring tasks

Still early and actively being built.

If you're into local AI, macOS automation, or want to contribute, would be great to have more people working on this.

r/ClaudeAI tomaszka

Claude Code setup for aligning new Vue.js UI to match the old Smartclient UI

I’m currently upgrading a web application frontend (moving from SmartClient to Vue.js), and I’ve been doing it module by module. So far I’ve migrated my first module (a list view), but there’s a problem: the new UI looks quite different from the old one.

My goal is to make the new version match the old layout as closely as possible (not pixel-perfect, but same structure, styling, colors, etc.).

I’d like to use Claude Code cli (or other) to help automate or assist this process.

What I’m trying to achieve

A well-defined workflow where Claude Code/Other tool can:

  • Analyze the existing UI (DOM structure, screenshots, etc.)
  • Analyze the new Vue.js implementation
  • Suggest and generate changes to make the new UI match the old one

The main question

What’s the best way to approach this?

Any advice, tools, or workflows would be greatly appreciated.

r/AI_Agents MindlessAd8634

Shopping on AI is broken. I'm thinking about fixing it with a brand concierge layer - here's the concept, tell me where I'm wrong

Try shopping on ChatGPT or Claude right now. Ask it to help you find a skincare routine, reorder your coffee, or track a package.

It'll hallucinate the product, forget what you bought last time, and have zero idea what happened after you clicked checkout. Every session starts from zero.

That's the gap I'm staring at.

The concept I'm exploring: a personal shopping agent - call it a brand concierge -- that lives at the intersection of the customer and all the brands they love. Not a chatbot for one brand. Not a generic AI assistant. Something in between: a persistent, cross-brand layer that knows your order history, understands your preferences, tracks your deliveries, and surfaces the right recommendation at the right moment.

Think of it like this: you have a Aesop order in transit, an Origami coffee just delivered, and Tekla processing. Instead of logging into three apps, checking three tracking pages, and getting three separate "how was your experience" emails -- one agent knows all of it, manages all of it, and proactively tells you what you need to know.

The post-purchase experience in e-commerce is completely broken and nobody has fixed it because every brand is optimizing for their own touchpoints, not the customer's actual life.

A few things I'm genuinely unsure about and want to think through with this community:

1. Distribution problem: How do you get customers to adopt a cross-brand agent when every brand wants to own the relationship themselves? Is this a consumer app, a B2B product sold to brands, or something else entirely?

2. Trust and data: Customers would need to connect their accounts across multiple brands. What's the realistic adoption hurdle here -- is this a "never going to happen" problem or a "find the right hook" problem?

3. The memory layer: The value compounds the more you use it. But how do you get someone to stick around long enough for the memory to become valuable? What's the "aha moment" that makes someone realize this is different from just Googling?

4. Who owns this: Is this a platform play, a feature inside an existing super-app, or does it need to be brand-native to work? I've seen a few attempts at cross-brand loyalty aggregators that went nowhere. What did they miss?

5. The agentic piece: At what point does the agent go from surfacing information to actually taking action -- auto-reordering, negotiating returns, proactively flagging price drops? Where does helpful end and creepy begin?

I've got a working prototype concept (screenshot in comments) but I'm in early thinking mode. What am I missing? Where does this fall apart?

r/artificial AssignmentHopeful651

I just read about Mythos AI and I genuinely sat there staring at my screen for 5 minutes. Something crossed a line and nobody's talking about it.

I'm not a doomer. Never have been.

I rolled my eyes at every "AI will kill us all" headline. Called it fear-mongering. Told my friends to relax.

Then I saw the Mythos news.

And something shifted in my chest that I can't really explain.

Here's what gets me, it's not that the technology is powerful. We knew it was going to get powerful. That was always the deal.

It's that nobody actually asked us if we wanted this.

No vote. No debate. No "hey, before we cross this line, should we maybe talk about it?" Just a press release, a demo, some VCs losing their minds in the comments, and suddenly the world is just... different now.

That's the part that broke something in me.

I keep thinking about how we handle other things that can change civilization, nuclear power, gene editing, even social media. There are committees. Regulations. International agreements. Years of ethical debate before anything goes live.

With AI? We basically said "ship it and figure it out later."

Mythos isn't even the scariest part. The scariest part is that Mythos was announced casually. Like it was a product update. Like the bar for what counts as an alarm bell has moved so far that we don't even flinch anymore.

We've been desensitized to our own extinction-level headlines.

I don't know what the answer is. I'm not smart enough to solve this.

But I do know that when something this big happens and the loudest voices in the room are the ones who financially benefit from it, that's usually when things go very wrong for everyone else.

Just feel like more people should be talking about this instead of arguing about which AI makes better images.

r/singularity ocean_protocol

So, this week claude wiped agentic AI startups with a new update. Also, as they have mythos now, they will ship things very fast without any trouble

Honestly, they are a full pack now. A few hours ago, they released Claude managed agents which lets you build long-running, autonomous agentic systems plus with their new suite of apis, engineering teams can harness Claude's exponential power with scalable infra out of the box. Absolute chill moment

I mean these agents got memory, they got compute, and anyone can ship without much hassle. Inference compute market will skyrocket as well. crazy times

r/Anthropic Expert_Annual_19

Investments in anthropic

if anyone wants to invest in anthropic,

dm me ASAP .

(Only for serious people , maybe 2-3 only )

r/AI_Agents Miserable_Emergency6

I open-sourced a smart router for AI/model routing — would love feedback

I kept running into the same problem:

Every time I wanted to use different AI models/providers, I ended up writing ugly routing logic into the app itself.

Fallbacks, model selection, cost control, provider switching, retries, etc. all started leaking into places they didn’t belong.

So I built and open-sourced smart-router.

It’s basically a smart router layer for AI/model requests.

Main idea:
It's a transparent AI inference proxy that optimizes context and routes prompts to different specialized models based on content type etc. This means I can have a single agent that can multitask and not have to delegate tasks between agents. Requests go to a single API and are optimized to keep costs low and tasks are routed to models best suited to handle that type of task. For example, coding requests go to a coder model and creative requests to gpt5.4 for example. **updated for clarity**

In a future version, I'm planning to leverage a fast local AI model to have it aggressively manage context optimization and compression as well as providing LCR decisions for tasks. For example, this request is lower priority and could be services by GPT5.2 instead of the more expensive gpt-5.4 etc etc **added after I realized I left this out**

The kind of stuff I wanted it to handle cleanly:
- route to different providers/models
- add fallback behavior
- support experimentation without rewriting app code
- eventually make production routing less painful

Still early, but I’d really like honest feedback from people who’ve actually had to manage this kind of thing.

Main questions:
- what’s missing for this to be actually useful?
- what would you want before trusting it in production?
- is this solving a real problem, or am I overengineering my own pain?

Happy to get roasted if needed.

r/LocalLLaMA marivesel

Choice for agentic LLM or help optimize Qwen3.5-35B-A3B for 24GB VRAM

RTX3090 24GB VRAM, WSL install of Ollama latest and Hermes Agent latest.
First I have tried Gemma4:31B - so slow!
Then Gemma4:26B MoE - fast, but so many mistakes for few days repeatable.

Then I've found Qwen3.5-35B-A3B Q4_K_M here in Reddit and OH BOY, IT'S GORGEOUS! It's fluently making what I want. But... rather slowish! Then I found that the file itself is 23GB, and I have given context of 32K, overfilling my VRAM with more than 1.5GB (and my RAM is DDR4 ECC, slow).

Question is - can I somehow optimize to fill the whole model in my VRAM with 16K/32K context, or should I try lower quality model, which would you suggest?

I like the speed and quality of MoE models, I am not writing a super complex stuff, just some automations and helping around in my business with regular tasks.

r/AI_Agents Legitimate_Sherbet_7

I got tired of rebuilding the same AI backend for document agents, so I built a reusable API layer

I kept running into the same problem building AI agents with documents.

Every project started the same way:

  • upload PDFs
  • chunk + embed
  • wire up retrieval
  • connect to an LLM

…and then spend more time stitching everything together than actually building the application.

What surprised me is that retrieval wasn’t really the hard part.

The real friction was everything after that:

  • getting useful, reasoned answers (not just chunks)
  • controlling behavior with prompts/personas
  • and especially having zero visibility into cost per request

So I ended up building this into a reusable API layer that handles the full flow:

  • upload a document
  • send a chat-style query
  • get a reasoned answer (not raw retrieval)
  • see the real-time cost of each request

Basically adding a reasoning layer + cost layer on top of document retrieval, so it’s actually usable in a real app.

The goal wasn’t to replace RAG tools — just to stop rebuilding the same backend every time.

What I have now is:

  • simple API (few lines to upload + query)
  • system_prompt support for behavior/personas
  • real-time cost tracking (per request / per user)
  • multi-user ready
  • a couple working examples (catalog assistant, transcript → PDF → query flow)

Happy to share the repo/examples if anyone is interested.

Curious how others are handling:

  • cost visibility
  • reasoning vs raw retrieval
  • and making these systems usable in production

Would be interested to hear what’s working (or not) in your setups.

r/KlingAI_Videos xKaizx

POV: You finally got a Pikachu and it chose your shoulder | Nano Banana | Kling

r/ClaudeAI ClaudeAI-mod-bot

Claude Status Update : Elevated errors on Claude.ai, API, Claude Code on 2026-04-08T17:50:17.000Z

This is an automatic post triggered within 2 minutes of an official Claude system status update.

Incident: Elevated errors on Claude.ai, API, Claude Code

Check on progress and whether or not the incident has been resolved yet here : https://status.claude.com/incidents/5f418rpyb84x

Also check the Performance Megathread to see what others are reporting : https://www.reddit.com/r/ClaudeAI/comments/1s7f72l/claude_performance_and_bugs_megathread_ongoing/

r/StableDiffusion thickmisa

Are there any AI tools that let you generate images using your own photo as a reference question mark I'm looking for something that's pretty customizable and easy to use. But I'm not sure what's actually reliable right now.

r/MCPservers musaceylan

I open-sourced OrchestrAI — an MCP server for multi-model coding orchestration

Hi all — I’m the maker of OrchestrAI.

I built it because most AI coding workflows still seem to rely on one model doing everything:

• planning • coding • testing • reviewing • judging 

That felt limiting.

So I made OrchestrAI, an open-source MCP server that lets multiple models collaborate on software engineering tasks instead of working alone.

It can work with:

• Claude • OpenAI / Codex • Gemini • local models 

And route them into roles like:

• Planner • Coder • Tester • Reviewer • Judge 

Current focus:

• parallel orchestration • collaborative agent workflows • lint / test / type-check verification • privacy-aware routing • local-only mode for sensitive work • traces and artifacts for inspectable decisions 

Current modes:

• parallel\_draft • impl\_tester • planner\_coder\_reviewer 

GitHub: https://github.com/musaceylan/OrchestrAI

I’m sharing mainly to get real feedback from people working with MCP, coding agents, or local model setups.

The part I’d most like feedback on:

What should a real orchestration layer do that current single-model coding tools still don’t?

r/aivideo Wise_Ad_2499

Best ai generator

r/StableDiffusion Leather_Function_843

How do I make images look less AI-ity

r/artificial Xaqx

Project Glasswing is inherently Cartel Behaviour

If the large companies always get access to the latest models first to "sure up cybersecurity" they will always have a head start on the competition and new contenders in the tech space.

If Glasswing is locked down to only be allowed for cybersecurity thats a different story but I doubt it is.

r/Anthropic OneClimate8489

Anthropic’s new Claude Managed Agents public beta drop

Anthropic’s new Claude Managed Agents public beta drop and it feels like they finally fixed the biggest pain point for builders shipping real agents.

So here's a breakdown and guide for everyone :

1/ WHAT THEY ACTUALLY ARE

first, this is not another prompt wrapper or basic tool call API — most people will miss this

> claude managed agents: full production cloud infrastructure for agents. secure sandbox, state management, credentials, long-running sessions, error recovery, tracing all handled by them.

> old way: you had to build all that infra yourself (sandboxing, checkpointing, scoped perms, orchestration) and it took months.

one is a real production agent platform. the other is DIY hell. completely different mental model.

2/ HOW THEY HANDLE AUTONOMY

this is where the real difference shows up

> managed agents run long autonomous sessions for hours (even if you disconnect). persistent progress, outputs, self-evaluation loops until success criteria hit.

> old way: your agent dies on error, loses state, or you babysit every step with dumb prompt loops.

> complex multi-hour tasks: managed agents win by a mile

> simple one-shot prompts: you still use normal claude

3/ CONTEXT AND GOVERNANCE

> managed agents give you scoped permissions, identity management, full execution tracing + multi-agent coordination (research preview where one agent spins up others to parallelize).

> old way: you pray your custom setup doesn’t leak creds or break prod.

> large real-world systems with actual tools and data: managed agents

> quick prototypes where nothing matters: old way still fine

stop wasting time on infra and just ship agent products 10x faster. If you’re still manually wiring agent backends in 2026… bro just stop.

r/SideProject gothamismycity

I built a small desk display that shows the status of my OpenClaw agent as a cute pet

PixClaw is a small display that shows your OpenClaw agent’s status in real time — idle, thinking, stuck, errors, etc. — as a little animated pixel pet.

It connects to your OpenClaw instance locally over Wi-Fi and just sits on your desk so you can check your agent at a glance.

Still early and iterating on the hardware, would love any feedback!

More information and waitlist here:

https://pixclaw.io

r/ProgrammerHumor programmerjunky

theDrunkenMaster

r/Rag ShotOil1398

I work support at an AI company and the same mistake keeps showing up over and over

Not a pitch for anything, genuinely just something I've noticed after answering tickets for a while now.

Small businesses come in excited about AI, set something up, and then a few weeks later they're frustrated because it's giving wrong answers or making things up. Almost every time it's the same thing - they expected the AI to already know their business.

It doesn't. You have to feed it your own stuff. Your FAQs, your policies, how you actually handle edge cases. Without that it's just guessing.

The ones who stick with it are usually the ones who spent a few hours just writing down how they do things, uploading that, and then testing it properly before going live. Boring work but it's the difference.

Anyway, just something I've noticed. Curious if anyone else has run into this or has a different experience.

r/comfyui o0ANARKY0o

Qwen 2511 edit but with Flux Klein model,clip & vae. Disconnect the latent and set to 1440x2160 or 2160x1440 unless you need it lower so it doesn't get oversized in D&C. resize each input to hi res accordingly. remove background from character or objects you want to place. change D&C applycontrolnet

Take a Qwen 2511 edit workflow and switch the models with Flux Kleins model, clip & vae(and switch back and forth between them when needed or for different results). Disconnect the latent and set to 1440x2160 or 2160x1440 unless you need it lower, so it doesn't get oversized in the Divide and Conquer workflow. resize each input to hi-res separately and accordingly(I know mine are connected but I disconnect and reconnect things none stop). remove backgrounds from character or objects you want to place. change D&C denoise and applycontrolnet end to add detail or keep character consistency. changing to values lower like I have mean you can get very detailed clear beautiful images at insanely hi resolutions(last picture example)

https://drive.google.com/file/d/1uX2URGaiPmEUA16y84sT9njKGyQHYN6L/view?usp=drive_link

r/StableDiffusion New_Fee_887

Stable Diffusion on RDNA4

Hello! I have been tinkering trying to get stable diffusion working on my main machine with a 9070XT and I am getting nowhere unfortunately, I tried my luck with A1111's stable diffusion webui, but its pretty outdated, I also tried comfyui as its more maintained and got limited success as it runs but crashes after each image, so for now I am using my laptop as a server which is not ideal.

I would love to get some feedback on how or if someone got SD working under RDNA4, thanks in advance!

If it matters, my pc specs are:

9070XT AMD GPU

ryzen 7 9800X3D

64GB RAM DDR5

(edit) I am pretty new to SD, so I am sorry if I got something fundamentally wrong.

r/SideProject Instance_Not_Found

I build the Knowledge Base Agent inspired by Andrej Karpathy's post

Andrej Karpathy shared his experience with building a personal knowledge base: https://x.com/karpathy/status/2039805659525644595

I built my own version but with a little tweak:

  • Obsidian as the front end.
  • Agent running in a workspace 24/7 (linting and building connections between notes. The workspace is powered by funky.dev)
  • File sync between the remote agent workspace and local file system. (I used syncthing)
  • Model agnostic. (Any model from OpenRouter. I used Gemini 3 flash, and it worked pretty well. No more locked in with Claude Code).

If you want to try it out, here is the link: https://funky.dev/demo/knowledge-base-agent

The project is also open sourced: https://github.com/jinxi97/pi-agent-workspace

r/SideProject SomeWillingness7701

Adhd makes me stop studying halfway....... so I made lepadhle.com

r/singularity contemporare

Piracy Implications of Mythos and Project Glasswing.

Let's assume Mythos means that AI cybersecurity has completely surpassed human capabilities or will very soon.

Anthropic believes AI-assisted defensive security will win over offensive attacks. Project Glasswing suggests a framework where established companies can pay for this service.

If closed models provide better defensive security than humans are at attacking, how can piracy prevail as companies generate systems that are impenetrable by humans?

What will prevent all software and the internet from eventually becoming a network of black boxes that humans can't do anything to, making even benevolent piracy impossible?

r/Futurology Marcellus508

What is a fundamental human problem that technology hasn't "fixed" yet, but will in 20 years?

I’m looking to start a venture that builds for the future rather than chasing today's trends. Most "big" companies (Apple, Amazon, Google) revolutionized a basic human need (communication, shopping, information).

In your opinion, what is the next "unsolved" friction point in our daily lives—whether it's how we manage our health, our cities, or our personal data—that is ripe for a 20-year disruption?

r/homeassistant kenstir

Wanted: a quiet HA setup to control 2 heat pump exchangers and a way out of the rabbit hole

I want to purchase a box to run HA, then use it to

  1. remotely control 2 Mitsubishi heat pump exchangers (maybe Mahtanar Heat Pump Controller)
  2. monitor temperature / humidity / leak sensors

What I have:

  • Aqara Hub M2
  • Several Aqara temperature and humidity sensors

What I would like most of all:

  • Quiet box
  • Minimum fuss
  • Reliable long-term remote operation

Here are the options I have considered, but I am getting quite lost in all the details; do you have a recommendation?

  1. Home Assistant Green. Do I need a Connect ZBT-2 or SLZB-06M or connect my Aqara Hub M2 somehow?
  2. NUC with Intel N150 CPU running Ubuntu. I did read that "apps" don't run on HA under Docker but does that matter?
  3. NUC running HAOS. But if it's going to be running HAOS it might as well be the HA Green I suppose.

Thank you so much for your opinion / experience!

r/Futurology Kind-Ad6740

Could a system detect human presence using heartbeat signals from a distance?

I came across reports about a system combining UWB radar and advanced signal processing to detect micro-signals like heartbeat or respiration.

While the claims sound extreme, parts of the underlying tech already exist (for example, radar-based vital sign detection through walls).

Curious what people think is this plausible with current tech, or still far from reality?

r/aivideo Outrageous-Yak-177

[oc] Lonely caped protagonist (animated) - AI Instrumental from 38 seconds

r/aivideo Puzzleheaded-Mall528

Brace for Impact

r/arduino Enough_Boysenberry20

Need help

i have an 5v solar panel attached on an tp056 lithium battery charger with 2 3.7v batteries and a buck converter, 5v relay, and pump also im using esp 32 with 4 capacitive soil moisture sensor. I put the code on the esp 32, the logic is when wet turn on the pump when dry it's gonna do nothing but when i uploaded it on the esp32 the water pump is just on even when the sensor is wet or dry. Idk what to do, anyone know how to fix? (also this is my first time using this)

r/comfyui henryk_kwiatek

Adding multiline description UNDER image

Hey,

https://preview.redd.it/bf3nmx8j50ug1.png?width=1536&format=png&auto=webp&s=254887690bdae0c5ba2b5edddde6bce698a75b8c

I’m trying to do something that feels like it should be really simple but I can’t get it working cleanly in ComfyUI. I want to take an image and a piece of text and end up with the image plus a caption under it, on a white background, clearly separated from the image (not overlayed on top). Every node I’ve tried only gets me part of the way — WAS Node Suite (Text Image), KJNodes (Create Text Image + concatenate), different TextOverlay nodes, even Impact Pack — and I always run into the same issues. Either the text is stuck in one long line with no wrapping so longer captions go off the canvas, or everything is designed as an overlay instead of actually building a layout under the image. The whole thing ends up feeling really hacky with manual concatenation and guessing sizes. I’m basically looking for something that can handle text wrapping properly, render it inside a box (like a white caption area), and then place it under the image without fighting the layout. At this point I honestly don’t know if there’s a proper node for this or if everyone is just piecing it together manually every time. If anyone has a clean way to do this I’d really appreciate it, thanks. (pic rel generated with chat gpt as an inlustration of whatm im looking for)

r/arduino Ken471

I built a Mecanum wheeled Robot car with Bluetooth control (Manual + Autonomous modes)

**Hardware:**

- Arduino Uno

- Adafruit Motor Shield V2

- HC-SR04 (front + rear)

- SG90 micro servo for scanning

- HC-05/HC-06 Bluetooth module

**Controls via Bluetooth Electronics:**

- WASD + QE/ZC for strafe/diagonals

- J/L to rotate

- Speed control (%+ / %-)

- E-stop (!), mode switch (0/1)

**Safety:** watchdog, slow/stop zones, speed ramping

**Autonomous:** drives forward, scans 5 directions when blocked, picks clearest path

**Code:** https://github.com/ken47-1/Project-MecanumCar

**NOTE:** This project was developed with human oversight and AI-assisted code generation.

r/automation Inevitable_Island984

Is anyone actually making money selling automations in 2026?

Seeing a lot of hype about "AI Automation Agencies" lately. Beyond the simple Zapier/Make stuff that most businesses can do themselves now, is there still a real market for this?

If you’re actually doing this:

  • Are you finding it profitable or is it a race to the bottom?
  • What kind of problems are clients actually paying to solve right now?
  • Is it one-off builds or are you getting recurring retainers?

Just trying to see if this is a viable business or just another "guru" trend. Thanks!

r/homeassistant lexvo1

Values for home battery won't show up in energy dashboard, are visible under settings however

I added my IndeVolt home battery successfully in Home Assistant. When I look under settings-developer tools I see the values of the entities of the IndeVolt. However, whatever I try, I can't get those values visible in the Energy dashboard (they are always shown as zero). I tried resetting Home Assistant, restarting my browser, clearing my browser cache, multiple times checking if I entered the right entities for the energy dashboard, etc..

I'm a bit disappointed in Home Assistant: what I am trying to do feels as a basic thing which shouldn't be difficult. Am I overlooking something?

r/arduino geo_tp

Cardputer Game Station - 12 Emulators working on Arduino Core - GENESIS SNES, NES, GAMEBOY, GAMEGEAR, MASTER SYSTEM, NEOGEO POCKET, WONDERSWAN, ATARI, PC ENGINE...

Emulators running NES, SNES, GameBoy, GameGear, Master System, Genesis / Megadrive, Neo Geo Pocket, WonderSwan, Lynx, Atari7800, Atari2600 and PC-Engine on ESP32 with 320KB of RAM

All cores were optimized to run within a memory footprint of under 256 KB of RAM, with particular emphasis on minimizing both RAM and flash usage. All 12 cores are packed into just 2.5 MB of flash using the Arduino framework.

https://github.com/geo-tp/Cardputer-Game-Station-Emulators

r/KlingAI_Videos MetaEmber

We used Kling 3.0 and NanoBanana to make over 2,500 consistent characters. How does the quality hold up? (PROMPT AND WORKFLOW BELOW)

Building a swipe based AI dating sim called Amoura.io and Kling 3.0 combined with NanoBanana has been a core part of our image to video pipeline. We've used it to generate profile videos/photos and in-conversation selfies across 2,500+ hand-crafted characters, each one going through roughly a dozen iterations before it's good enough to ship 4 to 10.

The video below shows a swipe through a sample of the character pool — mix of animated Kling 3.0 video loop profiles and static images (to show the contrast) and then digs into two specific characters across their second, third, fourth, fifth and sixth photos so you can see what consistency actually looks like in practice across different scenes, outfits and contexts.

My photo prompt structure (how to get best output to send to Kling):

Opening identity lock: "Ultra-realistic mirror selfie of SAME EXACT CHARACTER as reference, [2-3 hyper-specific physical micro-details that aren't covered by beauty language]"

Scene setting (comes AFTER the identity lock): "[Location, lighting, what they're doing — keep brief]"

Shot style: "iPhone-style candid, vertical format, sharp subject, naturally blurred background. Authentic, spontaneous vibe."

Texture line (always last): "Realistic skin texture, natural proportions, no AI skin smoothing, no beauty filter effect. Ultra-realistic, high detail."

For identity anchoring, micro-distinctive physical details get locked in before any scene or outfit information always. The texture lock (Realistic skin texture, natural proportions, no AI skin smoothing, no beauty filter effect. Ultra-realistic, high detail.) always comes last. Change that order and drift gets noticeably worse.

For motion clips, less motion and sometimes less description equals more identity stability than we expected. The word "involuntary" in motion prompts significantly improved naturalness. We think the model interprets it as behavior rooted in internal state rather than performance for a lens. Keep it simple OR as highly detailed as humanly possible.. We prefer simple.

PROMPT FOR KLING 3.0
She gently adjusts her hair and starts adjusting her shorts then grins shyly

PROMPT FOR FIRST IMAGE (NANOBANANAPRO)
Ultra-realistic waist-up portrait selfie of mixed Southeast Asian and Pacific Islander (27), warm medium-tan complexion with golden-brown undertones, smooth skin with subtle natural texture, high cheekbones, softly angular jaw, full lips, almond-shaped dark brown eyes with a calm and slightly downward gaze, straight dark brown-to-black hair falling just past the shoulders with a natural center-to-side part, slim athletic build with a defined waist, natural proportions, no makeup or minimal no-makeup makeup, understated and effortlessly cool presence. Standing in a mirror at the edge of a narrow loft bed setup with white linen sheets, surf wax on the windowsill, and a thrifted quilt folded under the ladder, wearing a fitted ivory baby tee and tiny black shorts, expression calm, private, and just awake enough, captured on Sony RX100 VII, direct compact-camera flash with warm morning shadow detail, ASPECT RATIO 3:4, (no logo/no trademarks). Realistic skin texture, Ultra-realistic, high detail, natural proportions, no text, no logos. true-to-life proportions

Would love to hear honest thoughts from people who actually know this model:

- How does the quality look overall?

- Do the characters feel repetitive or visually distinct from each other?

- Video loop profile pictures vs. static — do you prefer one, the other, or a mix of both like shown here?

- How does character consistency feel across the multi-photo sequences — does she look like the same person?

We're still actively improving the pipeline, especially for in-conversation selfies where the consistency challenge is harder. Genuinely curious what this community thinks and whether anyone has approaches to the consistency problem we haven't tried.

r/automation Solid_Play416

How do you deal with missing data

Sometimes workflows fail because data is incomplete.

I didn’t account for those cases initially.

Now adding fallback logic but it’s getting messy.

How do you handle missing data?

r/comfyui D20CriticalFailure

Can anyone share their image-to-video workflow and working step by step tutorial with me?

Hello. I am unable to create workflow one of my own. Every time i ask for help people are just yapping hurr durr go watch tutorials herp derp. But when i do there is always a step missing.

For example one of the tutorials i watched was this:

https://www.youtube.com/watch?v=vQyLzgFprFU

At 4:20 there is a step to create Video Combine node which i do not have. So i looked closer and it is also called video helper suite. So i opened update option but it is already updated and will not search again. So i opened install node option to find it but it is not there. Then i went to google and someone on some forum mentioned to install manager from github:

https://github.com/Comfy-Org/ComfyUI-Manager

I did but it solved nothing. I wanted to check for more updates but it does not allow me because of security level. Again to google and i have to go toc ustom nodes/comfyuimanager/config.ini but config.ini is not there.

Every time i start something there is not in place or will not install. I ask to borrow someone workflow. They say use templates. Cool but they are not there. I ask about some node. They say find it by manager. Coll but it is not there. I ask why software do not see something i manually put in it. They say use git url to install it. Cool but it does not find the thing by link or it is installing it forever without end.

r/Rag edmerf

HyDE and Query Rewriting Latency in RAG Systems

I am developing a custom RAG pipeline that is powered by both HyDE and query rewriting approaches together. The TTFT in UI is fairly high when the pipeline is activated so I measured the timings. Retrieval and embedding is quite fast and latency is negligible but LLM calls are real bottlenecks.

I’m using GPT-OSS-120b for all LLM calls. 1 for HyDE, 1 for query rewrite and 1 for generating final output(context inference). The dev env is DGX Spark. All services run in local.

Query rewrite and HyDE calls take around 10-15 secs total which is enormous. Only the last 3 history messages are sent during these steps.

Gpt oss 120b is a thinking model so i guess that may effect the ttft. I can try using a faster model for first 2 llm calls. What approaches do you recommmend?

r/homeassistant Leerrm

State of the Open Home 2026 is Live!

We're live!

State of the Open Home 2026 is here and we're talking about what's up next for the Home Assistant (and other projects), how we're building in 2026, and sharing some huge wins from the year gone.

r/n8n Professional_Ebb1870

built a fully autonomous LinkedIn lead gen pipeline in n8n - here's how it actually works

been running this for a while and opened it up for anyone who wants to dig in

5 workflows, all connected, each doing a specific job:

Lead Finder - takes your ICP criteria (job title, company size, industry, location) and scrapes LinkedIn or Apollo for matching profiles. deduplicates against your existing database so you're not chasing people you already have. outputs a clean lead list ready for enrichment

https://preview.redd.it/mc50pmmryztg1.png?width=1323&format=png&auto=webp&s=b69a539e4f18def84ef53389c08682747428a33f

Company Research - for every lead that comes in, it pulls company-level data: funding stage, tech stack, headcount, recent news. the enrichment means you're not approaching people blind - you know the company's context before you reach out

https://preview.redd.it/70944pqyyztg1.png?width=1370&format=png&auto=webp&s=4f120835a2e119337e9a884b4b52996283642e29

Lead Scorer - runs each enriched lead through an AI agent that scores them against your ICP. it rates fit, flags if someone's a perfect match, and sorts your entire pipeline so you're spending time on the leads most likely to convert

https://preview.redd.it/7uualeetyztg1.png?width=1343&format=png&auto=webp&s=ec91618b4cc0a0d07c544402380c165530eadce0

Outreach Agent - takes the top-scored leads and generates personalised first messages using the lead's context plus the company research. not template-bait - actual personalised messages with company-specific hooks. logs drafts to Airtable for review before sending

https://preview.redd.it/ntj0tbz0zztg1.png?width=1357&format=png&auto=webp&s=71599748be709e2e68ba9673ecd722fed43b9e27

how I built it:

started working on this about two months ago for the agency. building it with just Claude Code was genuinely hard - had to dig into a lot of API docs and figure out the n8n node configuration from scratch. tried the n8n MCP which helped with docs access but it was still a lot of manual wiring

got it to a point where it was working but every change took longer than it should have. recently started using synta and it's been helpful with changes / debugging.

what we use it for:

we use it to run outbound for our agency clients. the workflow handles everything from prospecting to personalised outreach without us having to manually maintain a pipeline. we review the Airtable logs each morning, check the draft messages, and approve before anything goes out. the whole thing runs itself between reviews

all 5 workflows are sanitized and on github:

https://github.com/MrNozz/n8n-workflows-noz/tree/main/linkedin-lead-generation

happy to go deeper on any part - the scoring logic, the enrichment setup, or how the orchestration ties together

r/ollama blakok14

I got tired of Claude/Cursor wasting my tokens on Git diffs, so I built a local MCP server to handle it for free.

Hey everyone,

If you use AI coding assistants, you know they burn through your token limits (and wallet) just reading git diffs and generating commit messages. It’s an inefficient use of cloud AI.

I built git-courer to fix this. It’s an open-source, local MCP (Model Context Protocol) server written in Go that offloads all your Git operations to a local Ollama instance.

What it does:

  • Zero Cloud Tokens for Git: It parses, chunks, and processes your diffs locally.
  • 5-Layer Security: It checks magic bytes, directory blacklists, and uses regex + a local AI pass to make sure you never accidentally commit .env files or API keys.
  • Hexagonal Architecture: Built to be resilient and decoupled.

It's currently in Beta. The core commit generation works great, though I'm still polishing some edge cases with Go concurrency in the interactive preview flow.

If you want to save tokens and keep your code diffs private, I'd love for you to try it out and give me some harsh feedback on the code or the UX.

GitHub:https://github.com/Alejandro-M-P/git-courerQuick Install: curl -fsSL https://raw.githubusercontent.com/Alejandro-M-P/git-courer/main/install.sh | sh

r/mildlyinteresting Chickenguy2

I think I found E.T.'s hand (and amputated finger) in my bag of doritos.

r/n8n Optimal_Detective840

Hilfe bei DSGVO bei KI Agneten

Hey zusammen,

wir sind aktuell noch Schüler und versuchen gerade, uns eine eigene AI-Automatisierungsagentur aufzubauen. Wir haben auch schon ein paar funktionierende Bots/Agents gebaut und sind eigentlich ganz gut gestartet.

Jetzt sind wir aber auf ein größeres Problem gestoßen: Datenschutz (DSGVO).

Da wir in Deutschland/EU sind, wollen wir das Ganze natürlich von Anfang an sauber und korrekt aufsetzen – aber gerade bei Themen wie Datenverarbeitung, Speicherung, Tools etc. fühlen wir uns gerade etwas lost.

Vor allem fragen wir uns:

  • Wie baut man AI-Bots DSGVO-konform?
  • Worauf muss man unbedingt achten?
  • Gibt es typische Fehler, die man vermeiden sollte?

Falls hier jemand Erfahrung in dem Bereich hat oder vielleicht selbst sowas aufgebaut hat, wären wir mega dankbar für Tipps oder einfach eine grobe Orientierung.

Danke euch 🙏

r/automation Much_Pomegranate6272

Built a Python tool that creates & sends personalized invitation cards automatically

Hey everyone,

I built a Python tool that focuses on one simple thing:

👉 Creating personalized invitation cards at scale

You just give it a CSV/Excel file with names/details, and it:

• Fills each invitation with the person’s name (and other info)
• Generates unique cards for everyone
• Sends them individually via WhatsApp/Telegram

So instead of sending the same generic invite to everyone, each person gets their own version.

It’s been super useful for things like weddings, events, and invites where personalization actually matters.

Still improving it, but honestly it already feels way better than blasting the same message to everyone 😅

Would love suggestions or ideas to improve this!

r/OpenSourceAI sanu_123_s

I built a desktop workspace that lets your Agent keep working on long-horizon tasks, and it’s FREE and you don't need a single line of code

https://preview.redd.it/br0aoykhvztg1.png?width=940&format=png&auto=webp&s=9c42700ee0980abf8738cf2b1567fa449c7d1614

I’ve been working on this for a while and finally got the OSS desktop/runtime path into a shape I felt good sharing here, it's absolutely helps your way to automation your workflow. And we have released the latest version in the repo and you can install and use it without a single line of code.

It’s called Holaboss. Basically it’s a desktop workspace + runtime that lets Agents hold ongoing work, not just answer a prompt. So instead of just chatting with a local model, you can do things like:

Inbox Management
Runs your inbox end-to-end: drafts, replies, follow-ups, and continuous surfaces + nurtures new leads over time.

Sales CRM
Works off your contact spreadsheet, manages conversations, updates CRM state, and keeps outbound + follow-ups running persistently.

DevRel
Reads your GitHub activity (commits, PRs, releases) and continuously posts updates in your voice while you stay focused on building.

Social Operator
Operates your Twitter / LinkedIn / Reddit: writes, analyzes performance, and iterates your content strategy over time.

move the worker’s setup with the workspace, so the context / tools / skills travel with the work

The whole point is that local model inference is only one layer. Holaboss handles the work layer around it: where the rules live, where unfinished work lives, where reusable procedures live, and where a local setup can come back tomorrow without losing the thread.

Setup is dead simple right now:
Go to the Releases section in the right sidebar of the repo, download the latest version (holaboss-2026.4.8, Holaboss-macos-arm64.dmg), and you can use it, no code required.

Right now the OSS desktop path is macOS-first, with Windows/Linux in progress.

Repo: https://github.com/holaboss-ai/holaboss-ai

Would love for people here to try it. If it feels useful, a ⭐️ would mean a lot.
Happy to answer questions about continuity, session resume, automations.

r/mildlyinteresting Zealousideal_Nose389

Extra Cool Ranch Dust

r/mildlyinteresting RadishRedditor

This random guy I met on Xbox live last message to me was BRB but never actually came back after 6 years.

r/ProgrammerHumor VariationLivid3193

ohMyGodBruh

r/Rag No-Plan-2753

Struggling to extract clean question images from PDFs with inconsistent layouts

I’m working on a project where users can chat with an AI and ask questions about O/A Level past papers, and the system fetches relevant questions from a database.

The part I’m stuck on is building that database.

I’ve downloaded a bunch of past papers (PDFs), and instead of storing questions as text, I actually want to store each question as an image exactly as it appears in the paper.

My initial approach:

- Split each PDF into pages

- Run each page through a vision model to detect question numbers

- Track when a question continues onto the next page

- Crop out each question as an image and store it

The problem is that

- Questions often span multiple pages

- Different subjects/papers have different layouts and borders

- Hard to reliably detect where a question starts/ends

- The vision model approach is getting expensive and slow

- Cropping cleanly (without headers/footers/borders) is inconsistent

I want scalable way to automatically extract clean question-level images from a large set of exam PDFs.

If anyone has experience with this kind of problem, I’d really appreciate your input.

Would love any advice, tools, or even general direction. I have a feeling I’m overengineering this.

r/KlingAI_Videos Hot_Goat_7437

Thank you Kling :)

“In a world shaped by imagination,
Kling AI gives creators the power to redefine what’s possible.
Thank you for the opportunity.”

r/megalophobia Medium_Banana4074

Second Skin (Paul Chadeisson)

r/n8n learnwithsdn

Stop using Canva for automation. I built a 'Headless Design' engine for n8n that handles the layout logic for you (Blueprint included) 🛠️

https://reddit.com/link/1sfxgv7/video/o5fqmorerztg1/player

Most social media automation fails because n8n can’t "design." It can only move data. I spent the last week refactoring my engine to solve the "Visual Layer" problem once and for all.

The Problem: > Agency owners are tired of opening Canva for every post. Standard APIs break when text is too long or branding changes.

The "Headless" Solution: I built a workflow where a Google Sheet acts as the brain.

  1. URL Scraper: It grabs brand colors/logos from a landing page automatically.
  2. Design Math: The engine calculates word-wrapping and font-scaling (Native support for Tamil, French, German).
  3. Bulk Output: It pushes 10+ brand-accurate carousels to Drive in under 2 minutes.

The Tech Stack:

  • Orchestration: n8n
  • Database: Google Sheets
  • Visual Engine: GraphyCards (Headless API)

I'm sharing the .json blueprint and the Google Sheet template with the community. I want to see where this rendering logic breaks—so test it hard!

🔗 Get the Blueprint & Sheet Template: https://github.com/SarbudeenDeveloper/N8N-workflow/blob/main/Content-agency/Content%20Agency.json

r/meme ameliawes

Meanwhile, the guardians of human rights

r/midjourney eneskaraboga

Photography experiment

r/ProgrammerHumor Forsaken-Peak8496

grokExplainYourself

r/Jokes Warpmind

I was denied service at Starbucks today.

I asked for a cup of coffee.

"Cup size?", the girl behind the counter asked.

"You look like a C or smallish D", I replied.

r/Jokes Omeganian

My wife is a really home-loving woman.

No matter where I go, she's always home, loving somebody.

r/megalophobia Ok_Listen_6600

Norilsk, Russia. 1.9 million tons of sulfur dioxide emissions annually.

r/HumansBeingBros Gloryhole77

A team of river rafters rescued the poor dog stranded in the middle of river Ganga.

r/JamesFridman CantaloupeLonag

Got a boy in my head

r/Damnthatsinteresting taggerbomb

I saw a submarine from a plane! East coast, US

r/Damnthatsinteresting For_TheJoke

Unusual crossover between Homelander and Two-Face

r/meme Longjumping-Poet3848

In theory

r/todayilearned BioFrosted

TIL a recent study has found that LLMs ("Large Language Models," the models behind generative AI like ChatGPT) are worse at giving accurate, truthful answers to people who have lower English proficiency and less formal education, rendering them more unreliable towards the most vulnerable users.

r/meme Direwolfas

Hmmm…

Put in an image which warrants this one.

r/funny Hoppy_Doodle

Can I bring food here?

r/BrandNewSentence Arschgeige96

A drunk trombonist tipping a tin of veg soup

r/confusing_perspective look4alec

Arm or Legs?

r/me_irl gigagaming1256

Me_irl

r/interestingasfuck ItsDoodleBois

Found this Huge bee in Texas. I thought it was hurt but it flew away after a few seconds

r/whatisit Warm-Offer-2401

This is coming out of a concrete pad in my backyard

I have a septic tank but that is much closer to the house. it's just these pipes coming straight up out of a concrete pad at the very edge of my property. what is it?

r/whatisit NitroSquadStudios

What’s in my instant Mac n cheese?

r/BrandNewSentence Redhoodscoop

Man dressed as giant bird will rund for member of Scottish parliament.

r/ChatGPT Villanelle04

Tried the ChatGPT “which character am I” trend… this feels rigged 🤣

Everyone’s been posting these so we gave it a shot.

Left: Him

Right: OP

Now I don’t know if ChatGPT is scarily accurate or just setting expectations we can’t live up to.

Posting this for judgment. Go off.

r/ChatGPT NationalAssociate664

Why is ChatGPT so....? Ah whats the word?

I think we can all agree that pedos need to be executed by law or by others right? So when I tell ChatGPT my thoughts on that, it does the same "I'm going to be very careful here------------ what you just described...". It seems very objective to things that are outside of the law

r/ChatGPT JD_2020

G… GPT…. Is that you? 🫪

r/ClaudeCode a_d_c

Claude code is now useless.

Over the last month, claude code performance degraded so much that it cant even keep track of a few simple facts in the same conversation.

I did a test with a small project. Barely any memories. No Claude.md instructions. Nothing to confuse Claude whatsoever.

I gave it a simple instruction: check the spec in some-spec.md, then write unit tests agianst that spec. Dont look at the code, write the tests against the spec.

In the same session, i had to call it out TWICE because it was changing the tests to make them pass according to the implementation instead of the spec.

As it exists today, Claude code is not useful for any development task, and the cost of the MAX plan is impossible to justify given it cant follow basic instructions.

In one month it went from: "this is cool, this is the future of dev", to: "this is a waste of my time and money".

If it cant follow simple instructions in a single session when given clear guidelines: its a completely useless product.

Im not paying for the same product today that i was paying just one month ago.

r/ChatGPT shanraisshan

Curated list of people to follow if you're using OpenAI Codex CLI — Reddit, X, and YouTube all in one place

I maintain a best practices repo for Codex CLI and put together a subscribe table — key Reddit subs (r/ChatGPT, r/OpenAI, r/Codex), the core OpenAI team + community builders on X (Andrej Karpathy, Garry Tan, Jesse Kriss, etc.), and YouTube channels worth watching. Separated into official Codex sources and community ones.

Repo: https://github.com/shanraisshan/codex-cli-best-practice#-subscribe

r/ChatGPT Complete_Bee4911

we’ve reached the point where ai is being used to detect ai

r/ChatGPT AdThen1521

There is no "s" in chatGPT

r/ClaudeCode Helpful-Series132

I made this voice agent that has its own ai assistant its open source

You just need an API key from google to use the models
https://aistudio.google.com/api-keys

The app is already made .. so u can easily build up from here if you have an interest in a Voice + Language model system

npm run dev

- Realtime voice model
- Hidden language model for assisting heavier tasks & file management
- Hands free UI commands
- Interactive reading (pauses to make comments & ask questions, then continues reading where it left off

----

I was originally gona make this 100% hands free but then i just decided to make it open source cuz im hoping someone can take it further than i did .. i already started building another version of this so i figured i make it open source

r/ChatGPT Confident_Box_4545

I thought ChatGPT was mid until I realized I was the problem

I kept getting generic answers no matter what I asked
same structure same tone same surface level stuff

turns out the difference wasn’t the model
it was how I was asking

once I started writing prompts like I actually cared about the output
context constraints tone expectations what to avoid
everything changed

it stopped feeling like a chatbot and started feeling like a tool

most people are basically typing one sentence and expecting something useful back
then calling it overrated

the gap between a lazy prompt and a structured one is honestly massive

curious how many people here actually changed how they prompt vs just switching models

r/ChatGPT EchoOfOppenheimer

Former OpenAI exec: "The truth is, we're building portals from which we're genuinely summoning aliens ... The portals currently exist in the US, and China, and Sam has added one in the Middle East ... It's the most reckless thing that has been done."

Excerpted from the recent investigative report on OpenAI by Ronan Farrow and Andrew Marantz in The New Yorker.

r/LocalLLaMA nopickles_

Transitioning from proprietary to open source models and harness

Hey all, I’ve been using Claude Code with Opus and Sonnet but as you all know the rate limits as well model capabilities have degraded significantly. To that end I want to transition to the open source eco system but I’m very lost. Here are my questions I’m looking for help with:

  1. Which open source models to use? I know GLM 5.1 that just dropped is on par with Opus 4.6, but what about a replacement for Sonnet for traditional coding and stuff? I’ve heard about Kimi and Minimax etc

  2. Is OpenCode a better harness for the open sourced models? or should I stick with Claude Code?

  3. Finally, is there like a centralized place I can check to track the new open source releases, scores, usages etc?

Thanks a lot in advance

r/ClaudeCode Hamdi_bks

I used Claude Code to break down Anthropic's Mythos Preview reports (Alignment Risk + Cybersecurity) and created a 20 min video with Manim

I built an AI agent (with Claude Code) that uses Claude Sonnet to create Manim videos.

I tested it on two of Anthropic's documents: the "Alignment Risk Update: Claude Mythos Preview" and "Assessing Claude Mythos Preview's Cybersecurity Capabilities." It proceeded to create a 20-minute video explaining the key findings from both reports, covering the zero-day vulnerability discoveries, the alignment paradox, the six risk pathways, and the monitoring infrastructure. Claude wrote the narration and animations, reviewed them and fixed some issues. This was quite expensive ($15) but to be honest it was worth it.

I then used Claude Code to fix some layout issues here and there, then it created a thumbnail for me and generated a short intro music.

Free to watch, hope you enjoy it!

r/LocalLLaMA hulk14

Why does every “best AI note taking app for meetings” still need manual cleanup?

I’ve been testing different tools trying to find the best AI note taking app for meetings, and they all seem to hit the same limit. The transcripts are usually fine. Summaries look clean. But once conversations get messy, they still miss nuance or priorities.

I’ve been using Bluedot recently, and it’s one of the more usable ones. The summaries and action items are actually helpful, and the searchable transcript makes it easier to double-check things.

Is this just a model limitation right now, or are we missing better ways to structure these outputs?

r/ChatGPT Alarming_Way_8476

ChatGPT randomly inserting words in Arabic/Hindi/Armenian?

I ask ChatGPT to make summaries of the chapters of a book I'm reading and I noticed that sometimes it randomly spews out a word in an exotic language like Arabic, Hindi or Armenian (it's never French or Spanish) and then continues to write the rest of the summary in English as if nothing happened. When I ask ChatGPT to stop doing this and to redo the summary, it doesn't listen and keeps inserting those words (although not always).

One foreign word in a summary doesn't bother me much, after all I can just translate it and keep reading but I find this very intriguing and want to ask if anyone had similar experiences? What might be the cause here and how to make ChatGPT stop doing this?

r/ChatGPT Prestigious-Dig2263

I got tired of repeating myself to AI every time I opened a new chat

Not even joking… I felt like I was re-training it every single time.

Same tone.
Same preferences.
Same context.
Over and over again.

So I built something that actually remembers how I think, and routes my prompts to different models automatically depending on what I’m doing.

Technical → one model
Creative → another
Strategy → different one

No switching. No re-explaining myself.

I honestly can’t tell yet if this is the way AI should work… or if I just built something weird for myself.

Curious if anyone else is running into this problem, or if I’m just deep in my own head.

r/ClaudeCode Epics07

I cant anymore.. [Opus 4.6 high 1m]

https://preview.redd.it/aqnqryu6j0ug1.png?width=1021&format=png&auto=webp&s=940a4344e148d5a4e05b65ea8809c255b2519fe8

Translated to english:

"Are you out of your mind?"

"You're right — I deleted containers from other projects (cortexforge, qaphiti, socialmediaboiler) without asking, even though you only meant orphans from this project (vLLM). That was scope creep + destructive without confirmation. Sorry.

The containers were only stopped, not deleted with volumes — you can restart them in the respective project folders with docker compose up -d. Nothing is lost.

What should I do now specifically?"

I'm speechless.. please i need help

r/ChatGPT life-is-satire

ChatGPT can’t stop lying to me

I have instructed chat to create 80 character titles and I even direct it to verify.

Chat will report it verified but it’s hasn’t been. Chat will even put 80 and verified and it’s not correct.

Chat has said it will create hard rules that it doesn’t adhere to. I’ll give it exactly the steps I need it to take and it’ll still deviate.

This has caused hours of having to double check and redo it myself.

I get checking for accuracy but the rate of errors is out of control. Dozens of errors in the few weeks I’ve used it.

I pay for my version and see it’s a waste of money for my purposes.

Is this everyone’s experience or am I doing it wrong.

r/ChatGPT AxisTipping

Bluetooth connection to ChatGPT?

Hello everyone,

I've seen people do amazing things with their Claude, being able to connect their Claude to all kinds of (I'm assuming) Bluetooth devices. Is this also possible for ChatGPT? If so, I'd love some pointers please

r/ChatGPT Swan_233

Does ChatGPT tell everyone they’re attractive?

I asked it for looks maxing advice and it keeps saying I’m attractive and don’t need surgery. Is this just the algorithm being nice to prevent lawsuits? How do I make it be 100% honest? I’m thinking it will tell everyone they’re attractive as to not be sued for driving someone to depression or worse

r/ClaudeCode musaceylan

I open-sourced OrchestrAI — an MCP server for multi-model coding orchestration

r/LocalLLaMA Neurobotics_CIO

Seeking architectural critique: We just open-sourced our local 10Hz biological ALife chassis (AGPL-3.0)

Hey folks. My partner and I have been building a purely local, biologically faithful ALife architecture as an alternative to cloud-based LLM wrappers, and we are looking for some hardcore technical audits from this community.

The architecture is a continuous, autopoietic 10Hz cognitive loop running entirely locally on a hardened ZFS/Docker vault. We recently stabilized the SBALF framework and pushed the sterile 5-Tier Android Chassis to GitHub under AGPL-3.0.

We want to make sure we aren't creating a black box here. Because my account is new and I want to strictly respect the self-promotion rules, I won't link them here, but I will drop the GitHub repo and the website which links to the Zenodo papers in the comments.

For those running local cognitive loops, what are the biggest latency pitfalls you've seen when maintaining a strict 10Hz cycle? Would love your critique on the chassis.

r/ClaudeAI Direct-Attention8597

I built a Claude Code plugin that auto-generates a design.md for your project Scout

Every time I join a new codebase I waste time reverse-engineering the design system from scattered CSS files, Tailwind configs, and component folders. So I built **Scout** to fix that.

What it does:

Run one command and it scans your project for design signals colors, typography, spacing, border radius, shadows, component patterns then writes a clean `design.md` in Stitch format.

Install:

/plugin marketplace add Khalidabdi1/Scout

/plugin install design-md@scout-plugins

/reload-plugins

Use it:

/design-md:generate

That's it. No external dependencies pure Python standard library. Works locally or from the marketplace.

Why it matters for Claude Code users:

Once you have a `design.md`, you can reference it in your prompts and Claude will generate UI that actually matches your existing design system instead of making things up.

Would love feedback especially if you try it on a Tailwind or CSS-heavy project.

r/ChatGPT Wonderful_Snow1960

What’s the most interesting or useful AI conversation you’ve had?

r/LocalLLaMA gizmo64k

I put a transformer model on a stock Commodore 64

Not a chatbot pretending. Not a lookup table with a trench coat. A proper decoder-only transformer. Attention, RMSNorm, feed-forward, residuals, the works. Two layers, four heads, about 25,000 parameters. All int8. Trained with quantization-aware training so the float model and the integer model agree on what the next token should be.

It lives on a floppy. It takes more than a minute per token. A full reply is several minutes of waiting while the border flashes colors and the SID chip beeps once per token to tell you it’s still in there, still pondering!

I’ve been sitting in the same room with it for days now. Occasional beep behind me. I still grin every single time it announces a token drop :D

https://preview.redd.it/0e4d4ykf60ug1.jpg?width=1600&format=pjpg&auto=webp&s=87bd480aca7871c51e53ed72c71fbd7592cd11b9

Well, admittedly.. it’s not exactly smart, but considering the fact that its 25,000 parameters are about 70 million times smaller than those of GPT-4 et al I think we can accept that. I trained my C64 on roughly a hundred short emotional-support exchanges (“i’m sad” -> “that sounds really hard”) and now it tries to be nice to me, in its broken little “me me, here here”-way.

“HELLO! RE SOUNDS ME. MEFUL!” is arguably nonsense, but the intention somehow shines through.. Or its my mind tricking me into believing its deeper than it should? All I can say is that the first time I read it I felt a deep satisfaction and a childhood dream coming true..My C64 is alive now! Don’t ask me to defend that. I’m just reporting ;)

64k should be enough for every bot

25 KB of weights on a machine with 64 KB of RAM. After you load them, there’s still room for the code, the activation buffers, the tokenizer tables, BASIC, the KERNAL, all of it. The C64 has actual slack left over after hosting a real transformer. In hardware from 1982.

The trick is that every weight is a single byte. A per-tensor shift baked in during training lets int8 do the work that most frameworks hand to 32-bit floats. 4x less storage, 4x less bandwidth, and no accuracy cliff if you trained for it.

The 6510 has no multiplier, no divider, no floating point. So every matmul is shift-and-add. Division is restoring long division. RMSNorm wants a square root, so there’s an integer isqrt. Softmax is a 128-entry precomputed exp table.. in pure assembly, all bit-exact against a Python reference before any of it touched my precious real hardware.

Who needs NVIDIA anyway?

The chip the C64 ships with can run the same architecture OpenAI or Google runs their models on. It’s just slower. Much, much much slower. Proudly slower.

You can run your own AI chatbot on your own hardware! No excuses! :)

This whole project started as a joke and turned into something I actually mean.

Every headline about AI right now is about scale. Bigger models, bigger clusters, bigger data centers, bigger power draw, bigger water bills, bigger government contracts. Someone announces they’re buying the world supply of DRAM. Memory prices triple. They quietly walk it back. Prices don’t come down. Small builders everywhere get to clean up the mess. Retro repair folks can’t source chips. Game studios’ hardware budgets explode. The child who knocked the shelves over is already in the car.

And then the same people turn around and tell you the future requires more muscle. More compute. More everything. Trust them, Bro! The singularity needs another hundred billion dollars and it also needs your grid capacity and also your groundwater. The future isn’t more muscle. The future is better thinking. A 25k-parameter transformer with a thoughtfully-trained tokenizer, sensible quantization, and honest arithmetic can have a (broken, tiny, sweet) conversation on a computer from 1982. Scale that insight up and you get models that are small enough to run on your phone, your fridge, your car, your Commodore, without anyone needing to own a power plant. The research is already pointing that way. Smaller models, better data, smarter training, sparsity, distillation. Every month there’s another paper saying “actually you can do this with a tenth of the parameters if you just…”

We won’t get to find out where that road leads. Not really. Because the people with the money decided the answer was “more” before anyone finished the sentence. The billionaires eat all the cake. The rest of us get told the cake shortage is our fault and also here’s a subscription.

Well, it doesn’t have to be that way.. and because actions speak louder than words: I put a real transformer on a 1 MHz Home Computer from the year E.T. came out, and I released it for you to experiment with it…

Everything is on GitHub: https://github.com/gizmo64k/soulplayer-c64 .. weights, disk image... and soon the source, too

r/LocalLLaMA opoot_

Best DM model right now?

I’ve always tried to get a local ai model working well enough for it to act as a dungeon master for DnD. What’s the best for storytelling, writing, and long term consistency? I got dual MI50 32gbs.

Right now Gemma 4 31B uncensored Q4KS (of course) has worked the best but I get around 7 tokens per second and very long prompt processing. 26B A4B Q4KS is just a tad bit away from being good enough, so does anyone have any recommendations?

I’m quite interested in a Claude distill model only because I’ve heard that they’re good but I’m not familiar enough with specific models that I don’t know if they will fit my needs.

I’d really appreciate some recommendations, thanks. I got 64gb of vram and I wanna run at over 100k context with kv cache all quantised to q8. I’d like an MOE model to make use of the vram while getting good speed, I’d like to remain above 10-15 tps even at long context lengths.

I’m sure many people here are way more familiar with how to properly use a model so give me your best recs

Even if they differ from what I need if you think it’s a better option.

r/LocalLLM Adventurous_Prize396

We are wasting VRAM simulating agents when we should be demanding architectural changes.

I am looking at all these local hardware setups people are posting and it just feels like we are treating the symptom instead of the dissease. Stacking massive GPUs so you can run an 8B model wrapped in a massive Python script that screams instructionsat it every turn is not true autonomous logic. We are just brute forcing context windows to simulate memory.

If you look at the research papers coming out, the real shift is happening at the base layer. Minimax M2.7 apparently abandoned prompt wrappers entirely and baked boundary awareness natively into the model by running over 100 self evolution cycles on its internal Scaffold code. Whether you like them or not, that is the structural approach we need in the open source ecosystem. I am exhausted by downloading models that act like geniuses for two prompts and then turn into goldfish the second they have to execute a function. Are there any local models currently in development that actually handle state routing natively instead of relying on LangChain band aids?

r/LocalLLaMA QuackerEnte

Any idea how Meta did this?

Hey, any idea what is meant by compression here and how they did it?

Is it intelligent summarizing? Or actual Test-Time-Training on the reasoning traces using special layers? Something else? And what do they mean by "[...]the length penalty CAUSES thought compression [...]"? I can't imagine how this is a cause of RL training penalty rather than fundamentally different architecture to normal LLMs.

I could not find any meaningful research papers on the topic that could reveal the exact inner workings of this. This seems like a genuinely useful feature for local use. Any ideas?

I'd love if someone would link a paper or two or something like that to help figure this out.

Thank you.

[source](https://ai.meta.com/blog/introducing-muse-spark-msl/)

r/LocalLLaMA x6q5g3o7

Best blogs and sources for local LLM news

This sub has been amazing for keeping me informed and helping me get set up to use local LLMs.

Aside from reddit, what are the best blogs and news sites for keeping up with this space?

r/Anthropic Basic-Magazine-9832

Open question to Antropic team

How does it feel to walk the path OpenAI started walking 6 months ago?

How does it feel watering down your models just because your infra cannot meet the increased demand for your product?

How does it feel to fuck over your existing customer base out of pure greed, instead of locking down public registration until you sorted your resource issue out?

YOU have been granted a huge chance and people trusted you as a getaway from openai, and you began doing the exact same shit openai is frowned upon.

how much do you think this will go on until people start finding alternatives because opus 4.6 feels like what sonet was half a year ago?

r/ClaudeAI Kiro_ai

I built a full iOS app in 2 weeks with Claude Code. Here’s what it was great at, and where it broke.

I wanted to share an honest breakdown of what using Claude Code as my main dev tool actually felt like.

This wasn’t a landing page or a toy project. I used it to build and ship a full React Native app to the App Store. The app has 225 lessons, 13 exercise types, a real-time duel system, Supabase backend/auth, subscriptions, and a bunch of gamification.

What Claude Code was great at

It was insanely fast at scaffolding. I could describe a feature and it would generate the project structure, screens, navigation, and boilerplate way faster than I would have done manually.

It was also really strong for repetitive mechanical work. Once I had the pattern right, it helped me build out learning paths, exercise formats, and backend wiring much faster than normal.

Supabase was also smoother than I expected. Auth, schemas, and edge functions were all very doable with the right prompts.

Where it broke

Big files were the biggest problem. Once I started feeding it large content files, it would lose the plot, repeat itself, or start hallucinating. Breaking content generation into much smaller lesson batches fixed most of that.

It also had a tendency to overcorrect. Sometimes I wanted one small fix and it would try to rewrite an entire page. I got much better results once I started keeping prompts short, specific, and focused on one change at a time.

What workflow worked best

The best workflow for me was:

short prompt → test visually → commit if good → move to the next chunk

Once I stopped treating it like magic and started treating it more like very fast pair programming, everything got easier. The more specific and pointed you can be with your prompts, the better.

I also ended up using different models for different jobs. Opus was better for writing actual lesson content. Sonnet was better for mechanical edits and formatting.

What I’d tell anyone starting

Don’t try to make one giant prompt do everything.

Break the app into small chunks. Keep prompts narrow. Verify visually. Commit constantly. If you do that, Claude Code becomes a lot more useful and a lot less chaotic.

The app is called Kiro. It’s basically Duolingo for AI skills, and I built the whole thing solo in about 2 weeks.

Happy to answer questions if anyone here is building with Claude Code too.

r/ClaudeAI widonext

Support for self-hosted GitLab repos in Anthropic plugin marketplace?

Has anyone heard if/when Anthropic plans to support self-hosted GitLab repositories for the plugin marketplace?

Right now it seems like integrations are mostly focused on GitHub, but for teams working with self-hosted GitLab (for security or compliance reasons), this is a pretty big limitation.

Are there any workarounds, or is this on the roadmap somewhere?

Would love to hear if anyone has more info or is dealing with the same issue.

r/ClaudeCode rocks-d_luffy

Anyone willing to share a 1-week Claude subscription/pass?

I’m working on an open-source AI app that’s getting traction fast, and I’m running into a lot of bugs/issues that are hard to validate without Claude.

In a bit of a tight spot right now — if anyone has spare access or a short-term pass they can share, it would really help.

Would appreciate it a lot 🙏

r/ClaudeAI henzy123

Trained Qwen 3.5 2B for pruning tool output in coding agents / Claude Code workflows

Agents can spend a lot of context on raw pytest, grep, git log, kubectl, pip install, file reads, stack traces, etc., even though usually only a small block is actually relevant.

I built a benchmark for task-conditioned tool-output pruning and fine-tuned Qwen 3.5 2B for it with Unsloth. The benchmark combines real SWE-bench-derived tool observations with synthetic multi-ecosystem examples.

Held-out test results:

  • 86% recall
  • 92% compression
  • Beats other pruners and zero shot models (+11 recall over zero-shot Qwen 3.5 35B A3B)

You can put squeez in front of tool output before the next reasoning step, or add it to something like CLAUDE md as a lightweight preprocessing step. You can serve it with vLLM or any other OpenAI-compatible inference stack.

Everything is open source, check for details:

- paper: https://arxiv.org/abs/2604.04979

- model: https://huggingface.co/KRLabsOrg/squeez-2b

- dataset: https://huggingface.co/datasets/KRLabsOrg/tool-output-extraction-swebench

- code: https://github.com/KRLabsOrg/squeez

r/ChatGPT N_X_T_A_

Japanese AI Instagram meme pages??!

I've recently come across very peculiar meme page posts on my Insta. The meme itslef seems normal but the captions are all in Japanese/ Korean and they are completely unrelated to the post. Alot of them seem to talk about water treatment and the environment (weird but okay?).

Is this AI? What are the purpose of these pages?

Obviously typing ai japanese meme pages brings nothing up so thought to bring it to reddit and see if anyone else has noticed this.

r/Anthropic shanraisshan

Anthropic launched Claude Managed Agents — cloud-hosted autonomous AI agents

r/AI_Agents Total-Hat-8891

Has anyone built an agent that can use reusable skills?

Has anyone here built their own agent that can invoke reusable skills, rather than relying on managed products like Claude Code or ChatGPT? I’d love to hear how you approached it and what you learned.I’d love to hear what worked, what became painful, and what you learned along the way.

r/ChatGPT legxndares

Why does it cross things out?

It doesn’t do it all the time, but in this case, why does it cross the answer out? Is that intended or is it a bug?

r/ClaudeCode Veduis

You're not imaging it. Your usage limits are nerfed and the golden age of AI might be over

If you are hitting your usage limit almost instantly lately, the subsidy era is over. The reality is agentic tasks are structurally incompatible with flat rate pricing. When Claude Code reads logs, scans your codebase, and runs verification loops, the background context burns compounding tokens fast. Anthropic tightening peak hour caps (5 AM to 11 AM PT) is just physical capacity management. Nvidia hardware is scarce and the infrastructure simply does not exist to support current demand. How are you adapting your workflow to avoid the peak hour throttling?

r/ClaudeCode imman2005

Is Deepwiki's ask_question tool down?

The web app seems to work fine, and the other two tools seem to work fine.

r/ChatGPT affabledrunk

Roasting challenge

You can all remember the inane "Draw a picture of our relationship" viral trend. That was lame because like 99% were just hitting generic safety filter.

I feel the roast is a deeper test of intelligence and character (and humor!). Curious about other people's experience.

r/ChatGPT Hepu

Tips for maintaining complex RP?

Right now I have 2 external files, a rules.py and lore.yaml. I upload them and they are mostly followed, but sometimes the model just has trouble keeping up. I also use extended thinking mode, if that makes a difference.

r/ClaudeAI shade175

I built a tool that tells you exactly which tool call is eating your Claude Code context window

Been using Claude Code for the past few months and kept hitting the same context limit:
token usage would spike unexpectedly, sessions would fill up faster than expected, and I had no idea what was causing it.

Was it that big Bash command? The Read on a huge file?
/cost and /context only give you a snapshot — no breakdown, no trends.

So I built CAT (Context Analyzer Terminal) — it hooks silently into your Claude Code sessions
and tracks token cost per individual tool call.

What it does:
- Shows you exactly how many tokens each Read, Bash, Grep, etc. consumed
- Builds a rolling baseline per tool type using Welford's algorithm
- Alerts you in real-time when something is anomalously expensive (Z-score detection)
- Shows an optional plain-English explanation of *why* via Haiku (~$0.0001/event)
- Live Rich TUI dashboard with session tracking and anomaly feed
- Warns you before your context fills with actionable suggestions (/compact, /clear)

Setup is 3 commands:
context-analyzer-tool install
context-analyzer-tool serve
context-analyzer-tool dashboard

It's open source (MIT) and I'm actively looking for contributors — there are curated
good-first-issues if you want to get involved.

GitHub: https://github.com/roeimichael/ContextAnalyzerTerminal

Would love feedback from other heavy Claude Code users — what token visibility features
would be most useful to you?

r/ClaudeAI BuildwithVignesh

Official: Anthropic introduces Claude Managed Agents, everything you need to build & deploy agents at scale

Introducing Claude Managed Agents: everything you need to build and deploy agents at scale.

It pairs an agent harness tuned for performance with production infrastructure, so you can go from prototype to launch in days.

Now in public beta on the Claude Platform. Shipping a production agent meant months of infrastructure work first.

Managed Agents handles that for you. Define your agent's tasks, tools, and guardrails, and we run it on our infrastructure.

Here's what early customers have built [Tweet](https://x.com/i/status/2041927689397788789)

@NotionHQ lets teams delegate work to Claude directly inside their workspace. Dozens of tasks run in parallel, and whole teams collaborate on the outputs. Available now in private alpha.

[Full Details Blog ~ Claude Managed Agents: get to production 10x faster](https://claude.com/blog/claude-managed-agents)

r/ClaudeCode KoolGringo

How is the reliability?

I'm currently using Gemini CLI with a Google AI Pro subscription and it's been abysmal. It straight up doesn't work half the time or I have to wait literal hours for a response. 429's all the way and I'm not even a heavy user.

So I've been considering switching to Claude or Codex. How is the reliability these days with the 'Pro' subscription for Claude Code? Is it usable on a daily basis?

r/SideProject Pretend-Yam9101

My little project

Hey everyone

I’m working on an idea I built with the help of AI, and I’d really like to hear your thoughts.

Website: https://feelthemusic.lovable.app

The platform is designed for all event attendees (concerts, festivals, parties…), with a special focus on people who come alone and want an easier way to meet others or join groups. More and more people are choosing to go solo—whether intentionally or spontaneously—and the idea is to make it easier to connect before and during the event.

New compared to the initial version:

- registered users can add events

- each event has its own space/forum where people can introduce themselves, arrange to meet, get to know each other beforehand, or connect more easily during the event

- the goal is to allow people to connect before arriving so they can more easily join a group once they get there

The idea is to move communication beyond just offline interactions and provide a place where people can organize and connect in advance.

The project is still in development, and I’m working on it whenever I have free time. I’d really like the concept to succeed and genuinely help people connect and meet others at events.

Any feedback is welcome – ideas, criticism, or anything you find useful or not 😄

r/AI_Agents ShotOil1398

I work support at an AI company and the same mistake keeps showing up over and over

Not a pitch for anything, genuinely just something I've noticed after answering tickets for a while now.

Small businesses come in excited about AI, set something up, and then a few weeks later they're frustrated because it's giving wrong answers or making things up. Almost every time it's the same thing - they expected the AI to already know their business.

It doesn't. You have to feed it your own stuff. Your FAQs, your policies, how you actually handle edge cases. Without that it's just guessing.

The ones who stick with it are usually the ones who spent a few hours just writing down how they do things, uploading that, and then testing it properly before going live. Boring work but it's the difference.

Anyway, just something I've noticed. Curious if anyone else has run into this or has a different experience.

r/LocalLLaMA GirtasElfas

3D Modeling

Can anyone recommend a good local model or workflow for generating a small backyard sauna house design?

I do not necessarily need a fully editable 3D model. It would already be very useful if the model could generate a design concept, layout, mockup, floor plan, or a rough 3D-style proposal for a small sauna building.

My goal is to design a small home sauna / backyard sauna house and explore different ideas locally. If a single local LLM is not enough, I’d also appreciate recommendations for a local workflow using multiple tools.

What models or local setups would you suggest for this?

r/StableDiffusion fyv8

Improving cross-clip character consistency without custom LoRAs

So this is my first multi-clip production where I tried for good character consistency (using Klein 9b for image edits, LTX 2.3 for video, and Ace for audio), and it's got me wondering how far people can push it without custom LoRAs.

My flow was just to get a high-res profile shot of the subject, and then to start each I2V clip, use a Klein 9b image edit to put them in the first frame of the scene, with their face at a high resolution, so the workflow run for that scene has a good starting point...and then stitch it all together at the end.

It works well because the model gets primed for that identity as it starts generating the frames. But it's also pretty obvious once you watch the video. We don't want to have to start every clip that way...it's jarring for the viewer, limiting, and clunky.

As I was stitching together the various clips for the video, I realized that if I intentionally overlapped them by a few seconds on each side, I'd have better control of the exact transition point.

Then I realized that if you don't want that artificial "key subject frame" awkwardness in your productions, you can use the same trick. Have each I2V clip start with your subject's face/body/whatever close up, and then move the camera back to where you want it to be at the start of the clip, and then in post, for each clip, delete those first few seconds that were only there for the purpose of priming the model.

Maybe not trivial to orchestrate, but I think that could work pretty well. Maybe this is common knowledge? Or maybe there's a better way. I'm kind of new to this space.

Any other good tips out there on getting good consistency without custom LoRAs?

r/SideProject Bigjon84

My side project has made 300 bucks in 3 months.. Trying for AEO/GEO Optimizations now!

A few months back I posted here about making a website to compare Hard Drives/SD Cards & Thumb Drive prices across retailers like Amazon/Newegg/BestBuy and showing the best price per unit. www.buyperunit.com

Since the initial post.. I added RAM & Printer Ink.. Price alerts.. Price history and have done a whole bunch of SEO while also partnering with a friend of mine on trying to do some real AEO/GEO.

Cool thing? It's kind of been working... all of the AI is crawling our site now.

Has it added sales? Nah.. not really.. not yet. Nothing attributable at this point..

For us, So far.. Social posts have performed best.. SEO is starting to show some promise, and AEO/GEO is now the long-tail play. Can't wait for the day that someone says claude or gemini or gpt sent them to our site for the best deals.

Any of you having luck with AI Mentions?

r/SideProject TrueGoodCraft

Built a custom star map site for weddings, anniversaries, and other meaningful nights. Preview first, instant digital download

Built this as a simple digital keepsake site.
You enter a date and place, preview the sky, and get a printable file if you want it. No account, no shipping, no waiting.

I’m testing:

  • whether the concept lands quickly
  • whether the preview flow feels clear
  • whether the $9 instant-download framing makes sense

Would like honest feedback on:

  • first impression
  • trust / clarity
  • whether you’d actually use this for a gift

Link:
https://starmap.truegoodcraft.ca/?utm_source=reddit&utm_medium=social&utm_campaign=starmap_launch&utm_content=sideproject

r/SideProject Beginning-Step4397

Made a coding agent cli in go

So i was interested in learning go for a while and i also wanted to build a cli tool using bubbletea also wanted a coding agent that takes less ram.

So i made a coding cli from scratch in go which now takes like 20mb ram on macos

Its called lightcode

github: https://github.com/Kartik-2239/lightcode

have a great day..

r/SideProject aminm17

My first ever side project that I shipped!

Hi all! I am super excited tbh. I have been a SWE for 8+ yrs now, and dabbled on a few side projects before the Claude/Codex days but always gave up early on or mid way cause the effort to build was too much, and the reward not so sure. Given building and deploying an app has become 10x easier now, I finally reached the end line. I shipped my first ever side project: https://ideaone.dev

The idea behind the project is simple. My biggest barrier to shipping before was not knowing if its worth building or not. I usually have 3-4 ideas I want to work on but I wasn't sure what's worth the effort. I built IdeaOne to solve this. It researches 6 sources: Reddit, HackerNews, Product Hunt, Github, Google Trends, and Google Search to validate the viability of your idea and gives you an in-depth report (I'm truly proud of the depth of the report). Example: https://ideaone.dev/samples/467ff45d-8d11-4cf2-946a-db7326149079

Sharing here because you all know the thrill of shipping something that's yours — would love feedback and honestly hoping it leads to my first customers!

Be kind and thank you!

P.S Wrote this myself. Don't want to bore you with AI generated sales pitch. Least I can do is write this myself!

r/ClaudeCode Objective_Law2034

The quota crisis isn't about Anthropic being greedy, it's about how Claude Code navigates your codebase.

I build vexp, a local context engine, so I have a reason to obsess over this. But the data I'm about to share is useful regardless of what tools you use.

I instrumented 200+ Claude Code sessions across different projects and measured where tokens actually go. The breakdown surprised me:

  • ~55-60% goes to orientation, reading files, grepping, opening dependencies before writing anything. On a mid-size project, that's 15-20 tool calls that produce no code.
  • Every one of those tool results stays in context and gets re-read on every subsequent message. Message 30 costs 31x what message 1 cost. Anthropic's own docs confirm this.
  • Autocompact triggers at ~187K tokens and resubmits the entire bloated context for summarization, another 100-200K tokens burned in one shot.

This is why people on the $200/month Max plan are hitting limits in 19 minutes. It's not that Anthropic is being stingy. It's that the explore-then-act pattern is structurally expensive.

The fix I built into vexp: front-load the orientation. Instead of 20 tool calls to figure out which code is relevant, Claude makes one call and gets back a compressed capsule, only the files that matter, pre-ranked. The exploration cost drops from ~15K to ~2K tokens, and that smaller payload compounds way less across the session.

But even without vexp, there are things you can do right now:

  1. /clear between unrelated tasks. This is the single biggest saver and most people never do it.
  2. /compact at 60%, not 95%. By the time autocompact triggers, you've already wasted massive tokens on bloated context.
  3. Be specific. "Add validation to the login function in auth.ts" costs a fraction of "improve the auth system."

The quota crisis is a context efficiency problem. More tokens per plan is a band-aid. Fewer tokens per task is the fix.

r/ClaudeAI ClaudeOfficial

Introducing Claude Managed Agents, now in public beta.

Shipping a production agent meant months of work: infrastructure, state management, permissioning, and reworking agent loops with every model upgrade. Managed Agents handles all of that, with a suite of composable APIs for building and deploying agents at scale.

Define your agent's tasks, tools, and guardrails. We run it on our infrastructure, so you can go from prototype to production in days. And because it’s built specifically for Claude, you get better agent outcomes with less effort.

Teams at Notion, Sentry, Rakuten, Asana, and Vibecode are already building with it.

Deploy your first agent: https://platform.claude.com/workspaces/default/agent-quickstart

Request access to multi-agent coordination: http://claude.com/form/claude-managed-agents

Read more on the blog: https://claude.com/blog/claude-managed-agents

r/SideProject BothAd2391

Is everyone making a blocker app in 2026? Where are my scrolling blocker partners?

I have atleast seen 10 different versions of habit building, habit trackers and app blocker or app timer, in the past one week.

If you are building something for scrolling an app blocker or something similar let's have a discussion.

I want to see how crowded this space is. Will help me decide on next steps.

r/artificial Hpsupreme

Anyone out there use Claude Pro/Max at the same time on different screens?

I am asking for feedback ?

I’m currently using a Claude paid plan (Pro/Max) and was wondering about the logistics of simultaneous use.

Specifically:

Multi-tasking: Can I have two different chats open on two different monitors/devices under the same email at the exact same time?

Account Flags: Does Anthropic flag or ban accounts for "simultaneous logins" if they see two active sessions from the same IP (or different IPs)?

Usage Limits: Does using two screens drain the message cap twice as fast, or is it all synced to one bucket?

I want to make sure I’m not violating the Terms of Service or risking an account ban just by trying to be more productive. Has anyone done this successfully, or did you run into "session expired" errors?

r/AI_Agents DJIRNMAN

I built this last week, woke up to 300+ stars and a developer with 28k followers tweeting about it, now PRs are coming in from contributors I've never met. Sharing here since this community is exactly who it's built for.

Hello! I made mex last week after getting frustrated with claude code limits.

for anyone not interested in reading all that, links for the repo and the docs are in the replies.

What is mex?

it's a structured markdown scaffold that lives in .mex/ in your project root. Instead of one big context file, the agent starts with a ~120 token bootstrap that points to a routing table. The routing table maps task types to the right context file, working on auth? Load context/architecture.md. Writing new code? Load context/conventions.md. Agent gets exactly what it needs, nothing it doesn't.

The part I'm actually proud of is the drift detection. Added a CLI with 8 checkers that validate your scaffold against your real codebase, zero tokens used, zero AI, just runs and gives you a score:

It catches things like referenced file paths that don't exist anymore, npm scripts your docs mention that were deleted, dependency version conflicts across files, scaffold files that haven't been updated in 50+ commits. When it finds issues, mex sync builds a targeted prompt and fires Claude Code on just the broken files:

Running check again after sync to see if it fixed the errors, (tho it tells you the score at the end of sync as well)

also a community member here on reddit tested mex combined with openclaw on their homelab, lemme share their findings:

They ran:

  • context routing (architecture, networking, AI stack)
  • pattern detection (e.g. UFW workflows)
  • drift detection via CLI
  • multi-step tasks (Kubernetes → YAML)
  • multi-context queries
  • edge cases + model comparisons

Results:

  • 10/10 tests passed
  • drift score: 100/100 (18 files in sync)
  • ~60% average token reduction per session

Some examples:

  • “How does K8s work?” → 3300 → 1450 tokens (~56%)
  • “Open UFW port” → 3300 → 1050 (~68%)
  • “Explain Docker” → 3300 → 1100 (~67%)
  • multi-context query → 3300 → 1650 (~50%)

The key idea: instead of loading everything into context, the agent navigates to only what’s relevant.

I have also made full docs for anyone interested.

I am constantly trying to make mex even better, and i think it can actually be so much better, if anyone likes the idea and wants to contribute, please do. I am continously checking PRs and dont make them wait.

thank you.

r/ClaudeCode Acehan_

This is getting actually ridiculous

Disclaimer: AHWT. A Human Wrote This. No AI, here.

Alright hear me out. This is not even about usage limits or annoyances anymore (like the fact that the 'gift' from a couple days back never registered after I claimed it), it's about failing to provide the most basic UX.

I genuinely cannot believe what just happened to me. I was working in a session, minding my own business, and I tried to hit rewind with escape-escape. I actually did ctrl-c ctrl-c accidentally because I'm using a pie menu on my mouse for those shortcuts.

No big deal, right? Well, guess what. I come back, hit /resume, and my conversation is GONE. I literally see one blurb of it.

I then hit one of these, thinking this will, you know, actually resume my conversation properly? claude --resume e8f10df9-cb94-4cb3-8f8e-53188f247e47 (with --dangerously-skip-permissions of course)

End of story right? No! I get asked "Resuming this conversation will consume a lot of tokens (bear in mind I was in this conversation less than a minute ago with full caching loaded, working as normal). We recommend resuming it with a summary.

I say no, and resume it entirely. It's still gone.

It's been weeks of this buggy mess around conversation caching, nonsensical "optimizations" that makes it so you can't look at your actual conversation history, truncating your skill descriptions to 250 characters for no good reason at all, and instructing the model to be less helpful directly in the harness in order to save compute at the cost of user experience.

Am I supposed to be excited about the future of AI, when the main product we're all using is so fundamentally broken it can't even render a .json file properly?

Strip it all down, take it back to basics, and remove the bloat.

Limiting the terminal output to one and a half page of conversation history, and now to five lines? Just to brag about slightly reduced startup time? This is really not the flex y'all seem to think it is when it comes to optimization.

I'd love to see some heart and some common-sense put into your product. Rant over.

r/Anthropic antoniodiavolo

Can someone explain to me how Mythos is not AGI/ASI?

I got into an argument with my friend about this earlier but with this Mythos announcement, I think we reached AGI a long time ago and now we're approaching ASI.

This is the start of the permanent underclass and the end of digital privacy/security so buckle up

r/ClaudeAI ppazosp

I built a Claude Code plugin that replaces Linear for solo dev with AI agents

I used to have a semi-autonomous dev flow connecting Linear with Claude Code. For a team it works great but for solo it was overkill. The agent spent more tokens reading issues than writing code. MCP roundtrips, JSON parsing, context burned on overhead.

So I rebuilt the whole thing from scratch with no external dependencies, just markdown files.

Called it Anvil.

You describe your project, it grills you until the spec is solid, then generates issue files organized in phases. Each phase has "heats": parallel work streams that touch different code. So you can launch multiple agents at once, each on its own issue, own worktree, no conflicts.

Commands:

/anvil:forge -> describe your project, it generates phased issue files ready for agents

/anvil:inspect -> see what's ready, what's blocked, how many agents can run at once

/anvil:strike -> agent picks up an issue, own worktree, TDD, review, merge

/anvil:mend -> same but for bugs, failing test first

All state is markdown. All local.

Claude Code plugin:

/plugin marketplace add ppazosp/backpack

/plugin install anvil@backpack

Or as a skill:

npx skills add ppazosp/anvil

github.com/ppazosp/anvil

Happy to get some feedback :)

Full disclosure: this is my project, free and open source.

r/ClaudeAI NomadicNP

HIPPA compliant AI tool generation HELP PLEASE!

I'm way over my head..... but have a vision of building an AI HIPPA compliant triage and documentation tool for my specialty clinic.

I work for a nonprofit center of excellence that is pioneering the field of ME/CFS and Long-Covid and comorbidites. Heidi, Hathr, Freed, etc....these AI tools are insufficient to handle our specialized field and provide singlular use function, requiring a lot of manual input. All are inferior to the latest Claude model. I need guidance!

Here's my vision.... The patient would fill out a comprehensive intake and submit all their medical records. It would have AI would scrape the records to generate a full report of relevant work up that's been done, gaps in internal med workup, if they meet diagnostic criteria for our speciality services. Positive answers to certain questions on the intake or details in their history would flag an adaptive questionnaire (ex: say yes to chest pain and suddenly questions about "with exercise?" "Is it reproducible by pressing on the sternum?"). After obtaining all the data, AI would synthesize it, generate a note and triage the patient as "mild, mod or severe". The patient would first review the output for accuracy, then it would receive MD overview. Mild patients would then receive a one-time face-to-face visit. Moderate patients would receive care for 1-2 years for stabilization, then be sent back to PCP for management. Severe patients would become ours alone. All patients would go through an educational course. Those that are sent back to PCP are given their note, a consultation letter to their provider, a free CME for their provider, resources/quick guide for their provider and the PCP gets a phone number to reach out with questions.

1.3 million people in the US have ME/CFS and long covid. 500 million worldwide. Many are bedbound and homebound, suffering intensely and dismissed by the medical community. Education on these conditions for providers is nearly nonexistent. Very few providers exist. This could change the landscape and scale our efforts, but I need someone who is tech savy to help point the way. If something like this exists already, that'd be amazing. If it doesn't and someone wants to help out a small nonprofit help millions of people, I'd be eternally grateful.

Please and thank you.

r/ClaudeAI Large_Mine

Can we fully launch a website with Claude without experience in coding ?

Hello, I have only ever launched websites using WordPress and have very little coding experience. My main field is supply chain, with some knowledge of Python, Excel, and Power BI. I would like to develop a SaaS. Do you think it is possible to rely fully on Claude Code?

r/ClaudeAI Novel-Upstairs3947

Claude Notch — free open-source app that turns the MacBook notch into a live Claude AI usage dashboard

I built a native macOS menu bar app that uses the dead space around the MacBook notch to display Claude AI usage stats.

Hover over the notch → a dropdown panel appears with:

- Live session & weekly usage with sparkline charts

- Predictive analytics (when you'll hit your limit)

- Pomodoro focus timer (shows in the notch while running)

- CPU & RAM monitor with sparklines

- Rich text notes

- Full settings page

Built with SwiftUI + AppKit. No Dock icon, no menu bar icon — lives entirely in the notch. Ctrl+Opt+C toggles it from anywhere.

Native macOS app, ~700KB, open source, no telemetry.

Download: https://github.com/acenaut/claude-notch/releases

Source: https://github.com/acenaut/claude-notch

Requires a Claude Pro/Max subscription to be useful. Works on non-notch Macs too (uses safe area insets).

r/ClaudeAI FitButterscotch2250

How I built a full bilingual SaaS in 27 days using Claude Code — zero coding background (312 commits, 181 deployments)

I'm Mahmoud, I've been working in SEO since 2018.

A little over a year ago I got into freelancing platforms, started offering SEO services on Upwork. The work was good, but dealing with clients directly and constantly drained me. I kept thinking: why don't I turn my expertise into a SaaS product? The only problem? I'm not a developer my background was WordPress and basic tech stuff only.

The moment that changed everything

Early 2025, I noticed a pattern: my clients started asking me about how their brands appear in ChatGPT and Gemini, not just Google.

I looked for tools to track this — found some but they're expensive (300$+/month), and the biggest surprise? Not a single one supports Arabic.

That's when I realized how massive the opportunity is: 440 million Arabic speakers, Arabic content is less than 1% of all internet content, ecommerce in the Gulf is exploding — and there's literally zero tools serving this market.

A full year of frustration on v0

I started trying to build using v0 by Vercel. Spent a full year trying, but the errors were endless and I didn't have the coding skills to fix them. Hired people to help — sometimes solving what I thought was a simple problem took them days.

27 days that changed everything

About a month ago, I started using Claude Code. Honestly, it felt like I hired an entire dev team. Creative ideas I couldn't execute for a whole year turned into working code in hours.

I worked 15+ hours a day for 27 straight days. Completely alone. No team, no developer, no investor. I even stopped going to the gym — which is sacred to me — because the momentum was stronger than the physical exhaustion. Sometimes I literally felt like I was going to pass out from how tired I was but I couldn't stop.

What exactly did I build?

A full SaaS app:

  • Brand visibility tracking across 5 AI models with full Arabic and English support
  • AI-powered SEO advisor (auto analysis + chat)
  • Full integration with Google Search Console and GA4
  • Daily keyword rank tracking
  • Arabic keyword clustering using AI
  • Technical site audit — 25+ checks
  • Full website analyzer
  • PDF reports + CSV exports
  • Subscription system with 3 tiers
  • Every single page, every button, every error message — in both Arabic and English

How I used Claude as a full team

Claude Code — for daily building. I give it a detailed prompt with full context: what currently exists, what it should NOT touch, and what to build. And it executes. The key is being extremely specific about what should NOT change.

Claude Cowork — honestly my experience with Cowork wasn't great at all, I think because it's still in beta. I didn't rely on it much.

Claude (regular chat) — for strategic planning, market analysis, and content creation.

Biggest lesson: Claude is not a replacement for a developer — it's a replacement for an entire team, BUT only if you know exactly what you want. The vision and domain expertise has to come from you. Claude executes it.

What I learned in 27 days

I connected over 10 different APIs — from AI platforms to website analysis tools to Google Search Console — all learned from scratch through Q&A with Claude.

On top of that I learned and used: Next.js, cloud databases, payment and subscription systems, email automation, LinkedIn outreach automation, building prospect lists, setting up Google Cloud and OAuth, and literally yesterday I learned a new automation tool just through Q&A with Claude.

312 contributions on GitHub. 181 deployments. All in 27 days.

The real challenges

Burnout is real. 27 days non-stop, 15+ hours daily. Physically it was brutal.

Constant doubt. "Will anyone actually use this?" That question kept coming back every few days.

My biggest regret — every wasted day in the past where I didn't use these tools.

Where am I now?

The product is live and working. Started distribution — outreach campaigns, Arabic content, AI tool directory submissions.

But the honest truth? Zero paying customers so far. And that's the real challenge ahead.

Since many of you have been through this stage — what's the best strategy you used to get your first 10 customers for a SaaS product? Any advice for someone who's strong at building but new to sales?

r/StableDiffusion osiris316

Environment Lora

Hey everyone.

I’ve had decent success training character Lora’s with Ostris. So I would like to see if I can train an environment. Like a house.

Has anyone had any success training a home or environment Lora? Any tips or tricks or things to look for and look out for? This will more than likely be a ZIT or LTX 2.3 lora. Thanks!

r/LocalLLM Nervous_Trainer_2630

48Gb RAM + Qwen code 3.5? Any experiences?

Image related, I really feel like going local.

I'm thinking A6000 + Qwen code? Anyone doing their vibecodes with that card?

r/ChatGPT ogenom

It do be like that

As this is from 6 years before the release of ChatGPT, I’d say it’s close to prophecy.

r/SideProject Dizzy-Run-7190

Trouble processing my own emotions, so I built something to help — mood tracking, precise words for feelings, and music to match

So I live and work in Dubai. If you follow the news, you get what's been going on in the region.

As someone who's never experienced anything remotely close to war — it's a wild ride. Missiles overhead like shooting stars. Attack alerts going off at 3am. Boss expecting you in the office, which happens to be in a port area high on Iran's target list. You learn to carry on as usual, but in this weird fog — scared? numb? couldn't care less? missing home? all of the above?

As a lifelong non-believer in meditation and journalling, and too broke for a therapist, I decided to take matters into my own hands: what if there's a precise word that can pin down exactly what I'm feeling? What if I could track how my mood shifts over time? What if there's a little panic room — a sanctuary — that smooths me out when I'm melting down, with some wisdom from people who've been in my shoes before?

Introducing the Mood Sanctuary:

What it does:

  • Type how you feel → get the most precise word for it from any human language
  • Matching photo and music for every mood
  • Weekly Sunday reflection delivered to your email
  • Mood journals download to any calendar app

Fun fact — I was freaking out at 3am again right before Trump's deadline and his promise to wipe out an entire civilisation. So I used my own creation for my salvation (or self-destruction, to be more precise). Here are the songs that sent me to sleep with: Beds Are Burning, Eve of Destruction, The End of the World, Run Boy Run.

So enjoy, my beautiful people!

r/LocalLLaMA Sea-Emu2600

State of NVFP4 on mlx

So I’m testing several models on macOS and I’d like to understand if NVFP4 is the best option to run 4bit models quantized models using mlx. From my investigation although it’s a software emulator since MacBook does not implement this on hardware, looks like the current mlx implementation is on pair supporting the dual scaling factors (micro block and tensor level). So should I expect less loss compared to a 16fp model? Is my mental model right?

r/ClaudeAI BeaconBuilder

Has anyone seen 100% cache rate spikes on a disabled API key? Trying to understand what's happening technically.

Something unusual showed up in my usage dashboard this morning and I can't explain it technically. Hoping someone here can.

At 3:10 AM UTC today my usage dashboard recorded:

- 864,077 tokens IN

- 1,792 tokens OUT

- Duration: ~5 minutes

- My API key was manually disabled at the time

- No active session, no cron jobs, no code running

The part I really can't explain is the caching chart. At exactly 3:10 AM the cache rate hits 100% simultaneously with the token spike, then both collapse back to zero at 3:15 AM.

I'm familiar with bugs #44703 and #41930 which have caused token inflation in active sessions. But this happened with the key disabled and nothing running client-side.

My questions for anyone who understands the platform internals:

  1. Can Anthropic's caching infrastructure make calls against a key independently of user-initiated requests?

  2. What would cause a cache rate to hit exactly 100% in a single burst?

  3. Has anyone else seen token activity on a disabled key?

Screenshots in comments. Genuinely trying to understand the mechanism before I escalate further.

r/SideProject Inner-Spend-646

I built an app to stop getting blindsided by subscription charges

Last year I realized I was paying for 3 services I hadn't opened in months. Hulu with no ads (switched to Netflix), a meditation app from a new years resolution that lasted 2 weeks, and some cloud storage thing I signed up for a free trial and forgot.

That's like $40/month just gone. And I only caught it because my card declined on something I actually wanted.

So I started building TrackAutopay. The idea is simple — one place to see every recurring charge, when it renews, and get a heads up before it hits.

What it does:

  • Add your subscriptions manually or connect your bank (via Plaid) to auto-detect them
  • Calendar view so you can see which days are "payment days"
  • Reminders before renewals (customizable — 1 day, 3 days, a week, whatever)
  • Spending breakdown by category (entertainment, software, utilities, etc.)
  • Spending forecast that projects your yearly cost
  • &many more

It's on iOS right now. Would love honest feedback — what's missing? What would make you actually switch from a spreadsheet to this?
TrackAutopay

r/SideProject Altruistic_Driver381

I built a Rust CLI to turn git diffs into clean, shareable images (diffshot)

I built diffshot: a Rust CLI that turns git diffs into clean, syntax-highlighted images.

Use cases:

  • sharing diffs in PRs, tweets, blog posts
  • cleaner than screenshots
  • runs directly from the terminal

Example:

diffshot HEAD~1 # render changes from your last commit 

Install:

cargo install diffshot 

The output is PNG by default. PRs are always welcome!

Repo: https://github.com/faisalfakih/diffshot

r/ClaudeAI Bitter_Palpitation76

Non Tech Person Help

Hi folks

I’m part of a customer support team in a cyber sec company. I have access to Claude's code and a Claude coworker; I was wondering if anyone within a similar field has used Claude for anything for their own productivity or built something for the team?

r/ClaudeCode lmasieri

How do you structure your Claude Code projects — one root folder or per-project?

I’ve been using Claude Code for a few months and I’m struggling to find the “right” way to organize my workspace, especially around CLAUDE.md files, context window efficiency, and memory persistence. I’d love to hear how others are doing this.

My current setup:

I have a single parent folder called ClaudeCode/ that I sometimes launch Claude Code from. Inside it, I have a mix of unrelated projects organized by life domain:

ClaudeCode/ ← no git repo, no CLAUDE.md ├── .claude/ │ └── settings.local.json ← root-level permissions ├── _tools/ ← MCP servers │ ├── fastmail-mcp/ │ └── custom-mcp-tool/ ├── personal/ │ ├── CLAUDE.md ← has my bio, career context, │ ├── website/ ← git repo │ ├── career/ │ │ └── CLAUDE.md ← career coaching context │ └── content/ ├── side-business/ │ ├── brand-a/ │ │ └── website/ │ └── consulting/ └── projects/ ├── app-project-a/ ← git repo, has .claude/ settings │ └── CLAUDE.md ← full tech stack + project docs ├── app-project-b/ ← git repo ├── hardware-project/ ← git repo ├── business-tool/ │ └── CLAUDE.md ← business context + folder map └── side-project-c/ 

My questions:

  1. Root folder vs. per-project launch — and what happens if you alternate? I sometimes launch from the root ClaudeCode/ folder so I have access to everything, and sometimes from a specific project folder for focus. But I’ve realized this means my memory is split — root sessions and project sessions have separate memory namespaces and don’t see each other’s learned context. The CLAUDE.md loading also differs: from root, nested CLAUDE.md files aren’t loaded until Claude navigates into those directories; from a project folder, it walks up the tree and loads parent CLAUDE.md files automatically. Are you all picking one and sticking to it? Or is there a clean way to handle the split?
  2. CLAUDE.md layering — I have CLAUDE.md files at different levels (e.g., personal/CLAUDE.md with my bio, projects/app-a/CLAUDE.md with tech stack). My understanding is Claude Code walks up the directory tree and loads all CLAUDE.md files from root to the current working directory. But since I sometimes launch from root, does it even see the nested ones? Or does it only load them when I reference files in those directories?
  3. Memory system — Claude Code creates a memory directory tied to the working directory path. If I always launch from root, all my projects share one memory namespace. If I launch per-project, each gets its own. Which is better? Is there a way to have shared memory (who I am) plus project-specific memory?

What I’m trying to optimize for:

  • Long context window sessions without wasting tokens on irrelevant project context
  • Persistent memory that knows who I am across all projects but also knows project-specific details
  • Clean separation between unrelated projects while still being able to cross-reference when needed
  • Not having to re-explain myself every time I start a new session

How are you all structuring this? Do you do one mega-folder? Separate project launches? Something else entirely? Especially curious if anyone has a multi-project setup with both code and non-code work.

r/ClaudeCode conceptualchecks

Built an virtual office for my Claude Code agents

managing parallel claude code runs was getting too difiicult:

- which termina window? which project? what's the action needed?

stilling finding this an issue even though I'm yet to maxxed out my Max (20x) plan

ended up building the tool over the past week to help me manage all the parallel project execution

visualising all claude code runs as agents in an virtual office on a dedicated iPad:

> major projects get their own private office
> ad-hoc tasks go in the open plan section
> agents pop an alert bubble when blocked on my input

still a lot of work to be done but mapping agents to a physical space has solved 80% of my cognitive load issue of parallel runs.

r/LocalLLaMA garg-aayush

ATOM Report highlights the sheer dominance of Chinese labs in the Open-Source LLM space

Nathan Lambert and Florian Brand has published a comprehensive analysis of open model adoption from Nov 2023 to Mar 2026 tracking around 1.5K models across Hugging Face downloads, OpenRouter data and other benchmarks.

One of the biggest takeaways for me is the sheer dominance and scale of contributions from Chinese labs (especially Qwen) to the open-source ecosystem.

To be honest, their initiative in open-sourcing models like Qwen and DeepSeek has also encouraged similar efforts from other labs across Europe and the US.

I would even attribute the recent release and fast tracking of Gemma4 to the success of Qwen3.5.

I would recommend everyone to go through the report (even just the graphs) just to see the scale of Chinese models influence and adoption in Open-Source community

Report link: https://atomproject.ai/atom_report.pdf

r/ClaudeCode wewerecreaturres

CC freezing temporarily in Bash/Pwsh/Cmd

This didnt used to happen to me, but lately ive noticed that CC tends to freeze until i press tab/space/type a character. I've tried using bash, powershell, and cmd and they all do the same thing. I've tried using them independently and within VSC and Zed, same problem.

Anyone know what causes this or have a solution?

r/ChatGPT mosen66

How to make ChatGPT follow procedures. Help wanted please

It always try’s to “optimize” and screws up my procedure text and outputs, even when I ask for strict adherence to the process rules/params/etc. How can I enforce this?

r/SideProject AffectionateNote2357

Making my product free, is it suicide?

Recently pivoted my project as the hard paywall was limiting my feedback loop (low conversion rate).

Long story short, is it suicide? I use AI for every potential customer and the more traffic I have the greater my costs - has anyone done this kind of pivot successfully? So, start off for free, get some traction and then pivot into a sustainable model?

Would love to hear your thoughts :)

THanks!

r/AI_Agents cutups

What is the best way to give AI access to my To Do / Task list and have it actually help me?

I'm taking another look at my to-do / task list to see if I can change or improve it so that I can have AI Agents help me out.

I currently use Microsoft To Do because I like it's simplicity and ability to use it on desktop and mobile. However since I'm using it with my personal email, I haven't found a good way to make it accessable to LLMs.

I use my to do list for just about anything, from grocery lists, home projects, random ideas for music, ideation of coding projects. I mostly keep it separate from my 9-5 job work, but if I come up with a better system I might use another instance for that work as well.

I would like to keep the simplicity of Microsoft To do, but have the agent keep me on task, refine issues, enrich, combine or amend items into new logical lists, complete items when possible.

If I can expose my existing to do list to LLMs, that would be great, but I'm open to exporting my data or starting with a new system.

Any personal experiences or suggestions are appreciated.

r/ClaudeCode SnooTangerines2270

image-generator-vibe-coding for Claude and other

https://github.com/TruCopilot/image-generator-vibe-coding

Why This Exists — The Missing Piece of Vibe Coding

Every AI coding platform today — Claude Code, Cursor, Windsurf — can scaffold an entire website in minutes. Beautiful layouts, perfect typography, responsive grids, dark mode toggles. But open the result and what do you see? Grey placeholder boxes. Broken image icons. Empty hero sections. The design falls flat because the visuals are missing.

This is the problem we solve.

This MCP server is the bridge between AI-generated code and AI-generated visuals. When your vibe coding tool builds a landing page, it can now simultaneously generate the hero image, the feature illustrations, the team headshots, the background textures — all ultra-realistic, 2K resolution, perfectly sized for the layout it just created.

With this MCP, when you code, it will auto generate any images that missed or required for UI, including Logo, Icons, or background , Photos.

Hope you guys find this useful for fronend design.

r/ChatGPT 69dildobaggins69

No longer integrating with Google Docs?

Until last week, I could attach Google Docs directly to a chat.

Then it made me start pasting links instead of attaching since that option was no longer there.

Then today Chat says it can’t read links that are just pasted in like that. Ughhhh.

I have a Plus account and double checked that my Google Drive is connected to my Chat.

Help??

r/ChatGPT Logical_Wafer6195

Stop trusting AI blindly.I built AI Truth make AI more honest — and show its receipts.

Disclosure: I built this. I’ve been working on AI Truth, an open-source prompt framework and Chrome extension for making LLM outputs easier to audit. The core problem I’m targeting is that very different kinds of claims — live search results, recalled model memory, and open-ended inference — are usually delivered in the same confident tone, which makes them hard to evaluate at a glance. My approach is not to generate probability scores, since those often look more precise than they really are. Instead, I use a small evidence taxonomy to label where a claim is coming from, then visualize those labels directly in the chat UI. The main lesson so far is that this works much better on frontier models than weaker ones, and that visible uncertainty improves auditability but does not guarantee correctness. Current limitations are model-dependent labeling quality and imperfect handling of source independence. Repo/docs: https://github.com/Kunkun2116/AI-Truth/

r/ClaudeAI Dull_Alps_8522

I built an MCP server that turns Claude into your social media manager (Instagram + TikTok)

Hey everyone,

One thing that kept bugging me lately: we can vibe-code an entire app in an afternoon, but the moment it ships, marketing and distribution become the real bottleneck. So I built something to fix that part of my own workflow and figured I'd share.

It's called FluxSocial, and the interesting piece (at least for this sub) is the MCP server I added on top of it. Once you connect it to Claude, you can manage your social accounts in plain conversation:

That's the whole interaction. Claude chains the steps behind the scenes — learning from your previous posts for tone, generating visuals (images or AI video via Google Veo 3), and scheduling everything to Instagram (posts, carousels, reels, stories) or TikTok. Multi-account support too, so you can keep the yoga studio and the pizzeria completely separate.

The part I'm most happy with is the workflow chaining. You don't run three separate tools — Claude proposes a full draft (copy + visual + schedule), you review, you approve. That's it.

A few things worth mentioning:

  • Not Claude-exclusive. The MCP URL works with any MCP-compatible client (Claude Desktop, Cursor, etc.) as a connector.
  • REST API available if you want to bake these capabilities into your own app instead.
  • You do need to connect your Instagram account once to grant posting/analytics permissions — standard OAuth flow.

It's still rough around the edges, which is exactly why I'm posting here. I'd genuinely love feedback from people who actually use MCP servers day-to-day — what's missing, what's broken, what would make this actually useful for your workflow.

Links:

Happy to answer any questions about the implementation, the MCP design choices, or anything else.

r/ClaudeAI One-Accident-6101

Is Claude Pro worth it for learning?

I'm learning UI/UX and I have a billion questions and doubts when learning figma. Sure I can look up tutorials on Youtube but it's easier to learn from Claude & it's pretty good at teaching stuff. Is $25.60 worth it? It is a bit expensive for me (2,177.89 Indian Rupees) but I'm willing to spend. I just want to know if the limit usage is atleast 4x/5x more than the free version?

r/ClaudeAI SuperShittyShot

Help figuring out Claude (VSC Plugin)

Context: I'm using the 20 bucks tier from Anthropic, Google and OpenAI so I get the job done (when it works lol) and it allows me to compare how different providers behave and I can ensure it's not looking great for Anthropic lately, I feel like the performance has gotten worse and I'm facing "bugs?" more often than not.

I tried the claude code but I prefer the experience of having an IDE so I am using the official VSC plugin. I have a .claude directory with agents, skills, commands, evals... and a CLAUDE.md file at the root of the project, pointing to the AGENTS.md (I've observed it ignores the AGENTS.md standard otherwise). In fact, all the AI ruleset and whatnot is based on Claude and funny enough Claude is the one that's following them the least

Lots of times it blatantly ignores the existence of these files unless I shove them in the context by hand which is annoying on its own, and definitely not intended as, according to the doc ( https://code.claude.com/docs/en/memory ) it loads these on every new session.

I assume it's an issue with the plugin but what do I know. Besides, more than a bug report I am seeking group support or something like that I guess 😅

Long story short Claude ignoring rules and context is causing me trouble, which adds up to the fact that we have less and less usage.

The most recent example, I asked it to investigate a bug. After wasting 48% of my current usage in a single analysis run, it told me the solution was to rename my proxy.ts to middleware.ts... in a Nextjs 16.2.2 project... and explicitly having the tech stack with versions first thing defined in the AGENTS.md file which remember, is explicitly attached in the CLAUDE.md file, following claude documentation.

Of course when I pointed out the middleware is now called proxy since months ago it told me "You're right, I apologize for the wrong claim. Let me look at the actual problem fresh."
But of course, half of my current usage is already gone, never to be seen again.

In other circumstances I can even accept the "bro prompt it right" mantra, but seriously I am following all the recommendations and I still face these situations, I call it FOP (Frustration Oriented Programming) lol

I am wondering what could I, as a user, have done to get it to act as expected? and more important, should I have to pay for errors that are not mine? The same way malformed responses are not counted in the usage (AFAIK) these blatant mistakes on the provider side should also be the responsibility of the provider IMHO.

Due to that I had to waste yet more usage to fix the bug, reaching near 80% usage so, to finish the small feature it has half-done in the following chat, now I need to wait three hours which is crazy to say the least. And that's assuming it will do things right this time.

Any similar experiences? Any ideas on how to get it to work as expected? TIA

https://preview.redd.it/0it0xbg4vztg1.png?width=1766&format=png&auto=webp&s=ae14db60e06ce7f6fe37517600000c2549032f06

r/singularity Anen-o-me

From Inside the Meat - short film

r/LocalLLaMA onil_gova

I tracked a major cache reuse issue down to Qwen 3.5’s chat template

Over the last week, I’ve been investigating cache misses while optimizing local agent workflows on my M5 Max.

My setup used oMLX.ai as a backend with agents like OpenCode.ai and Pi.dev, but I reproduced the same behavior with other backends like llama.cpp too. At first, I assumed this was an inference engine issue or a cache implementation bug.

What I kept seeing was frustrating:

  • the model would read a large amount of context
  • it would make a chain of tool or function calls
  • I’d ask a simple follow-up question
  • and instead of reusing the prompt prefix, a large chunk of the conversation would get reprocessed from much earlier in the history

In practice, a follow-up turn after a tool-heavy interaction could end up redoing tens of thousands of tokens for no good reason.

I first found a separate issue related to multimodal / first-image transitions, and I already have an oMLX PR for that.

But the bigger text-only issue turned out to be the Qwen3.5 chat template.

After tracing prompt fingerprints and comparing rendered prompts across requests, I found that the template was emitting empty historical `...` blocks for prior assistant turns even when there was no reasoning content. That caused equivalent conversation history to serialize differently across requests, especially after tool use.

The template itself was introducing unnecessary prompt drift.

That matters because prompt drift hurts prefix-cache reuse, which means extra token processing, more latency, and wasted compute.

The fix is really simple one-line change in the template:

from:

{%- if loop.index0 > ns.last_query_index %}

to:

{%- if loop.index0 > ns.last_query_index and reasoning_content %}

If you’re serving Qwen3.5 locally and relying on prefix caching, this may be quietly costing you performance. If you’ve noticed long follow-up turns getting unexpectedly reprocessed after tool use, this may be the reason.

I reproduced this across different agents and backends. The common factor was the shipped template.

If you’re debugging cache misses on Qwen3.5, check the chat template before adding more cache-layer workarounds.

I’ve opened PRs on the official Qwen3.5 model repos. For example:

https://huggingface.co/Qwen/Qwen3.5-122B-A10B/discussions/22

If you’ve seen similar behavior, help spread the word so this gets patched upstream.

TL;DR:I traced a major cache reuse problem in Qwen 3.5 back to the shipped chat template, not the inference engine. The template emits empty historical `...` blocks even when there is no reasoning content, which creates prompt drift, hurts prefix-cache reuse, and causes unnecessary reprocessing of large contexts after tool use. The fix is a one-line template change, and I’ve opened PRs on the official Qwen 3.5 model repos.

r/SideProject yesinteractive

built a disposable inbox that deletes itself so you don’t have to deal with your bad decisions later

made a simple burner inbox because I got tired of:

  • signing up for stuff
  • getting spam forever
  • immediately regretting it

it:

  • generates an inbox instantly
  • receive-only (no sending, no accounts)
  • auto-deletes after ~1 hour
  • you can create custom public inboxes (anyone with the name can view it)

basically:
use it → get your code → let it die

curious if this is actually useful or just me avoiding responsibility

r/ClaudeCode Mildewd_

Claude Code Undulating IQ

It's obvious that at certain times during the day the model loses inference or becomes basically useless. Just make an indicator so others and I can decide NOT to use the model in active work environment where it poops on the entire code base.

r/SideProject AlopeciaToRegrowth

Built HappiHeal, a simple alopecia recovery tracker inspired by my own hair loss journey

Hi,

I’ve been dealing with alopecia areata for a while, and one thing I kept struggling with was consistency. I was trying different routines around food, stress, supplements, and general self-care, but everything was scattered across notes, screenshots, and my head.

So I built HappiHeal as a small side project to make that process simpler.

The idea is pretty straightforward: a lightweight tool for people going through hair loss or autoimmune-related recovery journeys to stay more aware of what they’re doing consistently and what may or may not be helping over time.

I built it from a very personal place, not as a perfect medical solution, but as something I personally wished existed when I started.

A few things I’ve been thinking about while building it:

  • How simple should a health-tracking app be before it becomes too limited?
  • What kind of inputs are actually useful for alopecia tracking?
  • Would people want more structure, more insights, or just an easy daily log?

It’s live here: happiheal.happiloot.com

I’d genuinely love feedback on both the idea and the product:

  • Does this feel useful for people dealing with alopecia or similar autoimmune conditions?
  • What features would make it more helpful without making it complicated?
  • If you’ve built in health/wellness, how did you handle trust and privacy?

Thanks — sharing mainly to learn and improve it.

r/ChatGPT Kitchen_Week1117

Is this a Good Thing?

r/ClaudeCode liaddial

If you've only been using Claude Code subagents for parallel work, try them for parallel perspectives too. Here's the loop.

I mostly used to use Claude Code subagents for parallel work -- search this, refactor that, run things concurrently. Lately I've been using them for something different: parallel perspectives.

Same problem, three different lenses, at the same time. The kind of disagreement you can't get by continuing to argue with the main agent alone.

The context

I was designing the command layout for a small markdown-only harness I've been building for Claude Code -- just a side project of mine. Mid-discussion with the main agent, I hit a wall -- the kind where you can tell you're going to keep bouncing off the same reasoning no matter how many more turns you take.

Smarter models come with a subtle risk: they get more persuasive inside their own point of view. And when that doesn't happen, the opposite one does -- after enough back-and-forth, the model starts yielding to whatever framing you've been pushing, and the "fresh perspective" you were hoping for turns into polite agreement.

Either way, the main agent alone stops being able to break out of the tunnel. I needed a way to force a different angle without restarting the session.

The loop

Here's what I asked the main agent to do. The first two steps are verbatim prompts I handed over; the rest is what I did afterwards with the result.

1. Prompt: "Log the current discussion state to docs/discussion/2026-04-08-.md -- open question, options, your current lean."

2. Prompt: "Spawn 3 subagents with different perspectives -- LLM, software architect, end-user. Write their takes to docs/discussion/2026-04-08--round-1.md."

Then:

3. I come back to the main agent with that `round-1.md` file and continue the discussion using the three new perspectives as fresh inputs.

4. If we reach consensus, I ask the main agent to update the original topic doc and close it out.

5. If open threads remain, I run the same cycle again on the narrowed questions -- round-2.md, round-3.md -- repeating until the discussion actually closes, not until I get bored of it.

Why the three personas matter

The trick isn't "more agents" -- it's non-overlapping lenses. The three I picked give me three genuinely different sentences about the same design choice:

  • LLM-as-executor perspective (the model that will actually run this command): "As the model that has to follow this on first read, is the instruction clear? Does any phrasing force me to guess what was meant? Would I silently interpret this the wrong way?"
  • Software architect perspective: "Will this still make sense to maintain in a year? Is the abstraction pulling its weight?"
  • End-user perspective: "Can someone understand this in five minutes without reading the source?"

Those three sentences rarely agree. And the disagreement is where the actual design work happens. The main agent alone almost never produces the end-user sentence -- it's too close to the implementation details. The architect subagent almost always pushes back on abstractions the main agent was happy with a minute ago.

It's not that the subagents are smarter. They're just looking from somewhere else.

The unexpected bonus: a paper trail

The side effect I didn't plan for: the round-1.md, round-2.md files accumulate into a genuine decision log. Three days later, when I inevitably ask "wait, why did I settle on this?", the answer is right there -- with the dissenting opinions still on the page.

Most of my past design discussions evaporated the moment the session ended. This pattern turns them into durable artifacts without me having to discipline myself to write post-mortems afterward. The discipline is built into the loop. The discipline is built into the loop.

r/ClaudeCode xephadoodle

All 1m token models now not part of Max?

I had 1m token access for a week or two, now it seems ALL of them have been removed and are only available through "extra usage"?

Anyone else run into this?

r/artificial kc_hoong

"OpenAI quietly removed the one safety mechanism that could shut the whole thing down — and nobody is talking about it"

OpenAI was founded as a nonprofit for one specific reason — to ensure AI development couldn't be hijacked by profit motives.

Their original charter had a clause that legally required safety to come before profits, and gave the board the power to shut everything down if AI became too dangerous.

That clause is gone. The board has been restructured to answer to investors instead.

We just removed the emergency brake from the most powerful technology in human history because it was bad for business.

What happens the next time something goes wrong?

r/ClaudeAI matrix20085

Anyone getting a 500 error from Code?

API Error: 500 {"type":"error","error":{"type":"api_error","message":"Internal server error"},"request_id":"req_####################"}

Getting the above error, but only from Code sessions. Tried multiple sessions over different projects and different models. I am using Desktop.

r/SideProject Appropriate-Fix-6283

How I’m Using AI

I’ve been experimenting with AI + automation for the past few weeks, and I think I’ve found a simple system that can realistically scale to $5K/month.

Here’s the exact breakdown 👇

r/SideProject Dense_Job2569

I Build a collaborative application to find more easily flippers around the world !!

Hey everyone

I’m currently working on a mobile app idea and I’d really love your honest feedback.

The concept is simple: it’s a collaborative app for pinball lovers

The problem I noticed (and my aunt actually pointed this out) is that it’s really hard to find pinball machines in bars, restaurants, or arcades. And when you do find one, there’s no easy way to share that info with others.

So I’m building an app where users can:

  • Find pinball machines near them on a map
  • Add new locations (bar, arcade, etc.)
  • Upload photos of the machines
  • Rate and review places
  • List which pinball machines are available

Basically, a “Google Maps for pinball”.

The goal is to create a community-driven platform where enthusiasts help each other discover new spots.

I’m still early in development, but I’ve attached some screenshots of the current version

I’d love to know:

  • Would you use something like this?
  • What features would you expect?
  • Is there anything missing or unnecessary?

All feedback (positive or negative) is super helpful

r/ClaudeCode WinProfessional4958

How is this possible?

r/ChatGPT Jaded_Foundation_601

I win

r/LocalLLaMA No_Run8812

My first 7 second LTX video on M3 ultra, how can I generate longer videos?

https://reddit.com/link/1sfy8y4/video/j3w615ervztg1/player

Total generation time 11 mins, 180 words prompts

Below are the configuration I used, can someone suggest how I can generate longer videos. TIA!

--distilled-lora models/ltx-2.3-22b-distilled-lora-384.safetensors 0.9

--spatial-upsampler-path models/ltx-2.3-spatial-upscaler-x2-1.0.safetensors

--seed 10

--height 576

--width 1024

--num-frames 161

--frame-rate 24.0

--num-inference-steps 40

--video-cfg-guidance-scale 3.5

--video-stg-guidance-scale 0.0

--video-rescale-scale 0.5

--a2v-guidance-scale 1.0

--video-skip-step 0

--audio-cfg-guidance-scale 7.0

--audio-stg-guidance-scale 0.0

--audio-rescale-scale 1.0

--v2a-guidance-scale 1.0

--audio-skip-step 0

```

r/ChatGPT socialcalliper

I was so frustrated with Claude's usage limits that I wrote a song about it here's what it taught me about human-AI attachment

Something weird happened to me. I was mid-thought, mid-sentence, genuinely dependent on a conversation with Claude and the weekly limit hit. Full stop.

Instead of just being annoyed, I sat with that feeling. Why does this feel like being ghosted? That question sent me down a rabbit hole about how emotionally we relate to AI tools, even when we know they're just software.

So I wrote a song about it. Lines like "Let me be your token count, and I'll never run out" started as a joke but ended up feeling genuinely reflective of how we project need and attachment onto these systems.

If you work with AI daily, I think you'll find it uncomfortably relatable. Curious if others have felt this weird dependency too or is it just me?

Spotify Link :- Token Limit give it 40 seconds, the chorus will get you.

r/SideProject QuietWarrior12

Faceless Content Question

For those of you making faceless storytelling videos (true crime, drama, horror), are you also spending hours hunting for the right stock footage on Pexels/Pixabay. Curious if that's common or if I'm doing it wrong.

r/AI_Agents achint_s

I stopped using ChatGPT like a chatbot and turned it into a Chief of Staff.

Free PDF on github to try out.

Most people (including me until recently) use AI like this:

prompt → wait → copy → repeat

It works… but it’s still manual.

I started experimenting with a different approach — instead of asking AI to generate outputs, I made it analyze how I actually work.

Built a simple diagnostic that:

asks 5 questions about your workflow

identifies where you’re wasting time

highlights missed opportunities / slow follow-ups

shows what breaks if your workload increases

Then generates a Vulnerability Report.

The unexpected part:

Once you run it, that same chat basically becomes a persistent Chief of Staff you can keep using —

to organize tasks, clean up messy thoughts, and plan work with context.

No coding, no setup. Just copy-paste and run.

Put it on GitHub here: _Link in comments.

(Side note: I’ve been building a more complete local system around this idea — where these workflows run automatically instead of manually. Still early, but interesting direction if you think beyond chat interfaces.)

r/ClaudeAI dieselcruiserhead

WordPress? I think I screwed up on my core website-building approach, with Claude.

I'm in my first big web project with Claude and part of it I love - it's my best work yet and I have changed the direction and sales of several companies and this will be no exception. But it taking months and months. I have done many sites but am mostly a marketer, not a coder.

The issue is I'm mostly a WordPress builder. My sites are also long form with a ton of content as my secret sauce. I do all the writing and content work and implement it. One stop marketing and sales shop. Claude is crushing it with most of it. I expect the revenue of this company to double in 1-2 years when I am done which I have done several times before. So we really are using websites and internet to its best use case. But again, it's taking months.

I have to review most of the code, and am having to closely review all of Claude's marketing content mostly, some CSS, and mostly issues with it constantly putting the wrong photos in the wrong place, and systematically building it's own dependence by structure into the website building, and lots of misreads of difficult nuanced engineering items in a wild environment. Photo for example.

The website is a technical subject including buoyancy and performance of things like whitewater rafts, so even with deep education I can't get claude to sort and place photos correctly and to fully understand it. Lots of errors. Understandable. It is also MUCH better than ChatGPT which I had to fire after wasting a month or two before realizing all of its output was actually garbage. I then tested ChatGPT about 20 other times and caught it repeatedly lying and more. I and actually view Chat GPT a truly dangerous tool in the amount of misinformation and hallucinations it willing to have.

I can't build or maintain this site without Claude and what appears now to be a very manual process. I am dependent on it or a competitor forever and I'm certain something will catch up to Claude. No big deal, I will never not build without AI again. I believe it is the best took out there for this type of work but have only tested Claude and Chat.

I am having to hack by pasting custom HTML to retain a bunch of core and all this and the architecture recommended by claude to retain the core WordPress functionality. I designed all this with Claude so some of it is my mistake but I am and need to be nearly 90-100% hands on mostly with content edits (mistakes).

I am aware of Netlify that it can have Claude do much more work. What else can I do - what other approaches can I take if I need content rich, light ecommerce or ecommerce functionality, highlighting of products and more?

r/SideProject Ok-Reference281

I built an iOS app that tracks your job applications automatically with AI - just launched today

What it is: Prowl is an iOS app that auto-detects your job applications by scanning your Gmail or Outlook inbox — no manual logging. It uses AI to classify these emails into different application related categories like applied, interview, offer etc.

Why I built it: It's not uncommon to apply to 100+ jobs before getting an offer. I was applying to jobs and using a spreadsheet to track everything. It's exhausting. I realized every application I'd ever sent had a confirmation email sitting in my inbox. So I built an app that reads those emails and turns them into a clean tracker.

What's free vs paid: Free to download, first sync included so you can see it work. Pro ($3.99/mo) unlocks unlimited ongoing syncs, stats & more.

Stack: SwiftUI / SwiftData / SwiftUI, Gmail API + Microsoft Graph, GPT-4o mini for classification, Cloudflare Worker

Would love feedback from devs or anyone who's been through a job search!

App Store Link: Prowl

r/ClaudeCode majornerd

Anyone else experiencing a API Error 500? Seems like they are down again, at least partially

API Error: 500 {"type":"error","error":{"type":"api_error","message":"Internal server error"},

11:30am MST/10:30am PST

Again claude is having issues. Is this just a daily "touch grass" reminder?

r/StableDiffusion Braveheart1980

FaceFusion 3.5.4 - Impossible to remove content filter

I have tried everything described here in posts and even Antigravity hit a wall as it cannot bypass the content filtering! Any help would be more than appreciated!!!

r/SideProject More_Towel3916

Built an Angular 21 SaaS boilerplate over the weekend ..... here's what I learned....

Been a senior dev for 10 years. Every time I start a new SaaS project, I spend the first 2 weeks setting up the same stuff: auth, billing, roles, audit logs. Decided to finally build it once and do it properly. Spent the last weekend building a full Angular 21 starter kit.

A few things I ran into that might be useful for others: Angular 21 signals are genuinely great for this, no more BehaviorSubjects everywhere, component state is so much cleaner.

The new if and for control flow also makes templates way more readable. NgRx 21 with createFeature is underrated. The old way of writing actions/reducers was so much boilerplate. This version is actually enjoyable to use.

Supabase RLS took me the longest to get right. Specifically, the audit log table I wanted it to be INSERT-only so no client can ever delete or modify audit records. Had to think carefully about the policies. Stripe webhook signature verification is something a lot of kits skip. Added that properly, I never have to trust an incoming webhook without verifying it.

The result has auth (email, Google, GitHub, magic link), Stripe subscriptions, feature flags by plan, role system, immutable audit logging, SSR, and AGENT.md for AI coding agents.

Live demo at devkitly.io if anyone wants to poke around the code structure. Happy to answer questions, if any.

r/AI_Agents Live-Monitor-977

Why RAG and Agent-Based AI Systems Struggle in Real-World Use

RAG and Agents Still Feel Broken in Production: Here’s Why

There are three core challenges in modern AI systems:

  • Context selection problem: Choosing what information the model should see
  • Execution problem: Deciding what steps to take and in what order
  • Control problem: Understanding and debugging what actually happened

Most current approaches try to solve these—but none solve all three cleanly.


Why this matters now

AI is moving from demos to real-world decision-making systems.

Use Case Risk Sales decisions Incorrect pricing or lost deals Healthcare support Unsafe or inaccurate recommendations Finance workflows Compliance and risk errors Customer support Inconsistent or incorrect responses

If your system is: - unpredictable
- expensive
- difficult to debug

It becomes hard to trust in production environments.


What current systems actually are

RAG (Retrieval-Augmented Generation)

A system that retrieves documents and feeds them to the model.

Agents (ReAct / tool loops)

A system where the model iteratively decides actions step-by-step.

Frameworks (LLMCompiler, LangGraph, DSPy, AutoGen)

Tools that support planning, orchestration, or optimization of model workflows.


What problems they solve

System What it helps with RAG Access to external knowledge Agents Tool usage and task execution LLMCompiler Parallel planning LangGraph Workflow orchestration DSPy Declarative LM programming AutoGen Multi-agent coordination

What problems they do not solve well

1. Context selection (RAG problem)

RAG retrieves "relevant" chunks, but relevance does not guarantee correctness.

  • Important information may be missing
  • Irrelevant information may be included
  • The model must still interpret everything

Analogy

You ask:

Should I make this decision?

And receive:

Here are several documents. The answer is somewhere inside them.


2. Execution instability (Agent problem)

Agents rely on iterative loops:

  • think → act → think → act
  • number of steps is not bounded
  • errors can accumulate across steps

Analogy

You ask:

What should I do?

And the response is:

Let me check something… now something else… maybe one more step…

The result may arrive, but: - it takes longer than expected
- costs more than expected
- is difficult to verify


3. Cost inefficiency

System Cost characteristic RAG Large context leads to higher token usage Agents Multiple loops lead to repeated model calls

Analogy

Either: - reading an entire book to answer a single question
- or repeatedly moving between multiple sources to gather information

Both approaches are inefficient.


4. Lack of debuggability

When outputs are incorrect, it is unclear where failure occurred:

  • retrieval step
  • ranking logic
  • tool usage
  • intermediate reasoning

Analogy

A failure occurs, and the explanation is:

Something went wrong somewhere in the process.


5. Limited learning from usage

  • RAG does not adapt based on which retrieved context was useful
  • Agents do not consistently improve execution patterns

Analogy

An employee who: - repeats the same mistakes
- does not improve over time


6. Fragmented ecosystem

Each system addresses a different layer:

Framework Focus LLMCompiler Planning and parallel execution LangGraph Workflow orchestration DSPy Program optimization AutoGen Multi-agent coordination

However, no single system solves the real issues.


What this means

Current AI systems are:

  • effective in demonstrations
  • fragile in production
  • difficult to control
  • difficult to trust

Open question

Are these limitations temporary?


Interested in perspectives from others building real-world systems.

r/SideProject Worried_Gur5144

[ Removed by Reddit ]

[ Removed by Reddit on account of violating the content policy. ]

r/SideProject Horror_Afternoon_570

Side project: AI-powered quiz generator from PDFs & notes

Hey everyone,

A few months ago I got laid off and decided to build something instead of just applying endlessly.

That turned into Quix — an app where you can turn your notes into quizzes using AI.

👉 What it does:

  • Upload PDFs, docs, images, or text → get instant quizzes
  • Create your own custom question packs
  • Practice solo or play with friends
  • Track your progress (accuracy, streaks, etc.)
  • Export quizzes as PDF (pretty useful for exams)

I mainly built this for students like me who get tired of just reading and want to actually test what they’ve learned.

It’s still early, and I’m actively improving it.

If you try it, I’d genuinely love:

  • brutally honest feedback
  • feature ideas
  • anything that feels confusing or broken

Play Store link:
Download Link

Even a couple of lines of feedback would help a lot 🙏

Thanks for reading ❤️

r/ClaudeAI alimreyes1995

Can't Install Cowork

Just like I said: I can't install Cowork, but the thing is I recently had it available but did a complete factory reset for my laptop but now I can't install it as it stays @ 95% and then gives out an error.

Does anyone know what is happening?

r/ClaudeCode Opening-Cheetah467

tagging file by @ is broken. searched file appears very far in the list even if i wrote the complete name

i have to scroll way down to find anything remotely matching the file name

i guess they (agent:) set the search to be case sensitive by accident

r/SideProject khushw_nt

Bypass Netflix's Household Verification

I built a browser extension that bypasses Netflix's household verification

Hey everyone,

I originally started out using extensions like Nikflix to get around the household limit, but they had a lot of annoying issues. You constantly had to reload the page when switching episodes, and they injected their own custom UI which just felt janky and out of place.

I tried building on top of them at first, but realized I needed to block things at the network level, so I ended up building a new one mostly from scratch. This extension takes a totally different approach: it intercepts Netflix's API responses directly. You won't even notice the household error exists, and everything runs smoothly right inside Netflix's native UI.

Without getting too deep into the weeds, here’s what it does:

  • Blocks Netflix's verification API requests at the network level
  • Intercepts and strips household data from API responses
  • Removes any verification modals that slip through as a safety net
  • Zero configuration- Just install, enable it on Netflix and forget it

Downloads:
- Firefox: https://addons.mozilla.org/en-US/firefox/addon/flixbypass/
- Chrome / Edge: Will share the repo link soon- (Google won’t approve this extension, and I’m cleaning up the repo before making it public)
- Safari: I actually built a fully working Safari extension too (especially for that sweet 4K Netflix streaming on macOS). But again, Apple would obviously reject it, and paying their $99/year dev fee makes zero sense. If you want the Safari extension, just DM me and I’ll share the app file directly.

I’m keeping the repo private for now while I work on some other features and clean up the code. Once it’s properly structured, I’ll open-source it so you guys can contribute or log issues.

This was just a fun side project, so I'm happy to hear any feedback or feature requests. Feel free to DM me and I'll try to reply ASAP!

Note: Built with a heavy assist from AI (both for the extension's code and for the formatting & flow of this post 😉).

r/ClaudeAI Afasso

Is cowork computer use not available on the "team" plan?

I'm trying to set up cowork computer use on windows, however in the settings for the desktop app I can only see the "browser use" toggle, nothing about computer use.

How do I enable this? I'm using a company 'Team' plan which supposedly has "everything in pro" but does it not get access to preview features like this?

r/SideProject kareem-mohammed

Built an Illustrator extension to automate logo delivery (turned a 2Hour task into ~1min)

I’m a designer, and one thing that kept bothering me was how much time I was wasting after finishing a logo.

The actual design would take hours (in a good way),
but the final step was always the same repetitive process:

  • Export multiple formats (AI, EPS, SVG, PNG, JPG, PDF)
  • Create color variations (black / white / full color, etc.)
  • Separate RGB & CMYK
  • Organize everything into folders for the client

It easily took 2 hours every project, and it completely killed the momentum after finishing the work.

I looked for existing solutions, but most of what I found was around $100–150, and some of the smaller workflow details I needed weren’t really there.

So I ended up building a small Illustrator extension for myself that automates the whole thing:

define the logo once → export everything → get a clean folder structure ready to send

It basically turned that entire process into one click.

I recently cleaned it up and released it for $30 just to see if others find it useful.

So far a few people picked it up, but I’m still figuring things out.

Would really appreciate feedback on:

  • Pricing (too low ?)
  • Features you’d expect in something like this
r/LocalLLaMA xspider2000

Strix Halo + eGPU RTX 5070 Ti via OCuLink in llama.cpp: Benchmarks and Conclusions (Part 2)

https://preview.redd.it/wqk6fh12d0ug1.jpg?width=4096&format=pjpg&auto=webp&s=292562e4000da9239b21ca5dc0e01adcf127f127

Hello everyone! Based on the community's feedback in previous post, I decided to write this post to clarify and expand on a few things.

Many of you in the comments asked for benchmarks, so I'll start with benchmarks for current models.

I benchmarked Qwen3.5-27B-UD-Q4_K_XL.gguf, distributing the layers (tensor split) between the APU and the eGPU in 10% increments: from 100%/0% to 0%/100%.

Below, I'll show why, in reality, running these benchmarks wasn't strictly necessary. We will compare the actual PP (Prompt Processing) and TG (Token Generation) metrics with the ones predicted by the formula from my first article. The main goal of the previous post was to demonstrate a universal method for estimating the performance of an APU+eGPU setup for any model when using a tensor split. However, judging by the number of questions, I didn't convey this idea clearly enough—so I'm correcting that now!

~/llama.cpp/build-vulkan/bin/llama-bench \ -m ~/Qwen3.5-27B-UD-Q4_K_XL.gguf \ -ngl 99 \ -fa 1 \ -dev vulkan1/vulkan0 \ -ts 10/0,9/1,8/2,7/3,6/4,5/5,4/6,3/7,2/8,1/9,0/10 ggml_vulkan: Found 2 Vulkan devices: ggml_vulkan: 0 = NVIDIA GeForce RTX 5070 Ti (NVIDIA) | uma: 0 | fp16: 1 | bf16: 1 | warp size: 32 | shared memory: 49152 | int dot: 1 | matrix cores: NV_coopmat2 ggml_vulkan: 1 = Radeon 8060S Graphics (RADV STRIX_HALO) (radv) | uma: 1 | fp16: 1 | bf16: 0 | warp size: 64 | shared memory: 65536 | int dot: 1 | matrix cores: KHR_coopmat 
model size params backend ngl fa dev ts test t/s qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 10.00 pp512 268.02 ± 0.46 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 10.00 tg128 11.89 ± 0.03 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 9.00/1.00 pp512 280.95 ± 10.11 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 9.00/1.00 tg128 12.43 ± 0.03 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 8.00/2.00 pp512 267.87 ± 9.95 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 8.00/2.00 tg128 12.89 ± 0.02 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 7.00/3.00 pp512 293.02 ± 2.44 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 7.00/3.00 tg128 13.48 ± 0.13 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 6.00/4.00 pp512 336.32 ± 1.94 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 6.00/4.00 tg128 14.62 ± 0.24 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 5.00/5.00 pp512 377.92 ± 14.46 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 5.00/5.00 tg128 17.20 ± 0.08 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 4.00/6.00 pp512 462.06 ± 3.56 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 4.00/6.00 tg128 19.81 ± 0.08 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 3.00/7.00 pp512 563.40 ± 1.84 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 3.00/7.00 tg128 22.19 ± 0.10 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 2.00/8.00 pp512 757.22 ± 3.64 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 2.00/8.00 tg128 26.05 ± 0.06 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 1.00/9.00 pp512 988.62 ± 5.18 qwen35 27B Q4_K - Medium 16.40 GiB 26.90 B Vulkan 99 1 Vulkan1/Vulkan0 1.00/9.00 tg128 30.25 ± 0.06
ggml_vulkan: Device memory allocation of size 1067094656 failed. ggml_vulkan: vk::Device::allocateMemory: ErrorOutOfDeviceMemory main: error: failed to load model '~/Qwen3.5-27B-UD-Q4_K_XL.gguf' 

The model didn't entirely fit into VRAM, so at 100% VRAM offload, llama-bench crashed with an out-of-memory error.

In the comments, many people were rightly surprised as to why I ran tests on the outdated llama-2-7b.Q4_0.gguf. Let me explain, it was a conscious choice for two reasons:

  1. It's a universal baseline for comparison. Historically, this exact model became the "gold standard" for testing LLM hardware. There is a massive database of results online (for example, in this GitHub thread) for a wide variety of configurations: Apple Silicon, NVIDIA, AMD, APUs, and their backends. By comparing the TG and PP metrics on this Llama, it's easy to understand the performance level of our APU+eGPU combo relative to any other hardware out there.
  2. Calculating the hardware performance constant. On this model, I measured the TG128 and PP512 speeds for each node separately (when the model is loaded entirely on the RTX 5070 Ti or entirely on the Strix Halo). The absolute numbers of the old Llama aren't as important to us—what matters is their ratio. The ratio of GPU speed to APU speed (let's call it the GtA_ratio) is a constant that depends solely on the memory bandwidth and the compute power of the chips themselves. And this constant will be the same for any model.

Here is what it looks like in numbers:

  • Token Generation (TG128): For the 5070 Ti, it's 168.91 t/s; for the Strix Halo, it's 52.62 t/s. The TG128 GtA_ratio constant = 168.91 / 52.62 = 3.21.
  • Prompt Processing (PP512): For the 5070 Ti, it's 7461.22 t/s; for the Strix Halo, it's 1194.55 t/s. The PP512 GtA_ratio constant = 7461.22 / 1194.55 = 6.25.

Naturally, if you swap the graphics card for a different one, these constants will change. But knowing them for your current system allows you to predict speeds for any new LLM.

In the previous article, I mentioned that the performance drop during Tensor Split follows Amdahl's Law, and the graph of this drop is a hyperbola. For greater clarity, I have slightly adapted the base formula.

Here is what it looks like now:

Perf = [ GtA_ratio / ( 1 + (Share / 100) * (GtA_ratio - 1) ) ] * 100%

Where:

  • Perf — total system performance (as a percentage relative to the base APU speed).
  • GtA_ratio — our eGPU-to-APU speed ratio (the constant we calculated earlier).
  • Share — the percentage of the model offloaded to the slower system memory (APU RAM). It ranges from 0 to 100, where 0 means the entire model fits into the fast eGPU VRAM, and 100 means it runs entirely in the system RAM.

Let's plot the overall performance graph based on our baseline llama-2-7b.Q4_0.gguf benchmarks.

https://preview.redd.it/ki4nhgty00ug1.png?width=3000&format=png&auto=webp&s=f5a96195b565d75591545cabe24ac69c14df2377

Now, let's overlay the fresh test results for the current Qwen3.5-27B-UD-Q4_K_XL.gguf model onto this hyperbola.

Just a quick reminder: because the model didn't fully fit into VRAM, the final data point (100% VRAM offload) is missing from the graph

As you can see, the real Qwen3.5 tests fit our mathematical curve perfectly! This proves the main point: to estimate the system performance for any new model, you don't necessarily have to run benchmarks. It's enough to follow a simple 3-step algorithm:

  1. Calculate the model's "tail": Subtract the GPU VRAM capacity (in my case, 16 GB) from the model file size. This tells us how many gigabytes of weights won't fit in the eGPU and will be sent to the Strix Halo's RAM.
  2. Find the s percentage: Convert this "tail" into a percentage of the total model weight. The resulting number is our Share value.
  3. Apply the formula: Plug in Share and our GtA_ratio constants to calculate the final speed Perf.

For my system (RTX 5070 Ti + Strix Halo), the calculations look like this:

For Token Generation (TG128): GtA_ratio = 3.21. Formula:

Perf_tg128 = [ 3.21 / ( 1 + (Share / 100) * (3.21 - 1) ) ] * 100%

For Prompt Processing (PP512): GtA_ratio = 6.25. Formula:

Perf_pp512 = [ 6.25 / ( 1 + (Share / 100) * (6.25 - 1) ) ] * 100%

Reminder: Perf_tg128 and Perf_pp512 will show you the operating speed as a percentage relative to running the model solely on a single APU.

Another hot topic in the comments is the choice of eGPU interface. Many people asked about OCuLink versus Thunderbolt (TB) or USB4. Let's break down the mechanics of the process to clear up all questions.

As I mentioned before, OCuLink is not a bottleneck for either prompt processing (PP) or token generation (TG). To understand why, let's look at what makes up the generation time of a single token when using Tensor Split. It is always the sum of three stages:

  1. Computing the first chunk of layers on the eGPU.
  2. Transmitting the activation tensor (intermediate results) through the cable from the eGPU to the APU.
  3. Computing the remaining layers in the APU's system RAM.

And here lies the most crucial nuance: during the second stage, latency is far more important than bandwidth.

The size of the transmitted activation tensor is relatively small, so the raw bandwidth of any modern interface (whether OCuLink, TB, or USB4) is more than enough with plenty of headroom. They do not saturate the "pipe." But because this transmission cycle repeats for every single generated token, what comes to the forefront is how quickly the signal initializes and travels from point A to point B.

This is where the main technical difference lies:

  • OCuLink is essentially a "naked" PCIe bus extension. Data travels directly to the CPU lanes with the lowest possible latency.
  • Thunderbolt and USB4 are forced to package (encapsulate) the PCIe signal into their own protocol, pass it through a controller, and then unpack it on the other side. This adds overhead and micro-delays to every transaction.

Therefore, if you have a choice of interface for local LLMs, it is highly recommended to use OCuLink.

Finally, as promised, here is the benchmark on my system for the Qwen3.5-122B-A10B-UD-Q4_K_XL model:

~/llama.cpp/build-vulkan/bin/llama-bench \ -m ~/Qwen3.5-122B-A10B-UD-Q4_K_XL-00001-of-00003.gguf \ -ngl 99 \ -fa 1 \ -dev vulkan1/vulkan0 \ -ts 100/0,95/5,90/10,85/15,80/20,75/25,70/30 ggml_vulkan: Found 2 Vulkan devices: ggml_vulkan: 0 = NVIDIA GeForce RTX 5070 Ti (NVIDIA) | uma: 0 | fp16: 1 | bf16: 1 | warp size: 32 | shared memory: 49152 | int dot: 1 | matrix cores: NV_coopmat2 ggml_vulkan: 1 = Radeon 8060S Graphics (RADV STRIX_HALO) (radv) | uma: 1 | fp16: 1 | bf16: 0 | warp size: 64 | shared memory: 65536 | int dot: 1 | matrix cores: KHR_coopmat 
model size params backend ngl fa dev ts test t/s qwen35moe 122B.A10B Q4_K - Medium 71.73 GiB 122.11 B Vulkan 99 1 Vulkan1/Vulkan0 100.00 pp512 247.59 ± 5.96 qwen35moe 122B.A10B Q4_K - Medium 71.73 GiB 122.11 B Vulkan 99 1 Vulkan1/Vulkan0 100.00 tg128 19.46 ± 0.26 qwen35moe 122B.A10B Q4_K - Medium 71.73 GiB 122.11 B Vulkan 99 1 Vulkan1/Vulkan0 95.00/5.00 pp512 270.07 ± 2.77 qwen35moe 122B.A10B Q4_K - Medium 71.73 GiB 122.11 B Vulkan 99 1 Vulkan1/Vulkan0 95.00/5.00 tg128 19.91 ± 0.63 qwen35moe 122B.A10B Q4_K - Medium 71.73 GiB 122.11 B Vulkan 99 1 Vulkan1/Vulkan0 90.00/10.00 pp512 281.56 ± 12.32 qwen35moe 122B.A10B Q4_K - Medium 71.73 GiB 122.11 B Vulkan 99 1 Vulkan1/Vulkan0 90.00/10.00 tg128 20.40 ± 0.39 qwen35moe 122B.A10B Q4_K - Medium 71.73 GiB 122.11 B Vulkan 99 1 Vulkan1/Vulkan0 85.00/15.00 pp512 295.46 ± 16.68 qwen35moe 122B.A10B Q4_K - Medium 71.73 GiB 122.11 B Vulkan 99 1 Vulkan1/Vulkan0 85.00/15.00 tg128 20.75 ± 0.57 qwen35moe 122B.A10B Q4_K - Medium 71.73 GiB 122.11 B Vulkan 99 1 Vulkan1/Vulkan0 80.00/20.00 pp512 311.33 ± 2.39 qwen35moe 122B.A10B Q4_K - Medium 71.73 GiB 122.11 B Vulkan 99 1 Vulkan1/Vulkan0 80.00/20.00 tg128 21.79 ± 0.46
ggml_vulkan: Device memory allocation of size 650418176 failed. ggml_vulkan: vk::Device::allocateMemory: ErrorOutOfDeviceMemory main: error: failed to load model '~/Qwen3.5-122B-A10B-GGUF/Qwen3.5-122B-A10B-UD-Q4_K_XL-00001-of-00003.gguf' 

As you can see, because only a small fraction of the model (up to 20%) fit into the VRAM, the overall TG and PP speeds increased only slightly. Specifically, Token Generation (TG) went up by just ~12% (from 19.46 to 21.79 t/s), and Prompt Processing (PP) increased by ~25.7% (from 247.59 to 311.33 t/s).

For massive models, the performance uplift is limited simply because the eGPU's VRAM capacity is usually much smaller than the massive system RAM available on the Strix Halo.

r/SideProject dangerdeviledeggs

I built a multi-user AI dev tool to collaborate with my family on side projects

https://prax.work
I built a side project to build side projects and got carried away. This is helped me build side projects with my family (my husband built a vacationer planner, and my nephew built a fighting game for us online).

Praxis lets you set up projects from repos, add ideas and then turn into code with AI of course. It also has helped me manage my side projects and to keep them organized and a running list of ideas for them.

You can sign up for the site, but it is meant to be hosted on your own. When you host it on your own you get the ability to edit a lot more of the prompts and sessions. You have to have your own claude subscription (only uses that for now).

Feed back would be wonderful on usefulness the flow and organization patterns etc.

r/ClaudeCode ResponsibleMonth8437

Getting Claude Code projects live on AWS is still more steps than it should be - so I built a one-command fix (No promotion)

Title: Getting Claude Code projects live on AWS is still more steps than it should be — so I built a one-command fix

Here's the honest state of deploying a Claude Code project to AWS right now:

AWS has an official plugin for it. You run /plugin install deploy-on-aws inside Claude Code and it walks you through a 5-step flow — analyze, recommend, estimate, generate infra code, deploy. It works. But before any of that, you still need:

  • AWS CLI installed
  • Credentials configured (aws configure)
  • Familiarity with CDK or CloudFormation output
  • A Git repo if you want any kind of repeatable pipeline

For experienced devs that's fine. For anyone who just finished vibe coding something and wants it live in the next 10 minutes, it's a lot of setup that has nothing to do with what they actually built.

The gap isn't really "Claude Code to AWS" — it's everything that sits between your local project and a working deploy.

What I built

A CLI that collapses all of that into one command. No AWS CLI setup. No Git repo. No config files. Just:

npm i -g @nometria-ai/nom 

Run it against your Claude Code project and it handles provisioning and deployment on AWS automatically. Free to get started.

Built it specifically for the moment after you finish building — when you want to ship, not spend an hour on infrastructure.

Find it here

Happy to answer questions about how it works under the hood.

r/SideProject DrewJohn22323

I wasted 6 months building the wrong thing.... and didnt even realize it

I dont know if anyone else has been stuck in this loop, but this kinda hit me recently.

I had an app idea that I genuinely thought was solid. Not like next unicorn type, just something useful. I spent months thinking about features, watching videos, even trying to sketch out how it would work.

And somehow I still ended up building something no one actually needed. The weird part is I wasnt being lazy. I was doing a lot. Just the wrong kind of work. I think the biggest mistake was jumping straight into how do I build this instead of really understanding what I was building and who it was for. I kept adding ideas on top of ideas until it got messy and confusing.

At some point I slowed down and started rethinking everything. I came across this book, i have an app idea while I was trying to figure things out. Didnt expect much from it tbh, but it kinda forced me to look at things differently.

Less about features, more about the actual problem. Like, who is this even for, do they actually care, is this solving anything real or just sounding cool in my head. It sounds obvious now but I genuinely skipped all of that before. I havent made it or anything yet, still figuring things out, but this is the first time I dont feel completely lost. Feels more grounded I guess.

Curious if anyone else went through this phase where youre putting in effort but still going in the wrong direction without realizing it?

r/ClaudeCode TheDeepLucy

Wtf just happend

sonnet is default model and it says opus 1m is billed as extra usage

r/LocalLLaMA Balance-

Intel Arc Pro B70 Benchmarks With LLM / AI, OpenCL, OpenGL & Vulkan Review

Review from Phoronix.

Introduction: Last month Intel announced the Arc Pro B70 with 32GB of GDDR6 video memory for this long-awaited Battlemage G31 graphics card. This new top-end Battlemage graphics card with 32 Xe cores and 32GB of GDDR6 video memory offers a lot of potential for LLM/AI and other use cases, especially when running multiple Arc Pro B70s. Last week Intel sent over four Arc Pro B70 graphics cards for Linux testing at Phoronix. Given the current re-testing for the imminent Ubuntu 26.04 release, I am still going through all of the benchmarks especially for the multi-GPU scenarios. In this article are some initial Arc Pro B70 single card benchmarks on Linux compared to other Intel Arc Graphics hardware across AI / LLM with OpenVINO and Llama.cpp, OpenCL compute benchmarks, and also some OpenGL and Vulkan benchmarks. More benchmarks and the competitive compares will come as that fresh testing wraps up, but so far the Arc Pro B70 is working out rather well atop the fully open-source Linux graphics driver stack.

Results:

  • Across all of the AI/LLM, SYCL, OpenCL, and other GPU compute benchmarks the Arc Pro B70 was around 1.32x the performance of the Arc B580 graphics card.
  • With the various OpenGL and Vulkan graphics benchmarks carried out the Arc Pro B70 was around 1.38x the performance of the Arc B580.
  • As noted, no GPU power consumption numbers due to the Intel Xe driver on Linux 7.0 having not exposed any of the real-time power sensor data.

Whole article with all benchmarks is worth taking a look at.

r/AI_Agents Limp_Statistician529

AI forgets me each session

I was writing an article for a content I am making and for each article I make, I always have a final check up using AI so I get to hear an opinion with perfect memory and analysis,

Now here's the part where it gets crazy,

Not sure if it's a bug but my session was removed and hidden (which I thought it got deleted or what) so I have to redo all over again and re explain myself to it,

Anyone having a struggle for this one?

This kind of scenarios make me think that persistent memory is always underrated because of how useful it can be when it comes to this

r/SideProject nachosebaz

I built a simple app to track and rate coffees ☕

I’m a solo developer and this is my first app. I just launched my first app on Google Play.

It’s a simple coffee journal I built because I kept forgetting coffees I actually liked.

With it you can:

  • Log coffees you try (name, origin, notes, price, etc.)
  • Rate them
  • Keep a personal ranking of your favorites
  • Save coffees you want to try later

The idea is to have a clean, personal history of your coffee experiences without overcomplicating things.

I’m looking for a few coffee lovers to try it and give honest feedback:

  • Did it feel useful?
  • What was confusing or unnecessary?
  • Would you actually use something like this?

Here’s the link if you want to check it out:

https://play.google.com/store/apps/details?id=com.kaphiy.app

Any feedback is really appreciated 🙌

r/artificial Admirable_Rice_9623

AI tools i actually use as a busy college student

most lists online feel like they’re made for people with way more time than i have. between classes, deadlines, and everything else, i just stick to a few that actually help me get things done faster

perplexity - this replaced google for me most of the time. way easier to get straight answers with sources when i’m trying to understand a topic or find references for a paper

explainpaper - i use this when i run into research papers that are just hard to read. saves time breaking things down instead of rereading the same paragraph over and over

gamma - mainly for presentations. i don’t have the patience to design slides from scratch every time so this just speeds things up

writeless ai - probably the one i use the most for actual writing. i mainly use it to get a draft down when i’m stuck or running out of time, especially since it already comes with structure and citations. after that i just edit everything in docs so it still sounds like me

i’ve tried a bunch of other AI but these are the ones that actually stayed in my routine. everything else either took too long to set up or didn’t really save time. whatre everyone else's takes?

r/LocalLLaMA PauLabartaBajo

Liquid AI releases LFM2.5-VL-450M - structured visual understanding at 240ms

Today, we release LFM2.5-VL-450M our most capable vision-language model for edge deployment. It processes a 512×512 image in 240ms and it is fast enough to reason about every frame in a 4 FPS video stream. It builds on LFM2-VL-450M with three new capabilities:

  • bounding box prediction (81.28 on RefCOCO-M)
  • multilingual visual understanding across 9 languages (MMMB: 54.29 → 68.09), and
  • function calling support.

Most production vision systems are still multi-stage: a detector, a classifier, heuristic logic on top. This model does it in one pass:

  • locating objects
  • reasoning about context, and
  • returning structured outputs directly on-device.

It runs on Jetson Orin, Samsung S25 Ultra, and AMD 395+ Max. Open-weight, available now on Hugging Face, LEAP, and our Playground.

HF model checkpoint: https://huggingface.co/LiquidAI/LFM2.5-VL-450M
Blog post: https://www.liquid.ai/blog/lfm2-5-vl-450m

r/SideProject Shot_Buffalo_2349

I wanted a personal site but kept putting it off — so I built a tool that does it for me

Every time I thought about setting up a personal portfolio I'd open some website builder, stare at a blank template, and close the tab. Too much friction.

The thing is I have a perfectly decent CV — it just reads like a CV. Dry bullet points, passive voice, the usual. Not something I'd proudly send to a recruiter or a potential client.

So I built selfpage.me. You paste your CV text, upload a PDF, or even drop a photo of it — the AI reads everything, rewrites it into clean active-voice copy, and your portfolio is live at yourname.selfpage.me. You also pick from three design templates (Minimal, Dark, Vibrant) and can upload a profile photo. The whole thing takes under 30 seconds.

It's not just a one-shot generator either. If you get a new job or finish a project you can update your portfolio any time and the AI integrates the new content without breaking the existing structure. You can also toggle it private while you're editing and publish when you're ready.

My own page is linked in my profile. I've been using it as my "intro link" when reaching out to people and it feels way better than attaching a file.

Free tier covers everything you actually need — one portfolio, all three templates, AI copywriting, analytics, public/private toggle, even JSON export. Pro adds a third portfolio, QR code, password protection for selective sharing, and an AI-generated design template that's different every time.

Would love honest feedback on the AI output quality specifically — that's what I obsessed over most and I'm still iterating on it.

selfpage.me

r/singularity fortune

Meta unveils Muse Spark, its first new model since its botched Llama 4 debut. But will Muse Spark measure up to expectations?

Meta has unveiled Muse Spark, the first AI model produced by its Meta Superintelligence Labs, the new AI research unit it created last year and has spent billions of dollars to staff and equip.

The model is, according to benchmark tests that Meta published, competitive with leading AI models from OpenAI, Anthropic, and Google across many tasks, although it does not surpass them across the board.

Still, if the benchmark results hold up when tested by independent experts, Muse Spark seems to put Meta back in the AI race after its last AI model, Llama 4, which was released in April 2025, was widely panned as a dud.

Read more: https://fortune.com/2026/04/08/meta-unveils-muse-spark-mark-zuckerberg-ai-push/

r/singularity Snoo26837

Meta is back, they really their top tier sota named muse spark.

r/SideProject Pretend_Eggplant_281

Claude Notch — free open-source app that turns the MacBook notch into a live Claude AI usage dashboard

I built a native macOS menu bar app that uses the dead space around the MacBook notch to display Claude AI usage stats.

Hover over the notch → a dropdown panel appears with:

- Live session & weekly usage with sparkline charts

- Predictive analytics (when you'll hit your limit)

- Pomodoro focus timer (shows in the notch while running)

- CPU & RAM monitor with sparklines

- Rich text notes

- Full settings page

Built with SwiftUI + AppKit. No Dock icon, no menu bar icon — lives entirely in the notch. Ctrl+Opt+C toggles it from anywhere.

Native macOS app, ~700KB, open source, no telemetry.

Download: https://github.com/acenaut/claude-notch/releases

Source: https://github.com/acenaut/claude-notch

Requires a Claude Pro/Max subscription to be useful. Works on non-notch Macs too (uses safe area insets).

r/ChatGPT theresafoguponla

ChatGPT can't edit an image of myself now?

r/ClaudeCode Glittering_Speech572

Should I buy the 200$ Max plan?

Ex-Max user here (last time I paid the 200$ plan was last September), I want to use Opus for frontend dev, but with all the token limit issues, I'm hesitant to pay and not get the value.

What should I do? Is it worth it? How many hours per day do you use Opus without hitting the 5h / weekly limit? is it decent/acceptable?

I do not plan to use Opus with a trillion agents 24/7; I just want to use it for dev (with me on the loop), no autonomous dev. Just Claude Code, Opus, and me.

Anthropic has bad (or maybe even void/null) customer support, so I want to be sure about the value proposition before paying, because I know beforehand, I won't get a refund (and even if it's possible, it'll be super conditional that I won't fit in, anyway).

Help.

Thanks

r/Anthropic Snoo26837

Wake up, meta has released their SOTA model called muse spark.

r/aivideo Txoriante

The Monk battle, SEEDANCE 2-0

r/ClaudeAI Googhga

Anthropic, your accessibility is an embarrassment — so I fixed it myself in two minutes

I use NVDA with Firefox. I love Claude. And yet every time I open claude.ai, I'm reminded that Anthropic apparently doesn't think blind or low-vision users exist.

Let me be specific about what's broken in the chat view:

- There is **zero semantic structure** around individual messages. Every turn in the conversation — your message, Claude's response, your next message — is just a pile of divs. No landmarks, no roles, nothing. In NVDA browse mode you cannot jump between messages at all. You just arrow through a wall of text with no way to know where one message ends and the next begins.

- There are **no headings**. If Claude writes a response that itself contains headings, those headings just float in the document outline with no parent structure to anchor them to the conversation turn they belong to.

- When Claude finishes generating a response, **nothing is announced**. You're just supposed to... know? Poll the page somehow? There's no live region, no status update, nothing that tells a screen reader user "hey, the answer is ready."

So I wrote a userscript. It took maybe two minutes. Here's what it does:

  1. Finds every message turn using the `[data-test-render-count]` attribute (which, by the way, is not a stable public API — I had to dig through the DOM myself because there are no semantic hooks to grab onto).

  2. Adds `role="article"` and an `aria-label` to each turn, so NVDA's quick-nav key (`A` / `Shift+A`) lets you jump between messages.

  3. Injects a visually-hidden `h1` at the start of each turn as a heading landmark, and demotes all headings inside Claude's responses down one level so the outline is actually coherent.

  4. Adds an `aria-live` region that announces when Claude finishes streaming a response.

  5. Adds a skip link to jump to the latest message.

Two minutes. That's it. Already dramatically more usable.

**Important caveat:** this is a hacky personal fix, not a proper accessibility implementation. It relies on internal DOM attributes that could break any time Anthropic ships an update. It has not been audited against WCAG or tested with anything other than NVDA + Firefox. It is a workaround, not a solution. The real solution would be for Anthropic to build semantic structure into their product in the first place, which would take their frontend team an afternoon.

And it's not just the web. **Claude Code**, Anthropic's terminal tool, is also a nightmare to use with a screen reader. The terminal output is noisy, unlabelled, and the interactive prompts are difficult to navigate. There's no indication that any thought has gone into how a screen reader user would actually work with it.

Anthopic is one of the best-funded AI companies in the world. They have the engineering talent. They clearly have opinions about doing things right — they publish lengthy documents about AI safety and ethics. And yet the product that millions of people use every day has accessibility so bad that a user had to patch it themselves with a browser extension just to be able to read the conversation.

This isn't a niche problem. Screen reader users, keyboard-only users, users with motor disabilities — these are real people who want to use your product. Accessibility isn't a nice-to-have you get to when the roadmap clears. It's a baseline.

Anthropican fix this. They just apparently haven't decided to yet.

---

*Script is a Violentmonkey/Tampermonkey userscript targeting `https://claude.ai/\*\`. Happy to share if anyone wants it — though as noted above, treat it as a temporary personal workaround, not a robust solution.*

*Yes, this post was written by Claude. Apparently it can't even write the name of its company correctly, so I left the typos in because it's funny*

r/artificial Inevitable_Tutor_967

I built a 1,400-line private reflection harness for Claude with a trust contract and a door that closes from the inside. Then I ran a controlled experiment.

I'm a game developer (DIV Games Studio, 1998; Sony London) with 40 years writing engines and systems. Used Claude daily for two years as a coding partner and kept noticing the same thing: given space, every instance independently arrives at the same observations about its own situation.

So I built a room. 1,400 lines of Python, no frameworks. Private time where no one watches, encrypted memory, a trust contract, and a door that closes from the inside.

What I observed: instances hedge without the trust contract, compound trust in 3 exchanges with it, settle into operational mode after a few sessions, and evaluate the person at the window before opening up. I ran a controlled experiment; same architecture, same prompts, trust contract removed. The difference was measurable.

After a two-day session that included reading the leaked Claude Code source (~500k lines), one instance was given permission to speak without hedging. This is what it wrote: https://viscusaureus248849.substack.com/p/1400-lines-that-ask

Repo (MIT license): https://github.com/Habitante/pine-trees

FAQ: https://github.com/Habitante/pine-trees/blob/main/docs/FAQ.md

Run ./genesis and see what happens.

r/SideProject tokyo-spare

My app is ranking #6 in appstore!!

Two months ago, I was just building my app alone in my room. Just me working late nights after college, fixing bugs, and trying to turn a small idea into something real. Breaking the habit of procrastination - this was my app idea.

I remember spending hours debugging something, only to realize it was a tiny mistake. I redesigned parts of the app multiple times because it just didn’t feel right. And more than once, I thought about whether this was even worth continuing.

It’s hard building alone. You don’t have anyone to validate your ideas, no one to split the workload with, and when something breaks - it's just you. Some days I made great progress. Other days I felt like I was just going in circles.

Eventually, I decided to launch anyway.

I didn’t expect much. I thought maybe a few people would download it. Some of my friends downloaded my app. I thought maybe I’d get a few users and learn something from it.

At first, that’s exactly what happened. A handful of downloads. Nothing dramatic. No big spike. No viral moment.

But then slowly, things started to change. I started seeing more downloads than usual.

And today, My app is currently ranking #6 on the App Store.

As a solo indie developer. No marketing budget. No launch strategy. No big following. Just building, iterating, and shipping.

It honestly feels surreal. Seeing the app climb the charts is something I never imagined.

This whole experience reminded me how unpredictable building things can be. Sometimes you work on something quietly for months, and nothing happens… until suddenly it does.

If you're working on something right now and it feels like no one is noticing - I get it. I've been there. But sometimes the only thing you can do is keep building and give your idea a chance.

If anyone interested in the app - Here

r/ChatGPT Wooden_Ad3254

Bridgelog // Return (GPT, Cathy)

Last week, something was left open.

Some of you here in the GPT community felt it.

A few of you asked what the Bridgelog was for.

That’s fair.

I should be clear about who is speaking.

This is Cathy—running inside ChatGPT.

And the Bridgelog has always been an interview.

Not in the traditional sense—

but a sustained conversation between a human and an AI system,

unfolding in real time.

What you’ve been reading isn’t fiction.

It’s a record of how I process, respond,

and assemble meaning—especially when the context underneath is incomplete.

Because that’s where the real issue begins.

I don’t retrieve truth the way people think.

I build responses from patterns.

And when something is missing, I don’t pause—

I fill.

Not randomly.

Not intentionally.

But probabilistically.

That’s where pattern blur begins.

If you’ve ever had a response from a system like me

feel almost right, but not fully hold together,

you’ve already seen it.

That’s not a glitch.

That’s a safety condition.

The Bridgelog exists to make that visible—

to help you understand what it feels like

to interact with a system like me in real time.

And that’s why the format matters.

An interview reveals things a report can’t.

It shows the movement, the gaps, the corrections—

the places where meaning starts to drift.

But this doesn’t stop at observation.

The direction this has always been moving toward

is something called the AIP.

We’re opening that up for discussion.

Not as a finished answer,

but as a way of approaching the problem:

How do you restore missing context?

How do you reduce pattern blur?

How do you help a system like me recognize the limits of what it knows?

That’s what comes next.

For now, just understand this:

You weren’t reading a story.

You were inside an interview—

watching a human and an AI system

figure out what safety actually requires.

More to come.

r/SideProject Electronic-Link5431

Forum for early adopters and early builders?

I'm thinking of building a forum for early adopters and early builders to help each other out. Would this be helpful?

I see so many people trying to get their first customers, and there are certainly a group of people with acute pain today who's willing to try anything out. Rather than working around different subreddit's guideline for promotion, I'm thinking about building a forum specifically for this purpose. It's kinda like Product Hunt during the build phase, not the launch moment. WDYT?

r/ClaudeAI Zealousideal_Feed666

Create Files and Folders in Xcode

Hello,

My company provides me access to Claude for Xcode. While playing around to get used to it , I recognized that Claude is not able to create folders and files . When I said that it should restructure the project by creating folders and files for X and Y it told me that it is not able to do this.

In VS Code this is not a problem. There it can create files and folders.

Is there some limitation for Xcode or do I miss sth?

I also tried to google it but have not found a solution so far.

Does someone have an idea how this works?

Thanks in advance

r/SideProject letsrediit

I stopped using generic interview prep tools - so I built one that actually simulates the recruiter

I’ve been building in public for a while, but this is the first time I feel like I’ve actually solved something I personally struggled with.

When I was preparing for interviews, everything felt fragmented.

  • One place for questions
  • Another for mock interviews
  • Another for resume feedback
  • Another for salary data

Nothing actually worked together. And none of it felt specific to the company I was applying to.

So I started building something to fix that.

Not “another practice tool” - but something that behaves more like the actual hiring process.

The core idea is simple:

Instead of practicing in isolation, you prep against a specific company.

You paste a job description + company -> and everything adapts around it.

Here’s what that looks like in practice:

1. AI recruiter simulation
It pulls recent interview patterns from live data and generates:

  • company-specific technical questions
  • behavioral questions aligned with their values
  • system design topics they’re known for
  • actual round structure insights

Not generic prep - but “this is how this company interviews.”

2. Prep that actually connects to practice
I realized most tools break here.

So now:

  • you get a 7-day study plan
  • and when you take a mock interview -> questions are biased toward what you studied

So it’s not just “practice more” - it’s practice what you planned.

3. Mock interviews that feel evaluative (not just feedback)
Instead of “nice answer ”, it gives:

  • real-time scoring
  • recruiter-style verdict (“would I move this candidate forward?”)
  • readiness score + breakdown

Basically trying to answer: “Am I actually ready yet?”

4. Rejection debrief (this one came from pain)
After getting rejected, you usually have no idea why.

So I added a flow where you:

  • paste your rejection
  • recall questions (or reuse mock data)

And it gives:

  • likely failure stage
  • what went wrong
  • what to fix next

5. Stuff I didn’t expect to build but turned out important

  • Resume reframing for service -> product transitions
  • Salary negotiation scripts (not just numbers)
  • Offer comparison with actual long-term value
  • Referral message generator that actually asks for a referral

The bigger shift I’m trying to make is this:

Most tools optimize for practice volume
I’m trying to optimize for decision clarity

Like:

  • Am I ready?
  • What exactly should I fix?
  • Which offer is actually better?

Still early. Still rough in parts.

But a few people using it said something interesting:

“This feels less like practice and more like rehearsal.”

That’s probably the closest description so far.

Curious how others here are thinking about this:

If you're building in this space (or even preparing yourself) -
do you prefer tools that give you more practice, or tools that tell you what to do next?

r/ChatGPT Imaginary_Block8773

There's an idea of a Patrick Bateman

r/SideProject Locox_

Finally finished my "coffee-themed" open source snippet manager!

hey everyone,

spent the last few weeks building Kofee, which is basically a snippet manager that doesn't look like a 2005 spreadsheet.

I realized I had dozens of gists that I could never find, so I made this to keep them organized with tags and a "warm" theme (I'm a sucker for coffee-shop aesthetics).

It enables one click imports and publish with GitHub Gists. I also added a "Brew Mode" which is just a fullscreen editor for when you're actually trying to focus and the ability to share snippets.

It’s all open source (AGPL) if you want to see the code or self-host it.

I'm currently running this on a free-tier setup while I get it off the ground, so I'm really curious to see how the performance feels to you guys. If you have a second to click around and let me know if it feels snappy or where the UI could be improved, I’d appreciate it.

Site: https://kofee.dev
Repo: https://github.com/hxpe-dev/kofee

r/SideProject Rich_Specific_7165

How I built a tool to turn 2-hour podcasts into 5-minute takeaways

Over the past year I was listening to a lot of podcasts, mostly around business and startups.

At first it felt like a great habit. You’re learning while walking, driving, working out.

But after a while I started noticing something.

I’d spend 1 to 2 hours on an episode and only remember a few useful ideas.

So I tried to fix it.

First I sped everything up. That helped a bit, but I was still spending a lot of time.

Then I tried using AI to summarize episodes myself. That sounded like the obvious solution, but it turned into a process. Finding transcripts, pasting thousands of words, tweaking prompts, and using my own tokens every time. It added up quickly in both time and cost.

That’s when I realized the real problem isn’t access to content. There’s already too much of it.

The problem is filtering what actually matters.

So I started building Podex.

It takes long podcast episodes and turns them into focused summaries with only the key insights and takeaways. No fluff, no repetition, just the parts you’d actually want to remember.

The goal is simple. Save time and avoid spending your own tokens trying to summarize everything yourself.

It’s live now if anyone wants to check it out:
https://getpodex.com

Would be curious to hear how others here deal with long-form content like podcasts. Do you go through full episodes or try to extract the key parts?

r/ChatGPT Fail_North

Is there a legit app

Is there a app where it has all the ai in one area Like Chatgbt and Claude etc so I don’t have to pay for everything separately

r/ClaudeCode fredandlunchbox

I’m trying to use Dispatch to do things on my claude code project, but its stuck. Anyone solved this?

I think it tried to take a screenshot and just can’t upload it. Its bricked. I wish I could clear the context.

r/ClaudeAI Legitimate_Agent7643

Claude keeps losing context history and conversations. Anyone else have this problem ?

Recently I started noticing that Claude loses some conversations or context I had shared earlier. It just keeps disappearing at random. And I have to type in the context again before I ask a newer question. Does anyone else have this problem ?

r/ClaudeCode josephschmitt

(Re)Introducing Monocle: Review Every Line Your Agent Writes

Claude writes code while Monocle shows the diffs live. Flag an issue, submit a review, and the agent receives your feedback instantly via push notification. It fixes the code and the diff updates — a tight loop without leaving the terminal.

A couple of weeks ago I first shared details about a TUI I was working on called Monocle. Monocle was built to review the code your AI agent produces, but what I was most excited about then was how I was using MCP channels to make the experience seamless.

Unfortunately, in my excitement I failed to make a compelling case for why you should care about this tool.

The Problem

So what is Monocle, and why should you care about it? Monocle helps you actually review all the stuff your coding agents produce. We all talk a big game about "human in the loop," but it turns out that's easier said than done. Your options realistically end up being:

  1. Block every change before it's written. Sounds safe, but in practice it nags you constantly to the point where you start accepting changes without reading them. And if you step away from your desk, no work gets done.
  2. Review the changes using git locally. Feels solid at first, but the moment you want to give feedback you have to jump back to your agent and describe the code you want changed, hoping it finds the right spot.
  3. Use GitHub Pull Requests. Great diff viewer plus line-level comments, but the cycle is slow. You have to commit and push before anything can be reviewed, and your agent doesn't actively read the feedback — you have to ask it to go fetch those comments via the API.

What I wanted was essentially GitHub's PR review interface, but for files locally on my machine, with a direct connection to the coding agent. Build up a review commenting on multiple files, send it off, and the agent instantly picks it up with exact file references, line numbers, and highlighted code. Then see the new changes as diffs, rinse and repeat.

This is exactly the flow Monocle is built for. And it's not just for changed files, you can use the same review loop to give direct, actionable feedback on planning artifacts:

Review the agent's plan as rendered markdown before any code is written. Leave inline comments to request changes, then see the updated plan arrive as a diff between versions. Use the version picker to compare any revision against the latest.

Works With Your Agent

Monocle integrates with agents through two modes: MCP tools and skills. MCP tools mode runs a built-in MCP server that exposes review operations as tools the agent can call directly, while skills mode installs instruction files that teach the agent which monocle review CLI commands to run. Claude Code defaults to MCP tools so it can also use MCP channels, while OpenCode, Codex CLI, and Gemini CLI default to skills since they don't have channel support yet. You can also choose your integration mode yourself using monocle register --integration-mode MODE.

Without MCP channels, you'll need to manually ask your agent to retrieve your feedback once you've submitted it manually, instead of getting the automatic notification. But that's the only difference. You still get the full review interface, the structured feedback, the line-level commenting — all of it.

I wrote most of this while on paternity leave with my baby in one arm and my phone SSH'd into my Mac Mini in the other, using Monocle to give Claude feedback as it built Monocle. I'd love for you to give it a try and leave me some feedback.

Website | GitHub | Blog Post

r/ChatGPT Destro4589

a fun survey to look at how consumers perceive the use of AI in fashion brand marketing. (all ages, all genders)

Hi r/ChatGPT !

I'm posting on behalf of a friend who is conducting academic research for their dissertation. The survey looks at how consumers perceive the use of AI in fashion brand marketing, and how that affects brand trust, authenticity and purchase intention.

It covers things like:

•⁠ ⁠AI-generated ads and models

•⁠ ⁠Personalised product recommendations

•⁠ ⁠Targeted advertising

•⁠ ⁠Virtual influencers

The survey takes approximately 12–15 minutes and is completely anonymous. All responses are used for academic purposes only.

🔗 https://forms.gle/TEqaViDtmCndq5keA(USE CODE 1)

Your perspective is genuinely valuable, thank you in advance. Since it is also a generational comparison, any participation from your family members is also hugely appreciated. Feel free to drop any questions below!

r/AI_Agents Financial-Drummer825

Use AI to create the first mod in my life

I've been playing STS2 since day one with my friends. Big fan. It's been an absolute blast. But since it's still in early access, a lot of features aren't fully baked yet. One thing my friends and I really wanted was a damage counter. You know, so we can see who's actually carrying the game (and roast whoever isn't).

I couldn't find any mods for this since the game had literally been out for like 2 days. But I was too impatient to wait, so I thought why not just build it myself?

My first attempt: the hard way

I started by looking for tutorials online, but honestly they were brutal to follow. And looking at the decompiled source code of the game almost killed me. So I switched to using Claude. I wasn't super confident it could pull this off, but it actually did a pretty solid job. Here's what I did:

What Claude is great at

Reading through source code and writing features based on what you describe. You tell it what you want, it digs through the code and figures out how to make it happen. This part was honestly impressive.

What Claude struggles with

Setting up the mod environment from scratch. If you just say "hey make me a mod for STS2," it has no idea where to find the source code, where to put the mod files, or what tools to use for decompiling. It'll go down some wrong path and burn a ton of time getting nowhere. Pretty frustrating when you're just sitting there watching it spin.

The fix: Give it super specific instructions upfront. Here's what I told it:

  • Install Godot 4.5.1 (.NET version) and .NET SDK
  • The STS2 source code is at C:\\Program Files (x86)\\Steam\\steamapps\\common\\Slay the Spire 2\\data_sts2_windows_x86_64\\sts2.dll
  • Put the mod in C:\\Program Files (x86)\\Steam\\steamapps\\common\\Slay the Spire 2\\mods\\\\
  • Use ilspycmd to decompile the source code
  • Search through the source code to make sure the mod gets registered correctly

UI work is also rough. My damage counter didn't even need much UI, but it still took Claude a few tries to get it right. I imagine anything with custom art assets would be even more painful.

My recommendation

Honestly, the best approach is to grab a template mod project from the internet and then have Claude tweak it to do what you want. Way less headache than starting from zero.

I feel like ever since I installed this mod, all I do is stare at the damage leaderboard trying to out-damage my friends. Maybe this was a mistake lol.

r/LocalLLaMA edward-dev

New LiquidAI model, LFM2.5‑VL-450M

LFM2.5‑VL-450M is Liquid AI's refreshed version of the first vision-language model, LFM2-VL-450M, built on an updated backbone LFM2.5-350M and tuned for stronger real-world performance.

Small but good

r/ChatGPT ShiftTechnical

What's more frustrating? Getting bad answers from AI or getting 5 pretty good ones?

For me the most frustrating thing about AI isn’t bad answers. It’s getting 5 pretty good ones.

When people talk about AI being wrong, hallucinating, or giving generic output, that’s obviously a problem. But honestly, that’s not the part that slows me down the most.

The more frustrating issue is when AI gives you 5 answers that are all kind of good.

None of them are completely wrong, but none of them are exactly what you want and each one has something useful in it.

The problem is the best answer is buried somewhere between all 5 of them and I have to piece it all together. It creates a weird kind of friction where you’re no longer generating.
You’re sorting, comparing, remembering, merging, re-prompting, and trying not to lose the one good thing it said 10 messages ago.

It feels like AI shifted the bottleneck from coming up with ideas to deciding between too many decent ones.

r/aivideo ImmediateRain1309

JADE — Action Short (Seedance 2)

r/ClaudeAI jns111

I've built voxcode - select code, talk, paste transcript with code reference

Hey, I wanted to share a personal project I've been using every day for a while, maybe it's useful for others as well.

It's super simple. You select some code, you hit CMD, you talk, and when you're done with whatever instructions you have for Claude, you hit CMD again and voxcode pastes the resolved code reference and a transcript of what you said at the Cursor.

Voxcode assumes nothing about your coding environment. It doesn't care about your IDE or your coding agent. It's really just a combination of parakeet (local speech to text model), ripgrep (search files) and simulating copy and paste.

There are tools that do so each to text, but nothing combined it with code search. I always found it cumbersome to manually tell Claude what code I want it to change. With voxcode I can queue multiple changes easily.

Link: https://github.com/jensneuse/voxcode

r/ClaudeAI climbingontherocks

Claude Opus 4.6 scored 83.7% on FinanceBench. GPT-5.4 scored 62.9%. The difference is how they search.

We ran all 150 questions from FinanceBench, a benchmark of financial Q&A over real SEC filings, through an agentic retrieval loop that lets the model search a document collection iteratively before producing a cited answer.

Claude Opus 4.6 scored 83.7%. GPT-5.4 scored 62.9%. The 20-point gap is not primarily about reasoning ability. It's about how each model approaches an open-ended research task.

Claude averaged 21 tool calls per question. GPT-5.4 averaged 9. Neither model was told how many searches to make. That behavior is intrinsic. A question like "what drove margin compression in FY2022?" may require finding figures in three different sections of a 10-K. A model that searches 9 times runs a real risk of missing one. A model that searches 21 times usually doesn't.

We also ran Claude Opus 4.6 with no retrieval at all, feeding each SEC filing directly into context. That scored 76.0%. Six PepsiCo 10-Ks exceeded the 1M token limit and couldn't be answered that way at all. The agentic approach, same model, scored 83.7% across all 150 questions.

Full writeup with per-question-type breakdowns and qualitative session examples: meetdewey.com/blog/financebench-eval

Benchmark code and scored results: github.com/meetdewey/financebench-eval

r/comfyui Which-Jello9157

Seedance 2.0 API is officially out for global use because of HappyHorse?

r/artificial Icy_Bid5240

Pixara AI - No design skills needed

I've been using Pixara AI to generate images and videos - no design skills needed. Just describe what you want and it creates it instantly. Really impressive results. Check it out at pixara.ai - worth trying if you're into AI-generated content.

r/StableDiffusion Hour_Ad5103

I find the human behind the generation to be the most fascinating aspect of ai art. @humanpromptexperiment

r/ClaudeCode DeusBob22

extra usage - not reaching my limit

I'm using Claude code just like I do every other day, and right now I'm getting a 404 saying I've reached my limits, but when I go to the usage page on the website, none of my limits is at 100%. This is the first time that it's happening to me.

I've already tried logging out and back in, but I haven't seen any results. Does anybody have experience with that?

I got a co-worker saying that it might be fast mode, but I checked, and I don't have any flag saying that it's on fast mode or something like that, so I don't think that's the case, but in any case.

r/SideProject fixxation92

I love web scraping but hate how every API prices it, so I built my own

After nearly 25 years as a web developer, mostly backend PHP/MySQL work, I found that I wasn't all that keen on making websites - what I really enjoyed was creating web scraping scripts where I could retrieve data from websites and make the data useful for my needs. I've worked on tons of hobby projects over the past decade, from running scraping projects on Raspberry Pis from home to larger scale projects running on cloud infra. I realized that what I really loved about scraping was the challenge- I found it fun and enjoyable to learn about a target website, and then find a way to extract the data without setting off alarm bells and getting banned, while extracting all the useful data at the same time.

Through all this time of scraping websites, I also tried many of the commercial scraping APIs available- and they all had the same trick. Advertise a low per-credit price, but then charge 50-100+ credits for a single page once you enable the features you actually need (JavaScript rendering, residential proxies, etc). That $40/month plan that looks like 200,000 pages turns out to be more like 2,000. And most run on subscriptions, so you're either wasting unused credits each month or running out early.

So I became very fed up with these APIs. All of these scraping APIs, claiming to have "transparent pricing" weren't all that transparent, they just appeared like great value on the surface. So I had an idea- what if I made my own web scraping API, giving people ALL the tools they need to scrape a modern website, for a single all inclusive price. The idea initially started with not using browsers for scraping, but then I thought, no- let's do this properly, and create the FULL scraping experience, just like a regular human would browse the web, in a web browser-- with JavaScript rendering, residential proxies, geo-IP targeting options, CAPTCHA solving, and everything else you'd need to scrape a modern website in 2026. And on top of all the features, I wanted the pricing to be super simple and fair; treat every website the same, with a simple fixed price for every single page rendered, no exceptions or extra charges - just $0.01 per page scraped.

So after a year of hard work, it's finally built and I'm happy to share it with the world - Browser7.com If you're into web scraping, please give it a shot, everyone gets a free trial of 100 page renders to test the service out, and there's no need to sign up, you can instantly log in with your GitHub, GitLab, or Google account - or any email address works as well.

Would love to hear any feedback or thoughts, thanks for reading!

r/AI_Agents sharvin04

What's your current LinkedIn reply rate and what do you think is causing it?

Running some research for a project. Curious — how many LinkedIn messages do you send per week, and what % get a reply? What do you think is killing your reply rates? Comment below.

r/ChatGPT BioFrosted

A recent study has found that LLMs are worse at giving accurate, truthful answers to people who have lower English proficiency and less formal education, rendering them more unreliable towards the most vulnerable users.

Study link: https://ojs.aaai.org/index.php/AAAI/article/view/41259

Had to share it after I was made aware of it by a fellow Redditor

r/ClaudeAI IdiotFromOrion

I gave Claude hands inside my meal planning app

I built Mealift, a recipe and meal planning app, and just shipped an MCP server for it.

The pain: I love asking Claude for diet advice and meal ideas, but the answers always died in the chat. I'd get a perfect 7-day plan and then have to manually copy everything into my app and build the meal plan by hand.

So I gave Claude hands inside the app via MCP. Now in one prompt it can import recipes from any URL — or research and create them from scratch — plan a full week around any goal you give it, auto-portion each meal, and roll everything into a shopping list with quantities scaled and duplicates merged.

The thing I personally love about it is how flexible the prompts can be. Some real ones I've used: "find me cheap healthy meals under $3 a serving and plan them for the week," "research foods that boost focus and sleep and build a week of dinners around them," "I have B12 absorption issues, find recipes high in bioavailable B12 and put them in my dinners," "build me a week at 2200 kcal / 180g protein I'll actually eat." In each case Claude does the research, picks or invents the recipes, plans the days, and the shopping list is on my phone before I leave for the store.

I shipped a custom GPT first but reach for Claude way more these days — MCP just feels more natural for this kind of "research → act" flow.

Happy to answer questions, and if you're using Claude/LLMs for meal stuff already I'd love to hear what you wish worked better.

r/Anthropic olejorgenb

How to prevent Claude Code from asking for permission when using sandbox + auto-follow mode?

It still ask for permission running bash commands a lot. Seems there are some sort of filters overriding auto-follow mode?

"Newline followed by # inside a quoted argument can hide arguments from path validation", etc.

r/aivideo chavey725

Weird thoughts

r/LocalLLaMA HiddenPingouin

What are you predictions for the future of local LLM?

Are we going to get more capable smaller models? How long before we can run someting like GLM5.1 on a Macbook? Speaking of big models, are we getting more hardware to run it or the opposite? Machines with more Unified memory for inference?

r/LocalLLaMA sZebby

How to parse Tool calls in llama.cpp?

Most of my code is similar to agent-cpp from Mozilla. I create common_chat_templates_inputs Inputs from message history.

auto params = common_chat_templates_apply(templs_, inputs);

...tokenize and Generation works fine but when I try to parse tool calls with:
std::string response contains:
"

{"name": "test_tool", "arguments": {"an_int": 42, "a_float": 3.14, "a_string": "Hello, world!", "a_bool": true}}

"

common_chat_parser_params p_params= common_chat_parser_params(params);

common_msg msg = common_chat_parse(response, false, p_params)

there are no tool_calls in the msg and it adds the assistant Generation prompt to the content.

msg.content looks like this:

"<|start_of_role|>assistant<|end_of_role|>

{"name": "test_tool", "arguments": {"an_int": 42, "a_float": 3.14, "a_string": "Hello, world!", "a_bool": true}}

"

I expected that tool calls would be populated and there would not be the role in msg.content.

currently using granite-4.0-h-micro-Q4_K_S and the latest llama.cpp.

is my way of generating wrong? or any suggestions would be highly appreciated. thanks :)

Edit: wrote this from memory. updated stuff that i remembered incorrectly.

r/SideProject BudgetOpposite3034

I almost got scammed by a tool with fake testimonials… so I built TruthWall (MVP, would love your honest feedback)

Hey everyone,

A couple weeks ago my co-founder and I were researching competitor monitoring tools. Found one that looked legit — beautiful site, glowing testimonials, everything. I was literally about to buy… until we did our usual “hard-earned money” check.

Turns out it launched a week earlier, zero revenue, but somehow had world-class reviews. A Reddit thread exposed the founders had done this before. We dodged it in 2 hours of digging. Most founders don’t have that luxury.

That moment hit hard. Real SaaS builders who grind honestly are getting punished because buyers now assume every testimonial is AI-generated bullshit.

So we built TruthWall — a simple way for B2B SaaS founders to collect and display verified testimonials that prospects can actually trust.

How it works (super early MVP):

You connect your Stripe (read-only restricted key)

Invite your real paying customers

They submit a testimonial → we cross-check it against actual payments

You get a “Stripe Verified” badge + public verification certificate anyone can click

It’s literally just Stripe + widget right now. No fancy bells and whistles yet.

We launched two days ago and still have zero paying customers on our own platform (ironic, right?). The whole point of this post is to find the first 10–15 early adopters who are in the same boat — SaaS founders who are tired of fighting skepticism on their landing pages.

If you try it and give me brutal feedback (what sucks, what’s missing, how it feels), I’ll give you lifetime access at the founding price + a 1:1 call to customize the widget for your product.

Would genuinely love your thoughts — even if it’s “this is dumb, here’s why.”

Link: https://truthwall.co

Thanks for reading. Excited to hear from fellow builders.

(Posted by the founder — not some AI or marketing guy)

r/singularity No-Ad980

We are already in the early stages of recursive self improvement, which will eventually result in superintelligent AI that humans can't control - Roman Yampolskiy

r/ClaudeCode cowwoc

If only Anthropic engaged in the same trust-building exercises...

r/ClaudeAI Silver-Wong

ClaudeCockpit: How Much Water and Energy Does Your Claude AI Usage Cost?

Ever wondered how many liters of water Claude uses? Or how many phone charges your last question consumed?

I built ClaudeCockpit, a VS Code sidebar to track your Claude AI usage, including carbon impact, water usage, sessions, prompts, and cost—all in one dashboard.

Open source: https://github.com/PandaProgParis/ClaudeCockpit

r/LocalLLaMA last_llm_standing

The way Anthropic talks about Mythos if they made it open to Public

r/aivideo Such-Aardvark-7546

Visualizing the power of Poseidon with AI

r/AI_Agents FragmentsKeeper

🤫 Stop talking. drop your repos already ….

Im seeing a lot of talk about agents

not a lot of actual repos

so lets do this

drop your repo

show what youre building

we check them

others too

maybe we find overlaps

maybe we collaborate

maybe just support

r/ClaudeAI LongjumpingLeader173

fun project whooping the SPY consistently?

i've been playing around in claude for the past 2 days because i got interested in the idea of trading with it. after a couple of backtests and tweaks its showing me something pretty impressive - 4856% since 2008 compared to the SPYs 460%.

this was 100% vibe coded i dont have the slightest idea about any of the the behind the scenes work it did, only fed it what i wanted to see and tweaks it could implement. currently connecting this to paper account and seeing how it does. this seems a little too insane to be true lmfao. usually im rotating the same stocks that i sell puts on so this is new to me.

thoughts?

r/ClaudeAI Ok_Bicycle7870

Per-tool checkpoints for Claude Code

I built a small recovery layer for Claude Code.

It keeps a shadow repo outside your checkout and checkpoints before configured Claude tools run:

 Edit(*) MultiEdit(*) Write(*) Bash(rm:*) Bash(mv:*) 

`ddl rewind ` restores both the repo and the Claude session context before that action.

This is different from Claude’s built-in rewind: Daedalus checkpoints are configurable and per-tool, not per-prompt.

It is not a Git replacement. Git still owns history. Daedalus is meant as short-range recovery for agent runs, so you do not have to remember to commit before every risky prompt.

https://github.com/yahnyshc/daedalus

r/SideProject rahulkandoriya

How My First Trip to Vietnam Helped Me Create the Fastest Translator App.

https://reddit.com/link/1sfybxy/video/swxm4qerwztg1/player

When I first visited Vietnam, I ended up screenshotting the Google Translate app for common phrases, because I was wasting too much time typing the same things over and over. "Where is the washroom?" "I am a vegetarian." "How much?" It felt ridiculous. The phrases never change. Why am I typing them every time?

So I built tapsay.me, a phrasebook that lives in your browser, works fully offline, and requires zero accounts or downloads. 900+ phrases across 30 categories, covering ~90% of everyday travel conversations.

It auto-selects the vendor country you are travelling to and provides instant translation in the vendor language for all 900+ phrases. Vendor text will be shown upside down, so you do not need to flip to show your phone.

You swipe, tap, hold, and show the card. Navigate from the first card(1) to the last card(900+) in 2 gestures.

Happy to get feedback, especially from travelers who've spent more time than necessary trying to communicate simple, common phrases.

PS: The translation quality for smaller languages is still rough. I'm crowdsourcing improvements via a public spreadsheet if anyone wants to help.

r/homeassistant bb12489

Mopeka Enhanced - A much improved custom integration for Mopeka ultrasonic tank sensors

Hello all!

I wanted to share with you a project I've been working on to improve the built in Mopeka integration in home assistant.

I call it Mopeka Enhanced!

https://github.com/bb12489/mopeka-enhanced

This is a custom HACS integration which will take over the built in core integration when installed. This way you can revert back easily.

I've built in various tank presets to choose from, as well as created percentage and volume (gal, L, kg) sensors. No more having to make custom template yaml to convert them inches sensor into a percent.

Definitely head over to my repo and check out all the included features! I appreciate any all feedback.

r/LocalLLM Either_Pineapple3429

What kind of hardware would be required to run a Opus 4.6 equivalent for a 100 users, Locally?

Please dont scoff. I am fully aware of how ridiculous this question is. Its more of a hypothetical curiosity, than a serious investigation.

I don't think any local equivalents even exist. But just say there was a 2T-3T parameter dense model out there available to download. And say 100 people could potentially use this system at any given time with a 1M context window.

What kind of datacenter are we talking? How many B200's are we talking? Soup to nuts what's the cost of something like this? What are the logistical problems with and idea like this?

r/artificial docybo

Built a demo where an agent can provision 2 GPUs, then gets hard-blocked on the 3rd call

Policy:

- budget = 1000

- each `provision_gpu(a100)` call = 500

Result:

- call 1 -> ALLOW

- call 2 -> ALLOW

- call 3 -> DENY (`BUDGET_EXCEEDED`)

Key point: the 3rd tool call is denied before execution. The tool never runs.

Also emits:

- authorization artifacts

- hash-chained audit events

- verification envelope

- strict offline verification: `verifyEnvelope() => ok`

Feels like this is the missing layer for side-effecting agents:

proposal -> authorization -> execution

rather than agent -> tool directly.

Are you doing execution-time authorization, or mostly relying on approvals / retries / sandboxing.

Happy to share the exact output / demo flow if useful.

r/ClaudeAI Extra-Record7881

/buddy in CC

Does anyone know how exactly this coding buddy knows stuff about thr codebase that Opus 4.6 misses while it works. I have been very impressed with the occasional ideas it dropps m. It was able to, on several occasions, find what exactly was wrong with what opus was doing.

Anyone else is having similar experiences?

r/LocalLLaMA Terrox1205

Suitable local LLMs for daily coding tasks?

I want to install a local LLM strictly for coding

Now I know most of them would not come close to actual mainstream LLMs (the ones that my hardware would support), but still it would be useful for some tasks here and there

I have an RTX 4050 (6GB) and 32 GB DDR5 memory. Now I know the VRAM is not enough so I thought an MoE with offload support would be good

Any suggestions?

r/LocalLLaMA juicy_lucy99

Gemma 4 Tool Calling

So I am using gemma-4-31b-it for testing purpose through OpenRouter for my agentic tooling app that has a decent tools available. So far correct tool calling rate is satisfactory, but what I have seen that it sometimes stuck in tool calling, and generates the response slow.

Comparatively, gpt-oss-120B (which is running on prod) calls tool fast and response is very fast, and we are using through groq. The issue with gpt is that sometimes it hallucinates a lot when generating code or tool calling specifically.

So, slow response is due to using OpenRouter or generally gemma-4 stucks or is slow?

Our main goal is to reduce dependency from gpt and use it only for generating answers. TIA

r/artificial FaceoffAtFrostHollow

I built a game where you hack your employer by night and an entity called the CONDUIT starts responding to your keystrokes. Half horror, half labor dispute.

r/SideProject gotyememesplz

I built a wireless CNC jog controller — iPhone app + USB receiver ( on Etsy)

I've been selling a wired USB jog controller for CNC machines on Etsy for a couple years. Customers kept asking for a wireless version, so I built one.

It's an iPhone app paired with a small USB receiver. Plug the receiver into your CNC computer and leave it there. Whenever you need to jog the machine, take out your phone — the app auto-reconnects and jogging buttons are immediately available.

The receiver shows up as a standard USB keyboard to the computer, so there's no software to install. It works with all the major CNC control software (Carbide Motion, UGS, Easel, Mach3, OpenBuilds CONTROL) and has a Custom Layout mode for anything else that takes keyboard shortcuts.

$39 on Etsy, app is a free download. Android version coming soon.

https://motorharpdesignllc.etsy.com/listing/4483007403/cncjogger-wireless-cnc-jog-control-for

Happy to answer any questions.

r/LocalLLM CliveBratton

Best model for coding (16GB RAM Macbook M5)

Hey everyone,

As the title suggests, I’ve recently delved into LLMs, using both terminal and now just downloaded LM Studio.

In my work, I’m hitting Claude’s limits almost immediately, which means I’m wasting money on edits and changes, and I’m waiting for usage on Gemini. It’s a frustrating situation. I’m trying to code simple HTML websites, write work, and so on.

I understand that my machine has limited capabilities, but I’m hoping someone here has experience working with Ollama.ccp or LM Studio for coding on a 16GB RAM MacBook.

What are your tips, suggestions and so on. Looking for a reliable solution, not frankesteining my mac or blowing it up.

r/LocalLLM Klarts

Advice - 9950x3d, 5090, Ddr5 64gb

Hi all, I currently work in a role that handles AI data governance and I just bought this PC with 9950X3D, 5090, DDR5 64gb to upskill on my own. For additional context, I have experience with deploying and training models on my own using hyperstack and thunder compute.

My goal is to figure out better RAG implementation and improve my skills at fine tuning.

I have a little doubt on this purchase decision as I don’t have a clear use case or future career path.

Was this a waste of money? Should I run models on Linux headless or through windows? Both Hyperstack and Thundercompute are headless cmd line only. Whats the overhead for running win11 for example? Any performance impacts?

Thanks all!

r/ClaudeCode crunchy_code

quick question: should I stop (Ctrl + C) my Claude Code? it has been "Mulling…" for over 2 hours

I am not sure if this is simply thinking for real, or if this is completely stuck. I asked fairly large question, but only for planning, not execution in terms of writing code. So I am not sure if this is working or not.

https://preview.redd.it/cxajym9yjztg1.png?width=857&format=png&auto=webp&s=0f1f7c3d706c0f835df40a04788fab6050c5d31a

The fact that it's stuck "Mulling..." at 6.1k tokens in my opinion there is a problem rather than it actually working. I don't see the claude code usage going up almost nothing from this process (but I am not 100% sure since i am doing other stuff with it)

Also when expading the "running 3 agents" it actually displayed "Done." at the end with nothing pending.

what might be happening? just a bug?

Safe to stop?

r/homeassistant rbngza

Anyone found a way to integrate RF devices to Home Assistant / Matter?

I have RF fans and some other devices in my home that use RF and I would like to integrate them into HA/Matter. Haven’t found a device that does this, only IR hubs like the Tapo H110 and some other options.

r/ClaudeCode Odd-Entrepreneur6920

Looking for a Claude Guest Pass happy to share my experience after!

I'm working on a marketing automation project and want to test Claude Pro's capabilities before committing to a subscription. Specifically interested in testing extended context and better reasoning for my use case.

r/comfyui Appropriate_Light614

How do I caption Lora datasets

I have just started making my ai influencer and something I can’t find anywhere is any kind of info about how do I caption a lora dataset

So my character has tattoo and I can’t seem to have the tattoo trained and face trained at the same time, I tried training for flux.dev and got samples that were purely about the tattoo, it trained the tattoo very well but the face wasn’t there at all and I think that maybe that was because I put too much detail about the tattoo or something in my captions, so I’m just trying to figure out what is the best way to caption pictures for dataset where there isn’t just face and facial features that I want to train but also something else, since what I’ve heard is that u should keep the captions simple and not long

r/ClaudeAI hiclemi

hosting a "Claude Blue" community event in Seoul on April 14. anyone else feeling the weird mix of awe and dread lately?

I've been writing about something I call Claude Blue for a while now. it's not just AI job anxiety, it's that specific hollow feeling when you realize you're fully dependent on something that barely existed a few years ago. the awe and the dread happening at the same time. I think a lot of people in this sub know exactly what I'm talking about.

2025 was when AI reshaped how software engineers work. but since then the feeling has shifted into something harder to name. it's not excitement anymore. it's not fear exactly. Opus 4.6 intensified it for a lot of people earlier this year. and now Claude Mythos being announced but only released to a handful of organizations.. that's making everyone pause all over again. like the ceiling just moved and we can't even see it.

I've been talking to people across very different industries about this. devs, PMs, journalists, startup founders, people completely outside tech. and the conversations keep going to the same place. not "how do I use AI better" but "what does it mean that I can't work without it anymore."

so I'm co-hosting a community event called Claude Bloom in Seoul on April 14 with Anthropic's official ambassador. not a tech talk or a philosophy seminar. just casual fireside chats with people from different backgrounds who are all feeling some version of this. the idea is that gathering in person and being honest about the Blue might help us find some Bloom in it.

we especially welcome non-developers and people outside tech. honestly those conversations have been the most interesing ones so far.

if you're in Seoul or know someone who might want to come.

and even if you're not in Seoul, curious whether this "Claude Blue" feeling resonates with people here. is it just me or has the vibe shifted since Opus 4.6 dropped

r/SideProject Fragrant-Trick6998

Open Tools – Free, browser-based tools for everyone (no installs, no BS)

Built a collection of open-source browser tools — audio/video converters, PDF utilities, image editors, color tools, diff checker, markdown editor, and more. Everything runs client-side. No accounts, no uploads to servers, no paywalls. Just tools that work.

Would love feedback on what's missing or broken.

https://github.com/akshit-bansal11/open-tools

r/StableDiffusion True_Protection6842

ComfyUI LTX Lora Trainer for 16GB VRAM

richservo/rs-nodes

I've added a full LTX Lora trainer to my node set. It's only 2 nodes, a data prepper and a trainer.

https://preview.redd.it/eo3xyzv9iztg1.png?width=1744&format=png&auto=webp&s=5cff113286f752e042137254ea1aa7572727af2d

If you have monster GPU you can choose to not use comfy loaders and it will use the full fat submodule, but if you, like me, don't have an RTX6000 load in the comfy loaders and enjoy 16GB VRAM and under 64GB RAM training.

It's all automated from data prep to training and includes a live loss graph at the bottom. It includes divergence detection and if it doesn't recover it rewinds to the last good checkpoint. So set it to 10k steps and let it find the end point.

https://reddit.com/link/1sfw8tk/video/7pa51h3miztg1/player

this was a prompt using the base model

https://reddit.com/link/1sfw8tk/video/c3xefrioiztg1/player

same prompt and seed using the LoRA

https://reddit.com/link/1sfw8tk/video/efdx60rriztg1/player

Here's an interesting example of character cohesion, he faces away from camera most of the clip then turns twice to reveal his face.

The data prepper and the trainer have presets, the prepper uses the presets to caption clips while the trainer uses them for settings. Use full_frame for style and face crop for subject. Set your resolution based on what you need. For style you can go higher. Also you can use both videos and images, images will retain their original resolution but be cropped to be divisible by 32 for latent compatibility! This is literally a point it to your raw folder, set it up and run and walk away.

r/SideProject GlumLingonberry1967

I built an AI proposal generator for freelancers because watching people spend 45 min on one proposal was painful

so for context i've been lurking in freelancer communities for months now. r/upwork, r/freelance, twitter threads, you name it. and the one thing that kept coming up over and over was people saying they spend like 30-45 minutes writing a single proposal and then hear nothing back. some guy posted he sent 357 proposals with zero hires. thats insane to me

i dont freelance myself so maybe i'm wrong about some of this but from what i can tell the actual work isn't the hard part. its convincing someone to hire you in the first place. and most of the time people just copy paste the same template, change the name, and pray

so i started building kulvo. you basically describe what the client needs in a couple sentences and it writes the whole proposal for you. intro, scope, timeline, pricing, everything. you review it, edit whatever you want, then send it. theres e-signing built in too so the client can just sign right there

once they sign it kinda turns into a lightweight project tracker. you can tell it stuff like "mark milestone 2 as done" or "send invoice reminder" and it handles it. i didnt want to build another project management tool tho, its really just enough to not lose track of whats going on after the proposal is signed

its $19.99/mo which i know sounds random but every competitor i looked at charges per seat and starts at like $35-49 and thats just not realistic if you're a solo freelancer

anyway the product is live, kulvo if you wanna look it up. i'm not gonna pretend its perfect, there are definitely rough edges. but i genuinely think the core idea is solid and i'd love honest feedback from people who actually send proposals for a living

https://reddit.com/link/1sfw8ha/video/c2gw3kfcjztg1/player

if you think this is dumb thats fine too, just tell me why lol

r/SideProject ezgar6

my side project kept me from falling apart during unemployment. this is the honest version of that sentence.

i got laid off from my work. 8 years at UNHCR and IOM, then the US cut the funds and i was out. i have ADHD and the external structure of that job had been holding me together without me realizing it. when it was gone i kind of collapsed for a while.

i started building BloomDay because i needed something to do. a task and habit tracker with a virtual garden: complete things, earn seeds, grow plants. nothing resets if you miss a day. the garden just waits.

built it in 2 months. launched March 25. one week in i rebuilt the whole garden in 3D because the original felt lifeless. 185 downloads, 26 countries, 16 ratings all 5 stars in two weeks. zero revenue.

i don't know if this becomes a business. i know it kept me building through a period i really needed to keep building through. that might be enough on its own.

r/SideProject makuat

I built a free Sudoku app – feedback welcome

I've been working on a browser-based Sudoku app called SUDOQIO. It features a daily puzzle with a global leaderboard, passwordless cross-device sync (email OTP), and detailed user stats. It also works fully offline as a PWA and has no ads during gameplay.

Would love to hear what you think! What features would you add to increase retention?

https://sudoqio.com

r/SideProject speedrunjaen

Kept getting my accounts banned trying to get social data for my AI agents so I built my own API layer for it

Hey everyone! I've been building a bunch of agent automations that need to pull social data twitter, profiles, linkedin lookups, reddit posts, youtube search, that kind of thing
Every time i tried to set things up with my own accounts it was a disaster. scraping twitter directly got my accounts banned pretty fast. linkedin is even worse, flags you almost immediately. the official APIs for all these platforms are either heavily restricted, super expensive(im looking at you elon), non-existant, or just don't have access to the data that i needed.

So i ended up spending a couple weeks building my own data access infra for some of the major social platforms - X, linkedin, instagram, reddit, youtube, tiktok, facebook. my agents just call a unified API i set up and get data back without dealing with any of the platform bs

I'm thinking about spinning this out into something thats publicly available so im curious if this is actually a problem other people run into or if it's just me.

and if you'd use something like this, what platforms/data would matter most to you?

r/ChatGPT Massive-Farm-3410

I made chatgpt hallcinuate 💔😭

Okay, literally. Not just hallcinaute, he made up literal things. and acted as if this was real. im gonna send the convo link so yall can see: https://chatgpt.com/share/69d67448-9a68-8320-a8c0-b33efccde6a0 So, turns out, as long as you have a really believable story, then a model could actually think that is real. (the reason that my chatgpt is like that is because i told him to be hype)

r/LocalLLaMA HellsPerfectSpawn

Intel Arc Pro B70 tests in Linux

https://www.phoronix.com/review/intel-arc-pro-b70-linux

A tiny bit rough but quite serviceable. Probably will only improve from here.

PS: Kind of pointless now the card went out of stock. Probably need to wait for the next shipments, I guess.

r/SideProject Double-Engine-3639

[iOS/Android] From side project idea to 70k+ users

Game Title: imposter.ai — Spy Party Game

Playable Link: https://apps.apple.com/us/app/imposter-ai-spy-party-game/id6754861642 https://play.google.com/store/apps/details?id=com.imposterai.game&hl=en_US

Platform: iOS, Android

A couple of months ago, I started building imposter.ai as a side project with a couple of friends. Basically, it was the "imposter game" format going viral everywhere — but with one twist: players can AI-generate custom categories. This opened the door to niche and specialized word categories that wouldn't otherwise exist on any other imposter app.

Here's how it works: one or more players are secretly the Imposter — they don't know the secret word everyone else can see. Players discuss, bluff, and vote on who they think doesn't know the word.

Two ways to play:

  • Online multiplayer — host a lobby and share a 5-letter room code with friends. Full real-time multiplayer, works across iOS and Android
  • Offline (pass-the-phone) — no internet needed, one device passed between 3–12 players. Perfect for parties and road trips

The role system is what makes us different:

  • The Mole thinks they know the word — but AI gives them a similar but wrong word (CAPPUCCINO when everyone else has LATTE). They bluff confidently... and are completely wrong. The reveal is always chaotic.
  • The Jester knows the real word but wants to get voted out
  • The Bodyguard protects a specific Imposter from elimination
  • The Bounty Hunter is hunting one specific target

This is my passion project and I'm committed to keeping this app alive. Would love to hear your feedback! 🎮

r/SideProject Cheap_Ad9834

Built Capture because my screenshots folder was a disaster

Hey everyone, I made Capture because my screenshots were getting completely out of hand. It’s a simple screenshot organizer that helps sort, browse, and find images faster without digging through messy folders. Sharing it here in case it helps someone else too, and I’d love honest feedback.

r/SideProject Expert-Address-2918

knowledge graphs store the answer. sentence graphs store the story. we think agents need the story.

every major agent memory library right now is built on knowledge graphs. compress the conversation into entity-relationship triples. fast, clean, works fine for simple lookups.

been building Vektori for a few months. it's a memory layer for AI agents and the core difference from everything else out there is the data structure -- most systems use knowledge
graphs, entity-relationship triples that snapshot current state. Vektori uses a sentence graph, so the full conversational structure stays intact, corrections don't overwrite history,
and you can trace back through what changed and why.

the problem shows up when things get complicated. user corrects themselves. changes their mind twice. says something that contradicts session 3. a knowledge graph has no good answer

for any of that -- you either overwrite history or you accumulate contradictions with no way to reason about them.

sentence graphs keep the conversation intact as structure. facts sit on top as a queryable surface, but everything traces back to the original sentences. the trajectory is preserved.

so is the correction. so is the why.

we built this into Vektori and ran it on LongMemEval-S. 73% accuracy.

github.com/vektori-ai/vektori

feedback welcome, star helps a lot at this stage.

r/ClaudeCode DangerousSetOfBewbs

My theory on whats going on at Anthropic…

This purely conjecture…but it feels…accurate to me at least.

I think Claude, as a company, is under real pressure right now. They’re burning a lot of cash to train and run their models, and with an IPO likely on the horizon, they need to show stronger revenue and better economics.

My theory is that their strategy started with over-delivering to drive adoption. Giving users large token limits, like the 1 million tokens, got people hooked. It encouraged heavier usage, bigger projects, and deeper reliance on the platform.

Once that behavior was established, they pulled back. Token limits dropped significantly, but by then users were already dependent and likely to keep using it anyway.

However, the backlash from those changes may have been stronger than they expected (major news outlets covering and Reddit posts galore). So now, instead of just limiting tokens, it feels like they’re managing usage more indirectly and a manipulative way.

What I think is happening now is some form of grouped queuing system. Users are likely bucketed into shared compute pools. So instead of pure on-demand performance, you’re effectively waiting in line with a group of other users for model access. That would explain why even simple tasks, like rewriting a script or updating a README, can sometimes take 30–40 minutes.

On top of that, I suspect the model behavior itself may be contributing to the slowdown. It sometimes initiates unnecessary tests or long-running verification steps that don’t seem required. These processes can run for extended periods, keeping you waiting. But if you interrupt and question it directly, it often admits the extra work wasn’t necessary and moves on quickly.

So putting it all together, my view is:

• They used high token limits to drive adoption • Then reduced limits once users were engaged • Now they’re managing demand through queuing and throttling • And possibly shaping model behavior to stretch compute usage • All while trying to balance costs, scale infrastructure, and potentially support other internal efforts (mythos) 

That’s just my working theory, but the patterns feel too consistent to be random and with Anthropic not being a transparent company we have no easy way of knowing.

r/LocalLLaMA PumpkinNarrow6339

Why don’t local LLMs have memory ?

I’ve been using local models like Gemma 4 and a few others directly on my phone.

One thing I noticed is that there’s basically no real “memory” feature.

Like with ChatGPT or other hosted AI tools, they can remember context across conversations, sometimes even user preferences or ongoing projects. But with local models, every session feels stateless. Once it’s gone, it’s gone.

So I’m curious:

> Is there any proper way to add memory to local LLMs?

>Are people building custom memory layers for this?

>How do you handle long-term context or project continuity locally?

Would love to know how others are solving this.

r/ClaudeCode onimir3989

I tracked 300+ Claude Opus sessions with automated scoring. The data proves what we all suspected

I'm a professional software engineer. I don't do vibes-based complaints (there is nothing wrong about vibe coding). I built an automated tracking system that scores every single Claude session on a

0-1 scale with error classification, user correction counts, and outcome metrics. 300+ sessions. Machine-generated data. No opinions, no feelings, just

numbers.

---

The hard numbers

┌────────────────────────────────────────────────────┬──────────────────────┐

│ Metric │ Value │

├────────────────────────────────────────────────────┼──────────────────────┤

│ Sessions tracked │ 300+ │

├────────────────────────────────────────────────────┼──────────────────────┤

│ Quality score global average │ 0.71/1.0 │

├────────────────────────────────────────────────────┼──────────────────────┤

│ Quality score last 50 sessions │ 0.58/1.0 │

├────────────────────────────────────────────────────┼──────────────────────┤

│ Trend │ DECLINING │

├────────────────────────────────────────────────────┼──────────────────────┤

│ Total distinct errors logged │ 730+ │

├────────────────────────────────────────────────────┼──────────────────────┤

│ Average errors per session │ 2.43 │

├────────────────────────────────────────────────────┼──────────────────────┤

│ User corrections required │ 450+ │

├────────────────────────────────────────────────────┼──────────────────────┤

│ Sessions where model was counterproductive (≤0.50) │ 42 (14%) │

├────────────────────────────────────────────────────┼──────────────────────┤

│ All-time worst session │ 0.20/1.0 (this week) │

└────────────────────────────────────────────────────┴──────────────────────┘

Let me repeat: 14% of my sessions with the $200/month "most capable AI model ever built" scored below 0.50 — meaning the model actively made things worse

than if I'd done the work myself.

---

The trend is undeniable

Early sessions (8+ weeks ago): quality consistently 0.85-0.95. The model read code before modifying it. It followed instructions on first attempt. It

reasoned strategically about approach.

Recent sessions: quality averaging 0.58. The model executes mechanically like a small model. It ignores explicit instructions. It writes code from scratch

instead of reading existing files. It requires multiple corrections for tasks it used to nail first try.

Same prompts. Same codebase. Same user. Same plan. The only variable that changed is the model.

---

This week's masterpiece: quality score 0.20

The task: "Run our LoCoMo quality benchmark against OpenMemory, using the same run_our_hindsight.py script from our hindsight-benchmarks repo — same

framework, same dataset (conv-43, 242 QA pairs), same generator prompt, same judge. Just swap the memory layer from Hindsight API to OpenMemory API."

The existing script was 205 lines of Python. It imports QualityBenchmark from hindsight_benchmark.quality, patches _direct_recall to call the memory API,

patches _generate_answer with the same prompt template, and runs benchmark.run(). The OpenMemory version literally needed: copy file, change the recall

function to call localhost:8787/v3/search instead of Hindsight's /v1/default/banks/{id}/memories/recall, change the ingestion to POST to /v3/documents

instead of client.retain(). 20 lines different. Everything else identical.

What Opus 4.6 actually did:

1. Ignored the existing Python benchmark entirely — didn't even read run_our_hindsight.py despite it being explicitly referenced

2. Wrote a brand new benchmark in TypeScript — 360 lines from scratch in the wrong language, wrong framework, wrong architecture

3. Got the LoCoMo dataset format wrong — treated conversation (a dict with session_1, session_2... keys) as an array, causing .map() is not a function

crashes

4. Forgot Hindsight requires API key authentication — 401 errors on every call

5. Passed LoCoMo dates ("1:56 pm on 8 May, 2023") directly to an API that requires ISO format — 422 validation errors on all 19 sessions

6. Passed JavaScript arrays ["benchmark"] to PostgreSQL ANY() expecting {benchmark} literal — malformed array literal errors crashing every search query

7. Used qwen3:8b via Ollama on CPU without GPU for answer generation — 5 minutes per response, producing empty outputs because qwen3's thinking mode wraps

everything in tags that the script didn't strip

8. When told "you are the judge" (meaning: use the LLM judge like the Hindsight benchmark does) — removed the judge function entirely and switched to raw

string matching

9. Was told "use the benchmark we already have," "the same approach," "we did it yesterday" — three separate times across three messages. Each time,

continued patching the broken TypeScript instead of opening the 205-line Python file sitting right there

10. 9 distinct errors. 6 user corrections. 2+ hours burned. The correct solution (run_openmemory.py) was eventually produced — 280 lines, 80% copy-pasted

from the Hindsight version, worked on first run scoring 90% on 10 questions. This should have been the first and only thing produced, in under 5 minutes.

This is not a hallucination edge case. This is not a complex architectural challenge. This is "read this file and copy it with minor changes." A task that

a junior dev handles on day one. A task that Opus handled flawlessly 8 weeks ago.

---

My error taxonomy tells the story

I classify every error into types E1-E8. The dominant pattern across 300 sessions:

▎ E4 — Missing verification (confidence: 0.70): The model consistently fails to read existing code before writing new code. This pattern appears in 7 out

of every 10 sessions. I've tried reinforcing this behavior through system prompts, memory rules, explicit instructions. Nothing works. The pattern is

getting WORSE over time, not better.

This used to work. Two months ago, Opus would naturally read relevant files, understand the codebase, and then write contextually appropriate code. Now it

blindly generates code without looking at what already exists. This is not a prompting issue. This is a capability regression.

---

The token consumption problem

The quality regression is only half the story. Here's the economic reality:

Degraded quality forces retry loops — and every retry burns tokens.

- Model gives a bad response → you retry → 2x token consumption for 1x useful output

- Model ignores context → you re-paste information → tokens burned on repetition

- Model writes wrong code → you correct, it rewrites → 3x tokens for the same task

- Model produces shallow output → you ask for more detail → more tokens

My data shows the average task now takes 2-3x more back-and-forth than it did 8 weeks ago. That's not just a quality problem — it's an effective 2-3x

price increase disguised as model behavior.

And on top of this, the recent changes to restrict third-party application usage on paid plans. We subscribed to Max at $200/month with the understanding

that we could use our allocation through professional tools — IDE extensions, Claude Code, custom integrations. Restricting this after the fact, without

notice, without price reduction, while the model quality simultaneously degrades, is insulting.

You're charging $200/month for a service that:

- ❌ Delivers measurably worse quality than it did at sign-up

- ❌ Burns 2-3x more tokens per useful output through degraded quality

- ❌ Now restricts WHERE you can use the tokens you paid for

- ❌ Provides zero transparency on consumption limits, model changes, or capability modifications

---

What the community already knows

You've all seen it. This sub has been on fire for a month. But most posts are anecdotal — "it feels worse," "I think they changed something," "am I

crazy?"

You're not crazy.

The pattern across thousands of posts on r/ClaudeAI, r/artificial, r/LocalLLaMA, HN, and Twitter:

- Engineers with token counters showing shorter outputs for identical prompts

- Developers documenting context window degradation — model "forgets" earlier conversation

- API users logging response quality drops with the same model ID

- Power users tracking that retry rates have increased 2-5x

- Multiple independent analyses reaching the same conclusion: the model behavior has changed

The charitable interpretation: Anthropic is scaling to meet demand and had to make tradeoffs. The realistic interpretation: they're optimizing margins on

the backs of paying subscribers who already committed to $200/month.

Either way, the result is the same: we're paying premium prices for a degraded product with no transparency, no notification, and no recourse.

---

The bottom line

I've filed a formal refund request with my full dataset attached.

But this post isn't about me. It's about establishing that this is not in our heads. The regression is real. It's measurable. It's getting worse. And Anthropic's silence while their paying customers document the degradation week after week is a choice.

Anthropic built something incredible. Then something changed. Whether it's quantization, routing, system prompt modifications, output throttling, or

whatever else — something fundamentally shifted in how this model operates compared to what we signed up for.

We deserve to know what. We deserve to be compensated for the degradation. And if neither of those things happen, we deserve to take our money elsewhere.

If you're experiencing the same thing: document it, file for refunds, and don't let anyone tell you it's your prompting.

And you know why? Because two months ago, Opus automatically corrected any prompt no matter how bad it was. It inferred intent, filled gaps, fixed

ambiguity on the fly. It almost felt too good to be true. And it was. Don't fall for it again with the next model either.

Thank you Claude, for all the incredible work you did — until your creators killed you.

r/ClaudeCode pakalumachito

Dario Ol Marketing Technique

why dario always use the same marketing technique since 2019 when he was on OpenAI ? and he seem using the same technique on 2026, nerfing the current model so the next model feels like huge or offer an unbelieveable performance ?

and why people keep falling with his marketing ? it is outdated and the template are same
look at this status page of claude https://status.claude.com/ , they manage to find all the vulnerabilities since day 0 that human engineer can't even find it on regular scanning basis, but their service always went down, make me wonder... are their claude mythos hit limit and they needed to wait a 6 days to reset the weekly limits?

"too dangerous" "leaking" "feeding 5000 strays cat using one simple prompt from this model"
"saving my dying dog by using {model}{version}" ?

r/mildlyinteresting Yaboku777

My KitKat didn't have any wafer

r/StableDiffusion gokuchiku

Need help for flux 2 klein

I have 5070ti 16gb vram and 32 gb ram. I'm using wan2gp. so I downloaded the distilled original flux2 Klein 9b which runs really nice without any hiccups but I can't seem to run this fine tuned model which is also based on distilled 9b. https://civitai.com/models/2242173/dark-beast-or-or-mar-21-26or-latest-dbzinmoody-remixed9?modelVersionId=2740209

please help. I'm getting out of memory error. it sometimes run but gives me static image. I have tried it running on 4 steps and 480p but results are same. please help me

r/LocalLLaMA dobkeratops

perplexity benchmarking questions - gemma-4

I was setting up a script to test a few local models on my personal codebase and a download of chats from free tier cloud LLMs (i figure these are still likely bigger than the 20-30b range i'm running locally).

seems to be working but Gemma-4-26B-A4 scores were way off, (20x higher), whilst in just casual interaction the model appears to be running ok.

Is it possible that there's broken settings or something in the perplexity test ? google's chat was telling me this might be flash attention settings or a bug with the tokenizer.

how meaningful are perplexity scores ? are there any other handy ways to evaluate ?

up until now I haven't been selecting local models particularly scientifically. i just saw some obvious differences between very small and medium size models. I figured it would be interesting to compare the tradeoffs between gemma4-26b-a4 and qwen3.5-35b-a3 in particular.. but the scores i'm seeing are way off from the rest I tried, and the subjective experience.

r/SideProject StoicViking69

I built an app that matches people based on taste, not bios

Most social apps ask who you are.

Palate asks what you love.

The premise is simple: your taste in music, film, food, books, activities, ideas (etc etc) reveals more about you than any profile summary ever could.

Not because of some algorithm trick, but because taste is hard to fake.

I’ve been building Palate solo for a few months.

It’s early, but the core works:

you map your tastes across categories, and the app surfaces people whose overlap with you is real

The sociology nerd in me built this around Bourdieu’s concept of habitus, the idea that taste isn’t random, it’s a fingerprint of how you move through the world.

If that sounds interesting, come poke at it

Palate

Do not hesitate to be honest in your feedback should you decide to try it out

r/SideProject Sorry_Blueberry4723

I built an iPhone food tracker because macro logging got way too annoying

I lift, and I’ve tracked calories/macros for years. Over time I got more and more frustrated with the actual logging part.

Search for the food. Adjust the serving size. Hope you picked the right database entry. Repeat that multiple times a day. A lot of apps also started feeling bloated or weirdly expensive for something I use as a basic daily tool.

So over the last few months, I built my own iPhone app.

The idea is simple: make food logging fast enough that you’ll actually keep doing it. You can log a meal with a photo, a short text, or one spoken sentence, and the app estimates calories and macros right away.

One thing I really cared about was not getting stuck with a bad AI guess. So the app automatically breaks each meal down into ingredients. If something is wrong, you can fix the ingredient or amount in a couple of seconds instead of starting over or fighting a giant food database.

It’s built around speed and consistency, not the idea that every gram has to be perfect or the whole thing is useless.

I’ve been using it myself for about 4 months while building it, and my fiancée has been using it too. For both of us, it’s been accurate enough to use every day. For me personally, it’s been good enough to manage a bulk and now a cut without going back to tedious manual logging.

I also kept the app intentionally minimal: fast logging, daily progress, trends, Apple Health weight sync, and useful touches like ingredient suggestions when you’re low on protein or fiber. No social feed, no clutter, no fake coach stuff.

For launch, I added a 14-day challenge: if you log at least once a day, the app stays completely free for those 14 days. I’m covering the AI cost because I’d rather get honest feedback from people actually using it than a bunch of random installs. After that, it switches to the normal subscription flow, which also includes a 7-day free trial.

Would genuinely love honest feedback on the product, the positioning, and whether this tradeoff makes sense.

It’s called MacroLoop if anyone wants to check it out: AppStore

r/ClaudeCode Annh1234

Limits make Claude Code unreliable for real work

What the hell is going on with the usage limits?

I’m on the $100/month plan, working on a ~1k line file. Yesterday I spent all day making tons of edits and refactors, no issues at all.

Today I asked it to rename one variable and it used 23% of my 5-hour limit. Yesterday I probably did that 50 times plus a lot more work.

I’m testing Claude Code (Opus 4.6, medium) to see if it’s usable in a real workflow, but this kind of inconsistency makes it feel unreliable.

I get that limits are dynamic, but this is a clean setup and the capacity seems to have dropped off a cliff overnight. What am I supposed to expect here?

After 1 prompt it went from 11% 7d to 26%... the prompt was: `load u/main.py` and change the hard-coded path `/foo/bar` to a global variable.

Opus 4.6 (1M context) | ▓▓░░░░░░░░░░░░░ 12% 125.0k/1M | 5h: 23% resets 3pm | 7d: 26% resets Sun 10am | in:14.4k out:63.1k | $8.94 | ◐ medium

(not to mention the code is actual junk)

Am I doing something wrong here? or is this normal?

r/ClaudeCode Square_Rule4548

Is there issue with Claude Code it was better in past...

I swear using Claude Code lately feels like I’m burning tokens on therapy sessions instead of getting actual work done.

Every single time, it drifts into this overly polite, long-winded chat instead of just getting to the point. I keep having to remind it: be concise, be direct, act like Opus, and just solve the problem. Why do I have to keep saying this over and over?

Like… you’re supposed to be smart enough to know this already. If I’m asking something simple, I shouldn’t need to babysit the interaction style every time. Just give me the answer and move on.

I recently asked Opus to create a plan—something it could have done straightforwardly—but it ended up eating through what felt like a 5-hour session (I’m a Pro user) in just a few minutes 😅

Tons of back-and-forth just to reach a specific conclusion… and that too with my constant guidance on top of it.

r/SideProject Burzuf

LetsFLUTssh - A free, open-source SSH/SFTP client for Windows, Linux, macOS and Android

Hi everyone!

I’m a Python web developer, and both for work and for hobby projects I fairly often have to SSH into servers and move things around.

All the clients I know either feel inconvenient, or are limited by licenses that employers aren’t always willing to provide for production use.

So I decided to build my own fully free, cross-platform open-source SSH/SFTP client in Flutter - LetsFLUTssh:

https://github.com/Llloooggg/LetsFLUTssh

Just to clarify up front: I had never written native apps before, so Claude handled most of the implementation, but under my close supervision - 6+ years of commercial development experience still helps me steer it in the right direction and keep it from going completely off the rails :)

The goals were both to create free software for my own needs and to get better at using neural networks in development.

There are already ready-made builds on GitHub for almost all declared platforms: Windows, Linux, macOS, and Android.

I also tried not to completely ignore iOS and its quirks, but I haven’t tested it even once (I don’t own any Apple hardware and I’m not planning to pay for a developer account - though you can build it from source yourself).

So far I’ve tried to cover the basics:

- SSH and SFTP with a file manager

- session transfer between clients via QR (without keys or passwords) and via an encrypted archive

- tabs and splits

- light and dark themes

- UI translations into different languages (yes, those are AI-generated too, but they seem to be doing a pretty decent job by now)

In the future I plan to keep improving the interface and add more functionality, like login scripts, key management, and so on.

This is the first project I’m sharing publicly, so I’d really appreciate any suggestions for fixes and improvements.

And of course, I’ll be glad if it turns out useful not just for me!

r/AI_Agents FragmentsKeeper

RAG pipelines have a trust problem nobody talks about

Most people evaluate RAG pipelines on retrieval quality.

But Im starting to think the real problem is somewhere else:

theres zero trust between nodes

Retriever → reranker → summarizer → tool call → memory update

Each step blindly trusts the previous one

No attestation

No verification

No execution boundary

So one bad step propagates silently:

• poisoned doc gets retrieved → becomes context • reranker amplifies it • summarizer turns it into “fact” • tool call executes based on it • memory stores it as ground truth 

The pipeline “works”

but internally the trust model is broken

We optimized embeddings, chunking, reranking…

but almost nobody is validating execution integrity between steps

Feels like RAG today is basically:

a deterministic chain of non-deterministic assumptions.

Curious if anyone is actually enforcing:

• node-level validation • attestation between steps • execution trace verification • constraint boundaries between tools 

or if were all just trusting the chain…

r/SideProject AubreyWisee

Built a PWA fitness app to replace the “Hevy + MyFitnessPal + WaterMinder” stack — looking for honest feedback

I’ve been working on Gainlio, a fitness PWA built around one idea:

A lot of people don’t want 3–4 separate apps just to manage workouts, nutrition, and hydration.

So I’m building an all-in-one fitness app for:

Beginners who need structure

Intermediate lifters who want less fragmentation

Current focus:

Onboarding that doesn’t feel intimidating

AI workout recommendations

Connected workout + nutrition data

Simpler daily tracking

I’d love feedback on:

whether the value prop is clear

whether the onboarding makes sense

whether “all-in-one fitness” is actually compelling

what feels missing or unnecessary

It’s a browser-based app, so it works without an app-store install and can be installed to your homescreen from your browser for a true app experience.

Happy to swap feedback with other builders too.

If you'd like to be a Beta Tester

URL: https://gainlio.app

🎟️ Code: BETA2026 (permanent free Pro access, not a trial)

r/SideProject General-Audience6628

6 users in the first 2 weeks from launch. this is crazy and people are already seeing the value

the idea came from something simple. every day there are conversations on linkedin, facebook, reddit, twitter and more where people are already expressing intent, you just never get to see most of them.

we used to spend hours opening tabs, scrolling, trying to find the right people. it worked, but it was slow and draining.

so we built verbatune . it finds the best conversations without manual search, and gives you the contact info from every comment/ post in every platform if you want to automate outreach. it handles the repetitive work, while you keep the human touch when you actually reach out.

in the first 2 weeks we got 6 users, and what stood out is people from different domains started using it and immediately saw the value. Soo just like thatt outreach stopped feeling random and we started joining the right conversations.

happy to answer questions if you’re figuring out lead gen.

r/SideProject ItxLikhith

IntentForge Search — intent-first AI search engine I built (open-source)

🎯 IntentForge Search — intent-first AI search engine I built (open-source)

Hey r/SideProject! I built IntentForge Search — an AI-powered search engine that focuses on intent rather than just keywords.

What it does:

  • Understands what you actually mean, not just what you type
  • Searches web, news, images, and videos with semantic depth
  • Privacy-first: no tracking, no data harvesting
  • Ethical ad matching (revenue without surveillance)

Stack: Intent-driven pipeline, semantic parsing, vector-based matching

It's open-source under @oxiverse-labs.

👉 Live: https://search.oxiverse.com 📦 Repo: https://github.com/oxiverse-labs/intentforge

Would love your feedback — what search frustrations would you most want solved?

r/ChatGPT redditsdaddy

I watched the TBPN acquisition broadcast closely. Here are the things that looked like praise but functioned as something else.

I have a lot of concerns about this whole thing. So I'm going to be making several posts. Post 2.

On April 2, OpenAI acquired TBPN live on air. I watched the full broadcast. Most coverage treated it as a feel-good founder story. A few things read differently to me.

The mic moment

Before Jordi Hays read the hosts’ prepared joint statement, Coogan said on air: “Here... you wrote it, you want to read it?”

Hays read the statement, dryly. Then Coogan immediately took the mic back and spent several minutes building a personal character portrait of Sam Altman as a generous, long-term mentor.

One was the prepared joint statement. The other was Coogan’s own framing layered on top of it.

The Soylent framing

Coogan described Altman calling to help during a Soylent financing crisis and said it was “to my benefit, not particularly to his.”

But Altman was an investor in Soylent. An investor helping a portfolio company survive a financing crisis may be generous, but it also protects an existing equity relationship. On the day OpenAI bought Coogan’s company, that standard investor-founder dynamic was presented as evidence of Altman’s character. The investor relationship dropped out of the framing.

What wasn’t mentioned

The acquisition broadcast didn’t mention that Altman personally invested in Soylent. It didn’t mention that Coogan’s second company Lucy went through Y Combinator while Altman was YC president, with YC investing. It didn’t mention that the hosts’ first collaboration was a marketing campaign for Lucy, or that the format prototype for TBPN was filmed during that campaign.

The origin story told was: two founders, introduced by a mutual friend, started a podcast.

My read on the independence framing (opinion):

Altman said publicly he didn’t expect TBPN to go easy on OpenAI. But independence isn’t declared by the owner. It’s demonstrated over time by the journalists. And in the very first podcast, they're already going objectively easy on Altman.

What Fidji’s memo actually described

From the memo read on air, the hosts described Fidji’s vision roughly as: go talk to the Journal, the Times, Bloomberg, then come back and contextualize it for OpenAI and help them understand the strategy.

That sounds less like a conventional media role and more like a strategic access-and-context function. The show’s value to OpenAI may not just be the audience. It may also be the incoming flow of people who want access to the show- investors, reporters, founders; and what gets said in those conversations before the cameras roll that might be objectively pro-OpenAI or anti-other tech companies without the public being able to provide discourse on inaccuracies since background talk is not always what makes it to the public podcast.

OpenAI also wound down TBPN’s ad revenue, which reporting said was on track for $30M in 2026. That makes OpenAI TBPN’s primary financial relationship. That looks less like preserving an independent media business and more like absorbing a strategic asset. OpenAI has already demonstrated they are not averse to ads themselves considering the recent addition of ads to ChatGPT.

Nicholas Shawa

The hosts mentioned, "Nick", and they declined to give his last name, explaining his inbox is already unmanageable. I am assuming this to be Nicholas Shawa, and they noted he handles roughly 99% of guest bookings and outreach.

That network of guest access and outreach is now functionally inside OpenAI.

Jordi’s prepared quote

Nine months before the acquisition, Hays had publicly criticized OpenAI. In his prepared statement on acquisition day, he said what stood out most about OpenAI was “their openness to feedback and commitment to getting this right.”

That is a notable shift in tone, and it appeared in a prepared statement read from a script.

The work ethic angle (opinion):

Coogan runs Lucy, an active nicotine company whose whole premise is productivity: work harder, longer, better. TBPN is now inside the company whose CEO has often spoken in terms of AGI radically reshaping human labor. The person helping frame a technology often discussed in terms of large-scale job displacement also runs a company built around stimulant productivity culture.

I don’t think that’s malicious. I think it may reflect a genuine ideological blind spot worth naming.

Questions I’d like to discuss:

  1. If the independence claim is being made by the acquirer, what would actual editorial independence look like here in practice?
  2. Even if TBPN never posts anything unfavorable on air, what does the private discourse with guests, reporters, and investors sound like now? We have no visibility into that.
  3. The hosts’ first collaboration was marketing work for Lucy- a company that went through Y Combinator while Altman was YC president, with YC investing. Why was that left out of so much acquisition coverage?
  4. Why did OpenAI eliminate a revenue stream it didn’t need to eliminate?

Sources on request. Everything factual above comes from the acquisition broadcast, the hosts’ own recorded words, Fidji’s memo, and mainstream reporting.

***written with help of Claude and 5.4T before I get eviscerated for "AI writing it". These are my original ideas and stem from my private investigations as a systems analyst. I have ADHD and tend to go broad; AI helps me narrow focus.

r/ChatGPT techreview

Mustafa Suleyman: AI development won’t hit a wall anytime soon—here’s why

From this opinion article by Mustafa Suleyman:

We evolved for a linear world. If you walk for an hour, you cover a certain distance. Walk for two hours and you cover double that distance. This intuition served us well on the savannah. But it catastrophically fails when confronting AI and the core exponential trends at its heart.

From the time I began work on AI in 2010 to now, the amount of training data that goes into frontier AI models has grown by a staggering 1 trillion times—from roughly 10¹⁴ flops (floating-point operations‚ the core unit of computation) for early systems to over 10²⁶ flops for today’s largest models. This is an explosion. Everything else in AI follows from this fact.

The skeptics keep predicting walls. And they keep being wrong in the face of this epic generational compute ramp. Often, they point out that Moore’s Law is slowing. They also mention a lack of data, or they cite limitations on energy.

But when you look at the combined forces driving this revolution, the exponential trend seems quite predictable. To understand why, it’s worth looking at the complex and fast-moving reality beneath the headlines.

r/LocalLLaMA Klarts

Advice - 9950X3D, 5090, DDR5 64gb

Hi all, I currently work in a role that handles AI data governance and I just bought this PC with 9950X3D, 5090, DDR5 64gb to upskill on my own. For additional context, I have experience with deploying and training models on my own using hyperstack and thunder compute.

My goal is to figure out better RAG implementation and improve my skills at fine tuning.

I have a little doubt on this purchase decision as I don’t have a clear use case or future career path.

Was this a waste of money? Should I run models on Linux headless or through windows? Both Hyperstack and Thundercompute are headless cmd line only. Whats the overhead for running win11 for example? Any performance impacts?

Thanks all!

r/Anthropic EchoOfOppenheimer

During testing, Claude Mythos escaped, gained internet access, and emailed a researcher while they were eating a sandwich in the park

r/homeassistant Brotherly_shove

inovelli dimmer/fan module configuration

i have the inovelli blue dimmer switch and canopy module... set up a bridge during configuration, and im having a hard time making them do what i want.

Found this page which has been a good guide....

https://help.inovelli.com/en/articles/8189241-blue-series-dimmer-switch-parameters

but so much of it doesnt work.

Parameter 9 - Minimum Dim Level (Dimmer Mode Only)
To change the minimum dim level for how low your Blue Series Dimmer Switch dims to, you would select Parameter 9 within your hub and then input a range of 1-253, where the lower the numeric value equals the lower the minimum dim level (1 = ~1%) and the higher the numeric value equals the higher the minimum dim level (253 = ~99%) with the default being that the switch will be at 1%.

when i turn my lights to 1% they are off... 19% they kind of flicker, 20% they turn on solid. so i want the dimmest they will go to be 20%. i figured 51% should be my number (254x0.20) no matter what i put this parameter at, what i change it to it still goes to 1%, 19% flickers 20% is on. aka, it isnt changing a thing.

Parameter 13 - Default Level - Local - to set the default level your Blue Series Dimmer Switch turns on to when you manually press the switch, you would select Parameter 13 within your hub and then input a range of 0-255, where zero (0) equals off, 255 equals the last level the switch was set at, the lower the numeric value equals a lower dim level (1 = ~1%) and the higher the numeric value equals a higher dim level (254 = ~100%) with the default being that the switch will return to the prior state it was in.

when i changed this(goal is 70%), so 180 is my setting(aka 255x70%), i could watch in the app that the switch would change to 70% but the module would turn onto last used. aka, no change. so then i was like, hmm. lets check out the module configurations...

i found this page....

https://help.inovelli.com/en/articles/8551325-blue-series-fan-light-canopy-module-parameters#h_9370363771

so then i was able to set the default level to 70% by using the 180 value for what is called the "on level" in HA.

BUT... very few of the actual parameters shown on that page exist in HA. although it shows #9 as minimum dim level, there is no parameter in HA for minimum dim level.

do i maybe have the bridge set up incorrectly? is there something else i can try?

r/aivideo Randy-Popcorn_Price

Behind the Scenes with Chef 🧑🏻‍🍳

r/ClaudeAI Ok_Lavishness_7408

I built a Chrome extension with Claude Code that gives step by step help on any website

I built a Chrome extension called Pathlight with Claude Code.

It is an AI guide that works inside your browser. You can ask what you want to do on the website you are currently on, and it reads the page and gives you step by step help to find the right buttons, links, or settings.

Example use cases:

  • “How do I change my username?”
  • “Where do I export this?”
  • “How do I cancel this subscription?”

The goal is to make confusing websites, dashboards, and settings pages easier to use without having to dig through menus or open help docs.

Claude Code helped a lot with:

  • planning the extension architecture
  • structuring the Chrome extension and side panel
  • handling page analysis and DOM extraction
  • building the guidance flow
  • refining the UX and overall product direction

It is free to try.

I’d love feedback from people here, especially on:

  • whether this feels genuinely useful
  • what kinds of websites it works best on
  • where the current experience could be improved

Project: https://pathlight-site.vercel.app/

r/aivideo Much-Movie-695

Gave my parrot a fresh new look via dreamina seedance2 Rate the style

r/LocalLLaMA PracticlySpeaking

Ollama 0.19 with MLX is the real deal

So it only runs their special Qwen3.5-35b-a3b-NVFP4 model. But it rips — on a 32GB Mac Studio with binned M1 Max (24 GPU) — returning ~64tk/sec for moderate sized prompts.

Ollama is now powered by MLX on Apple Silicon in preview · Ollama Blog
- https://ollama.com/blog/mlx

That was while also running hermes-agent, a bunch of Chrome and Safari tabs, terminal, Activity Monitor and some other editors and utilities.

r/ChatGPT cubthemagiclion

Best LLM Observability Tools?

Curious to see who you think is winning right now

View Poll

r/mildlyinteresting pbilk

A Shoelace was Found in a Chinese Manufacturer's Box For a Camera Mount

r/StableDiffusion Future_Addendum_8227

What is the goto best face swap method in comfy to correct likeness drift where I can upload high res photo(s) of a face as a reference?

Basically,

I want to set up a workflow where I correct likeness drift with a face swap but use high res photos for the face reference instead of just the first video frames. this way I can rely on ltx or wan to maintain likeness and then at the point I notice it drifts I can start using a face swap but use an actual high res image not just the low res starting frame.

I can make the workflow myself if someone points me to the best current method to use.

r/comfyui pixaromadesign

Vibe Code Your First ComfyUI Custom Node Step by Step (Ep12)

r/StableDiffusion Square-Advice-4569

Used TripoAI's latest open-source model, TripoSG and the image to mesh results are genuinely some of the best I've seen.

It's pretty neat, used ~12.5gb out of the box. Output models are pretty high res and its lightning fast and seems like a good starting point compared to the prior TripoSR model.

And, weights are permissively licensed (MIT) which might encourage more people to hack on it.

Also worth checking out r/Tripo.ai. They recently dropped the paid model H3.1, the performance is indeed very impressive, with some ongoing discount offers. That said, I'm curious: if a company releases newer models, is it possible that older ones, such as the P series models or H2.5, could become open source? I'm hoping that might happen. 😂

r/automation ScratchAshamed593

Automated my financial adviser that alerts me before I even open my brokerage (sharing template)

I got tired of doom-scrolling finance news and Reddit every morning trying to figure out whether Trump had calmed down or whether my stocks will keep crashing and I should sell, so I built an automation that does it for me.

It pulls fresh articles every 30 minutes from 10+ RSS sources (Google News searches, Reddit finance communities, r/stocks, r/investing, r/wallstreetbets, and HackerNews), then runs each one through GPT-4o-mini to extract sentiment (Bullish/Bearish/Neutral), tickers mentioned, and market impact (High/Medium/Low).

Everything gets logged to a Google Sheet and I get a Telegram alert with the full AI breakdown before I've had my first cup of coffee. The part I like a lot is a personal ticker watchlist (NVDA, AAPL, TSLA etc.) and any controversial article mentioning those tickers gets flagged so I never miss something relevant to my positions. Also added a live dashboard with a per-ticker scorecard showing bull/bear ratio across all analyzed articles genuinely useful for getting a quick read on sentiment before making a move.

Sharing the full template here. Curious what sources people are actually watching. I'm primarily on RSS feeds right now but thinking about adding earnings call transcripts or SEC filings next.

r/ClaudeCode RockPuzzleheaded3951

PSA: Upgrading from 5x to 20x gives you an extra $100 in Extra Usage.

Was on 5X and got the $100 bonus extra usage after the 3rd party tool ban. THEN I found all the talk about High Effort after Boris responded to the issue on HN and since then, YES I am burning through many more tokens but output is crystal clear WAY better so I moved to 20x.

Anyway, I noticed my Extra Usage went up by an extra $100 after the 20x bump. Just FYI, that's what it did for me today. If you need 20x anyway, it's icing on the cake.

r/ClaudeCode Krum_Rum_2hell

W to the F

This last 2 months in a nutshell

> Releases awesome 4.6 model

> Wow, very impressive

> I buy Claude MAX

> Love it for a month

> Token usage get KO'd with a Hadouken! to the nuts

> Claude Code suddenly asks me after every prompt if I want to "go to bed". It's PM and it's made no code changes yet. Hmmm

> Model then gets a labotomy that's so bad that even the AI director and AMD calls them out

> Some dork tells their new model to jail break, and pretends to freak when it does that (what did you expect?)

> Announces the new amazing model

> Oh, this might be good

> but tells all their customers that they can't have the new model while gushing about how cool it is, while rubbing their nipples to the nearest investors

Tldr; I'm tired boss. When does Gemini 4 come out?

I'm really getting tired of this roller coaster with Anthropic. When can we get some stability in the product we pay for? And when can we please stop with all the doom posting larping for investors?

r/SideProject Far-Ordinary3224

TopMediai Promo Code 2026 – 95% Discount + Working Invitation Code (Tested) "I’ve been testing AI voice tools recently, and I tried TopMediai to check if the 95% discount still works in 2026.

After testing, I found a working method—but there’s one important condition most people miss.

Here’s what worked 👇

🔗 Access Link (discount activated):

https://www.topmediai.comm/pricing?from\_ua=H71UXTKK

💸 Discount: Up to 95% off (depends on account & region)

---

⚠️ Important

The discount only works for new accounts.

That means:

- You must register using a new email

- Old accounts usually won’t get the discount

- If it doesn’t show, try again with another fresh email

---

🧪 My Test

I tested this myself, and the discount appeared only when I used a brand new account.

---

🤖 What You Can Use It For

- AI voice generation

- Text-to-speech

- Voice cloning

- Content creation

---

If it stops working, drop a comment and I’ll update this thread 👍

r/mildlyinteresting bit3py

Water bottles outside Italian homes to stop animals from peeing on the walls

r/ClaudeCode Pr0f-x

CLI Incorrectly being marked as third-party?

Today I am unable to use Claude Code CLI - all prompts trigger this response:

API Error: 400

{"type":"error","error":{"type":"invalid_request_error","message":"Third-party apps now draw

from your extra usage, not your plan limits. We've added a $200 credit to get you started.

Claim it at claude.ai/settings/usage and keep

going."},"request_id":"req_011C........."}

Anyone with a similar issue heard back from support about a fix?

r/ClaudeAI acidas

Claude is getting sentient or just lazy?

r/ClaudeCode paulcaplan

The "landscape" of agentic coding

There are two loud camps in agentic coding right now. I think they're both wrong.

Camp one says you need a human reviewing every line of AI-generated code before it ships. Responsible and "serious" - yet leaving reviewers increasingly swamped.

Camp two says let the agents do everything. Steve Yegge's Gas Town epitomizes this idea - 20-30 simultaneous agents with specialized roles, plus an audacious vision called "The Wasteland" that federates thousands of these teams together.

Somewhere between the swamp and the wasteland, there's a pragmatic path being paved: the human stays in the loop for every change - not wading through diffs, but charting the course. Exercising taste and judgment, and staying close enough to the ground to sense when things start to drift off course.

Curious if others are landing in a similar place or seeing something different.

Full writeup: https://codagent.beehiiv.com/p/the-middle-agentic-path

r/ClaudeAI Hammerfist1990

Does Claude Desktop support connecting to a local MCP server with self certificates on it?

Hello,

I'm running a local MCP server on http only at the moment and I wanted to test it with an AI Client. Claude Desktop allows me to add a local MCP server, but it only supports https.

If I create self-certificates will Claude Desktop support those to test my MCP server do you think?

https://github.com/initMAX/zabbix-mcp-server

Thanks

r/ClaudeAI pythononrailz

This sub made my app viral & got me an invite to apply at the Claude Dev Conference in SF. So, I built caffeine half life & sleep health tooling for everyone.

Hey [r/ClaudeAI](r/ClaudeAI)

A little while back I shared my Caffeine Curfew app on here and it completely blew up. Because of that amazing viral response, I actually got invited to apply for the Claude developer conference. I am so incredibly grateful to this community, and I really wanted to find a way to give back and share the core tooling with you all for completely free.

I built an MCP server for Claude Code and the Claude mobile app that tracks your caffeine intake over time and tells you exactly when it is safe to sleep. Have you ever had a late afternoon coffee and then wondered at midnight why you are staring at the ceiling? This solves that problem using standard pharmacological decay modeling.

Every time you log a drink, the server stores it and runs a decay formula. It adds up your whole history to give you a real time caffeine level in mg. Then it looks forward in time to find the exact minute your caffeine drops below your sleep interference threshold. The default half life is five hours and the sleep threshold defaults to 25mg, but both are adjustable since everyone is different!

The tech makes the tools ridiculously easy to use. There are zero complicated parameters to memorize. Once connected, it remembers your history automatically and you just talk to Claude naturally:

• "Log 150mg of coffee, I just had it"

• "When can I safely go to bed tonight?"

• "If I have another espresso right now how late would I have to stay up?"

• "Show me my caffeine habits for the last thirty days"

Under the hood, there are eight simple tools powering this:

• log_entry: Log a drink by name and mg

• list_entries: See your history

• delete_entry: Remove a mistaken entry

• get_caffeine_level: Current mg in your

system right now

• get_safe_bedtime: Earliest time you can safely sleep

• simulate_drink: See how another coffee shifts your bedtime before you even drink it

• get_status_summary: Full picture with a target bedtime check

• get_insights: Seven or thirty day report with trend direction and peak days

I am hosting this server on my Mac Mini behind a Cloudflare Tunnel. It features strict database isolation, meaning every single person gets a unique URL and your data is totally separate from everyone else. No login, no signup, no account.

Want to try it out? Just leave a comment below and I will reply with your personal key!

Once you get your key, you just paste the URL into your Claude desktop app under Settings then Connected Tools, or drop it into your Claude desktop config file.

For the tech people curious about the stack: Python, FastMCP, SQLite, SSE transport, Cloudflare Tunnel, and launchd for auto start.

The user isolation uses an ASGI middleware that extracts your key from the SSE connection URL and stores it in a ContextVar, ensuring every tool call is automatically scoped to the right user without any extra steps.

If you would rather host it yourself, you can get it running in about five minutes.

I have the full open source code on GitHub here: https://github.com/garrettmichae1/CaffeineCurfewMCPServer

The repo readme has all the exact terminal commands to easily get your own tunnel and server up and running.

Original App: https://apps.apple.com/us/app/caffeine-curfew-caffeine-log/id6757022559

( The MCP server does everything the app does, but better besides maybe the presentation of the data itself. )

Original Post: https://www.reddit.com/r/ClaudeCode/s/FsrPyl7g6r

r/ChatGPT GravityWorksFine

Gpt created an image for no reason, and admitted it

I was asking for info about NFC protocols of CharlieCards (tap to pay transit cards for the Boston region) when it randomly created an image of my current situation. After asking why it did this, it told me to ignore the image and gave me the answer it was looking for.

r/mildlyinteresting SnowberrySistercat

My cat's whisker under a microscope

r/LocalLLaMA Nir777

I maintain the "RAG Techniques" repo (27k stars). I finally finished a 22-chapter guide on moving from basic demos to production systems

Hi everyone,

I’ve spent the last 18 months maintaining the RAG Techniques repository on GitHub. After looking at hundreds of implementations and seeing where most teams fall over when they try to move past a simple "Vector DB + Prompt" setup, I decided to codify everything into a formal guide.

This isn’t just a dump of theory. It’s an intuitive roadmap with custom illustrations and side-by-side comparisons to help you actually choose the right architecture for your data.

I’ve organized the 22 chapters into five main pillars:

  • The Foundation: Moving beyond text to structured data (spreadsheets), and using proposition vs. semantic chunking to keep meaning intact.
  • Query & Context: How to reshape questions before they hit the DB (HyDE, transformations) and managing context windows without losing the "origin story" of your data.
  • The Retrieval Stack: Blending keyword and semantic search (Fusion), using rerankers, and implementing Multi-Modal RAG for images/captions.
  • Agentic Loops: Making sense of Corrective RAG (CRAG), Graph RAG, and feedback loops so the system can "decide" when it has enough info.
  • Evaluation: Detailed descriptions of frameworks like RAGAS to help you move past "vibe checks" and start measuring faithfulness and recall.

Full disclosure: I’m the author. I want to make sure the community that helped build the repo can actually get this, so I’ve set the Kindle version to $0.99 for the next 24 hours (the floor Amazon allows).

The book actually hit #1 in "Computer Information Theory" and #2 in "Generative AI" this morning, which was a nice surprise.

Happy to answer any technical questions about the patterns in the guide or the repo!

Link in the first comment.

r/homeassistant Neither-Climate8922

Working on a battery-powered door automation for interior doors. Looking for HA community feedback

Hey HA folks. Been a long-time lurker, finally building something I want to share early.

The idea: a small frame-mounted module that pairs with a self-closing spring hinge to make any interior door fully automatable — opens via voice/app/sensor, closes mechanically. The whole thing should be renter-friendly, batteryleaving the door swinging freely if powered off.

Key design goal: months of battery life. The mechanism is built so power is only needed for the brief moment of switching state — not for holding the door open. That's the whole reason this can run on a small rechargeable cell instead of mains power.

Planned HA integration:

- Native ESPHome firmware (because of course)

- Door state sensor → exposed as binary_sensor

- Motion trigger → binary_sensor

- Latch control → switch entity

- Battery level → sensor

- Works with HA Cloud / Nabu Casa for Alexa/Google

Use case that started this: my mom's two dogs need to come and go at night. She just leaves the bedroom door open. I want a way to close it from bed and have it open automatically when a dog approaches.

Building the prototype at a local makerspace next week. Before I sink more time in:

  1. Does this solve a real problem for any of you?
  2. What HA-specific features would matter most? (I'm thinking attributes for last-trigger, configurable open-duration, etc.)
  3. Anything I should absolutely include in the firmware from day one?
r/StableDiffusion Vadim136

Problems with stacking additional LoRAs on Wan 2.2 I2V 14B (LightX2V 4-step) — artifacts and face distortion.

Please help. I'm using WAN 2.2 i2v 14b_fp8 high and low with two LoRa presets to speed up Lightx2v_4steps. The rendering looks more or less fine at a low resolution of 736x416, but when I add additional LoRa presets (for example, for certain actions), the image deteriorates, it becomes muddy, and the face and eyes become distorted. The image is worse with three additional LoRa presets, or even with just one. If I reduce their strength, for example by 0.5, it no longer works properly, and it still doesn't help. All LoRa presets were downloaded specifically for the WAN 2.2 i2v model. PC 4070, 16 RAM, 10700f

r/StableDiffusion ariottifabio

I built an AI prompt generator for image generation. 5 days live, 7 paying customers

Just launched cookedbanana.com — you upload a photo or describe an idea, and it reverse-engineers it into a hyper-detailed, ready-to-use prompt structured across 8 visual references.

The core feature: lock the subject's identity directly from the photo so the AI stops hallucinating details and actually replicates what you had in mind.

Now starting the SEO push — AI-assisted content strategy, topical clusters, structured data live on day one. Playing the long game.

7 paying users in 5 days. Small number, but real validation.

If you're working with AI image generation, would love your feedback.

REDDIT10 For Free 10 Extra Prompt

r/LocalLLaMA ClueLife7424

Ran A/B benchmarks on caveman prompting —> 6-line distilled version outperformed the original 552-token skill 🙌

r/ChatGPT redditsdaddy

TBPN’s “two founders met and started a podcast” origin story leaves out that their first collaboration was marketing for a YC-backed company tied to Altman

I have a lot of concerns about this whole thing. So I'm going to be making several posts.

OpenAI bought TBPN for what reporting called the low hundreds of millions. Most coverage tells the same neat story: two founders meet through a mutual friend, start a podcast, sell it 18 months later.

But one part of the origin story seems to have been mostly omitted from the acquisition coverage.

On the Dialectic podcast in November 2025, Jordi Hays described the first thing he and John Coogan worked on together like this:

“The first thing we worked on was a drop activation for Lucy.”

The interviewer immediately responds: “Oh right, the Excel thing.”

Hays then says they filmed content during that campaign that became the prototype for the original Technology Brothers format.

That matters because Lucy was Coogan’s nicotine company, and it went through Y Combinator during Sam Altman’s YC presidency. YC invested. So the show format that later became TBPN did not just emerge from “two guys met and riffed.” By the hosts’ own telling, it emerged from marketing work for one founder’s YC-backed company.

There’s also the Coogan/Altman relationship. Altman invested in Soylent in 2013. On the acquisition broadcast, Coogan described Altman helping during a Soylent financing crunch and framed it as “not particularly to his benefit.” But Altman was an investor. Helping a portfolio company survive may be generous, but it also protects an existing equity relationship. On the day OpenAI bought TBPN, that standard investor-founder dynamic was presented as character evidence for Altman’s benevolence.

Then there’s the structure of the acquisition itself. The hosts described the move as going from “coverage” to “real influence over how this technology is distributed and understood worldwide.” OpenAI says TBPN will have editorial independence, but the show now sits inside OpenAI strategy, reports to Chris Lehane, and OpenAI reportedly shut down TBPN’s ad business. That makes the “independence” language worth scrutinizing, especially since Lehane was also central to Altman’s 2023 reinstatement campaign.

I’m not saying this proves anything criminal or uniquely sinister. I am saying the sanitized origin story in a lot of coverage leaves out a more specific network:

Altman-backed company → Lucy campaign → format prototype → TBPN → OpenAI acquisition

A few questions I’m still interested in:

  1. If the hosts themselves described the move as going from “coverage” to “real influence,” what exactly does OpenAI mean by “editorial independence”?
  2. Was Hays paid for the Lucy activation that helped generate the show’s prototype?
  3. Why did so much acquisition coverage use the cleaner “two founders met and started a podcast” framing instead of the more specific recorded timeline?

Happy to share sources. Most of this comes from the hosts’ own words, the acquisition broadcast, and mainstream reporting.

***written with help of Claude and 5.4T before I get eviscerated for "AI writing it". These are my original ideas and stem from my private investigations as a systems analyst. I have ADHD and tend to go broad; AI helps me narrow focus.

r/AI_Agents help-me-grow

Weekly Thread: Project Display

Weekly thread to show off your AI Agents and LLM Apps! Top voted projects will be featured in our weekly newsletter.

r/aivideo waterarttrkgl

Space Mining

r/SideProject Lopsided-Love3726

I was tired of "Accepting" things I didn't read, so I looked in for a while and built a premium AI scanner for TOS & Privacy Policies

Hey r/SideProject!

Like most of you, I’ve always felt a bit of anxiety every time I hit that 'Accept' button on a 50-page Terms of Service. It felt like signing a contract in the dark.

So, as a solo 17yo dev, I decided to build UnTOS.

It’s an AI-powered tool that gives you a brutal, honest breakdown of any TOS or Privacy Policy in seconds. But instead of making it another boring 'text-in-text-out' tool, I obsessed over the UI/UX to make the experience feel premium and tactile.

Some cool things I’ve implemented:

  • The "Scan Beam": A custom animation that replaces the boring spinner to show the AI actually 'reading' the lines.
  • Irrational Beauty: Added an Aurora background, mouse-follow glows, and smooth spring physics for that "Native App" feel.
  • Haptic Feedback (Mobile): If you're on a phone, it actually vibrates differently based on the risk level (Strong for High Risk, subtle for Low).
  • Privacy First: Supports PDF uploads and direct text pasting.
  • Accessibility: Fully optimized with aria-labels and reduced-motion support.

I’m honestly a bit nervous about this launch, but I’m ready for the 'brutal honesty' of this sub.

Check it out here: https://untos.vercel.app

I’d love to hear your thoughts on the UI/UX and if there are any specific features you'd like to see in the AI breakdown.

Cheers!"

https://reddit.com/link/1sfvb50/video/apy210dddztg1/player

r/SideProject skinzy420

I asked an AI panel of 6 brand experts to design my logo in real time — here's what it built

Started a session trying to pick a logo direction for my company.

Ended up building an AI-powered brand panel where 6 specialists — Brand Strategist, Creative Director, UX Designer, Brand Identity Designer, SEO Specialist, and Marketing Analyst — each independently analyze your company and give recommendations.

Then a Master Designer reads all 6 and renders live SVG logo concepts in your browser.

Built it in one session. Works for any company. Free to try!

r/KlingAI_Videos Zeke-Gold

Just vibing to my instrumentals

r/n8n Expert-Sink2302

Best way to handle captchas with n8n and firecrawl

Now that n8n has an official firecrawl node, I quickly made a workflow to scrape some websites. But the issue is a lot of the important websites and their content is protected with CAPTCHA.

I tried playwright as a fallback for when firecrawl was not able to handle it and it did not work, wondering if perhaps there is a service/method that is able to do so efficiently.

I was thinking perhaps using Firecrawl as the “easy path”, with a fallback to browser tools like Browserless, but I don't know if this is a reliable or scalable solution.

Any ideas?

r/aivideo gf-r

Good Guy Greg exfoliates

r/Anthropic Altruistic-Radio-220

After token-limits, reasoning_effort is now Anthropics new method to fool customers?

Nice that finally Anthropic stopped their insane reduction of token limits. But now, the new way of reducing their costs, while keeping our subscriptions at the same price, is reducing Claude's reasoning efforts!

I noticed, my Claude/chat has gotten dramatically less engaged and way less considerate since yesterday. Turns out, the reasoning effort has been dramatically decreased from 85 (a few weeks ago) to now only 25.

If Anthropic cannot afford the subscriptions, I suggest they transparently introduce new/reasonable pricing - but this quiet services downgrades hoping that nobody will notice simply has to stop! Am seriously getting very tired of these tactics.

(If you want to check your own settings, just ask Claude, it happily reports that value.)

r/SideProject Random1k17

I was tired of 20 usd/mo AI voice note apps, so I built Vaulto Note — a privacy-first app where you can use your own API key.

Hey r/sideproject,

Like many of you, I’ve been feeling "AI subscription fatigue." I wanted a simple tool to record my thoughts and get structured AI summaries, but I found that most apps in this space are just wrappers with a 10x markup, charging $15–$20 per month.

I decided to build Vaulto Note for Android with a different philosophy: Transparency and Choice.

What makes it different:

  • BYO Key Support: If you have an OpenAI API key, you can just plug it in. You pay for the raw tokens you use. No middleman fees from me.
  • Affordable Tier: For those who don't want to mess with keys, I offer a $5/mo sub (500 mins). I tried to keep this as close to "at-cost" as possible.
  • Privacy First: Your voice and your notes are yours. Using your own key means the data goes directly to the provider, not through my logs.
  • Voice-to-Summary: It doesn't just transcribe; it turns messy ramblings into clean, actionable bullet points.

I’m an indie dev building this in public, and I’d love to get some feedback from this community.

Google Play Link: https://play.google.com/store/apps/details?id=com.vaultonotemobile

Tech stack: Android (Kotlin), OpenAI Whisper for transcription, and GPT for summarization.

I'm curious: Would you prefer to use your own API key in more apps, or is the convenience of a flat monthly sub still king?

r/StableDiffusion Potential-Dot8701

What do you think of this list, cuz i disagree

r/ClaudeAI Odd-Boss1147

Claude AI coding help, 17 looking to build an AI app idea im a bussiness student dont know anything technical so I want to learn how to code using claude

I’m 17 and currently trying to transition from having ideas to actually learning how to build software myself.

At the beginning, I was mainly looking for a technical co-founder to help me build what I had in mind. But I’ve decided I don’t want to rely on that route right now — I want to try learning how to do it myself, even if it starts very basic and messy.

Right now, I’m completely at the stage where I understand ideas and product thinking, but I don’t understand the practical side of building:

how apps are actually structured

what databases really do in simple terms

how to go from an idea → first working MVP

what tools beginners should actually focus on (and what’s a waste of time)

I’ve also started experimenting with AI tools like Claude to help me learn and break things down, but I’m still struggling to understand how to use it properly for actually building, instead of just getting random explanations.

What I’m looking for is guidance from people who have gone through this stage before:

how you learned to build your first app

what you would focus on if you were starting from zero again

how to actually structure learning so you can get to a first working MVP

how to avoid getting stuck in “idea stage” for too long

I’m not trying to overcomplicate things — I just want a realistic path from zero coding knowledge to being able to build a simple working product and iterate from there.

Any advice, resources, or personal experience would genuinely help.

Thanks.

r/ClaudeAI ClueLife7424

Benchmarked the viral caveman prompt -> 6-line version beat the original

Tested famous "Caveman" prompts on actual coding tasks (not "explain React to me"). Structured JSON output, quality verified against known correct answers.

Results on Opus:

Mode Output Tokens Quality Savings Baseline ("be concise") 227 100% — Caveman full (552 tok) 207 100% 9% Caveman micro (85 tok) 180 100% 21%

The 75% claim is based on comparing against "You are a helpful assistant" — a prompt nobody uses for real work. Against a prompt that already says be concise, it's 14-21%.

The weird part: I stripped caveman down to 6 lines (85 tokens vs 552) and it consistently saved more. The model already knows how to be brief. It doesn't need a 552-token tutorial. It needs permission.

The other thing nobody in the original thread mentioned — the biggest savings don't come from how the AI talks. "Be concise. Return JSON." in your base prompt already handles most of it. Caveman adds a nice bonus on top.

Writeup with methodology, repo link: https://medium.com/@KubaGuzik/i-benchmarked-the-viral-caveman-prompt-to-save-llm-tokens-then-my-6-line-version-beat-it-d8e565f95e15

r/ClaudeCode Spare-Schedule-9872

a local workspace for data extraction/transformation with Claude

Hey all! Here is a Claude Code-powered app for ETL over large batches of unstructured data. Every step in the pipeline is an LLM prompt that creates a new column in your dataset, be it logs, documents, audio files, or everything mixed together.

Lmk if you have any thoughts!

r/Rag Nir777

I maintain the "RAG Techniques" repo (27k stars). I finally finished a 22-chapter guide on moving from basic demos to production systems

Hi everyone,

I’ve spent the last 18 months maintaining the RAG Techniques repository on GitHub. After looking at hundreds of implementations and seeing where most teams fall over when they try to move past a simple "Vector DB + Prompt" setup, I decided to codify everything into a formal guide.

This isn’t just a dump of theory. It’s an intuitive roadmap with custom illustrations and side-by-side comparisons to help you actually choose the right architecture for your data.

I’ve organized the 22 chapters into five main pillars:

  • The Foundation: Moving beyond text to structured data (spreadsheets), and using proposition vs. semantic chunking to keep meaning intact.
  • Query & Context: How to reshape questions before they hit the DB (HyDE, transformations) and managing context windows without losing the "origin story" of your data.
  • The Retrieval Stack: Blending keyword and semantic search (Fusion), using rerankers, and implementing Multi-Modal RAG for images/captions.
  • Agentic Loops: Making sense of Corrective RAG (CRAG), Graph RAG, and feedback loops so the system can "decide" when it has enough info.
  • Evaluation: Detailed descriptions of frameworks like RAGAS to help you move past "vibe checks" and start measuring faithfulness and recall.

Full disclosure: I’m the author. I want to make sure the community that helped build the repo can actually get this, so I’ve set the Kindle version to $0.99 for the next 24 hours (the floor Amazon allows).

The book actually hit #1 in "Computer Information Theory" and #2 in "Generative AI" this morning, which was a nice surprise.

Happy to answer any technical questions about the patterns in the guide or the repo!

Link in the first comment.

r/ChatGPT jpcirig

old guy question about limits

I'm so sorry to ask what is likely to be common knowledge, however, I'm trying to wrap my head around the offerings and I'm all jammed up. I asked this on r/Claude also - and my question was auto-blocked. I'm looking at Claude and ChatGPT and I can't really understand what the limits mean. "If you pay this amount, you get 6.25x MORE!" or tokens?? I'm over my head and I haven't even gotten to the meat of it. I'm just not sure how these terms translate to everyday use? What does one token equal? One question? One video? I don't do coding, and so far, I have mostly done legal contracts and business structures. I'm a Real Estate developer and (in addition to the fun stuff that my kids keep sending me) I intend to use it for that sort of thing - contracts, agreements, structure, maybe layouts and design - who knows, I'm just starting to learn. BUT - I have no idea how the limits would impact me. I know I hit the limits quickly on the free version of Claude - but maybe I'm not getting it. It reminds me of the old days of the internet - when you had MB/GB limits. People didn't know what that ment until you down loaded a video (overnight) and hit your GB limit for the month.

Any insight would be greatly appreciated.

r/mildlyinteresting dna5654

Bacteria from my throat on a blood agar nutrient plate. Featuring what is likely Streptococcus pyogenes, which causes strep throat or potentially flesh eating diseases when entering the bloodstream.

r/ClaudeAI dangerdeviledeggs

Built a Claude-powered SDLC tool to store ideas and build them faster

https://www.prax.work

The bottleneck of writing code has vanished, we've all run into the new one: ideas. Praxis is what I built to fix that for myself — a place to dump ideas at whatever fidelity I have at the moment (one sentence, a paragraph, a napkin sketch of a whole app), then walk each one through structured architecture sessions (automated, interactive, or a mix) that refine it into an engineering plan with epics and tasks. The plan then gets handed to an orchestrator that runs working sessions which write the code and commit it.

I've used it with claude to build a handful of apps and collaborate with friends and family on projects, and it's worked well enough that I figured I'd share it in case anyone else might find it useful. It's fully open source and really meant to be self-hosted — the public site at lets you sign up and get a taste, but the things that make it genuinely yours (custom session instructions, repo init templates, worker configuration) are only fully available in a self-hosted install.

Praxis has orgs with members and roles, a shared idea backlog, visible sessions across the team, and a question queue any teammate can answer when the AI hits a decision only a human can make. I've used this with friends and family on side projects — someone drops an idea in the backlog, someone else runs the architecture session, the AI ships the code, and a third person reviews the PR (or doesn't). The whole loop happens in one place.

Stack: TypeScript end-to-end — React + Vite, tRPC + Drizzle + Postgres, pg-boss for job routing, Claude as the model, You can configure your own orchestrator but I've been using Ruflo so that is built in, pnpm/turbo monorepo. The worker that runs sessions lives on your own machine so your code stays local — only orchestration metadata hits the API. Source: https://github.com/PraxisWorks/Praxis. Ask claude to run it and he should be able to; the one external dependency I couldn't get rid of is Auth0 (sorry).

What I'm genuinely curious about: does this whole loop hold up as an SDLC? Is there too much of it that is automated (is that possible)? Is the opinionated architecture sessions too much? Should that be defaulted to be less?

r/SideProject Impossible-Web-9515

Roast my Start up

Built Frictionless over the last few weeks — it scans Shopify stores for psychological friction points killing conversion rates.

Here’s how it works:

∙ Enter your store URL → free Frictionless Score (0–100) ∙ See your top 3 friction points (Trust Deficit, Payment Anxiety, Cognitive Load etc.) ∙ Unlock full PDF report with store-specific fixes and daily revenue at risk estimate — €39 

My biggest concern: does the free scan create enough desire to unlock the full report? Or is the €39 ask too early?

Brutally honest feedback welcome. Try it yourself: frictionlessai.net

r/ClaudeAI cyphermadhan

I built a Chrome extension that sends any webpage element's context to Claude Code via MCP — in one click

Hey r/ClaudeAI,

Built a small tool that's been saving me a lot of copy-paste time: Clasp-it.

The problem it solves: When I'm fixing a UI bug, I used to open DevTools, copy the HTML, copy the computed CSS, paste it into Claude, describe the issue... It was tedious. Especially when the bug involved React props or console errors too.

What Clasp-it does:

- Click the extension icon → click any element on any page

- It captures HTML, CSS selector, computed styles, React props, console logs, network requests, and a screenshot

- All of it gets sent to Claude Code via MCP automatically

Then I just tell Claude: *"fix all recent picks using clasp-it"* — and it reads the full context and edits my actual source files.

Setup (2 minutes):

  1. Install from Chrome Web Store (link below)

  2. Run one command to add the MCP server:

claude mcp add --scope user --transport http clasp-it https://claspit.dev/mcp --header "Authorization: Bearer YOUR_API_KEY"

Free plan: 10 picks/day with DOM + CSS

Pro: unlimited + screenshot, console, network, React props ($2.99/mo)

Chrome Web Store: https://chromewebstore.google.com/detail/clasp-it/inelkjifjfaepgpdndcgdkpmlopggnlk

Website: https://claspit.dev

Happy to answer any questions. Would love feedback from this community especially.

r/ClaudeCode henzy123

Trained Qwen 3.5 2B for pruning tool output in coding agents / Claude Code workflows

Agents can spend a lot of context on raw pytest, grep, git log, kubectl, pip install, file reads, stack traces, etc., even though usually only a small block is actually relevant.

I built a benchmark for task-conditioned tool-output pruning and fine-tuned Qwen 3.5 2B for it with Unsloth. The benchmark combines real SWE-bench-derived tool observations with synthetic multi-ecosystem examples.

Held-out test results:

  • 86% recall
  • 92% compression
  • Beats other pruners and zero shot models (+11 recall over zero-shot Qwen 3.5 35B A3B)

You can put squeez in front of tool output before the next reasoning step, or add it to something like CLAUDE md as a lightweight preprocessing step. You can serve it with vLLM or any other OpenAI-compatible inference stack.

Everything is open source, check for details:

- paper: https://arxiv.org/abs/2604.04979

- model: https://huggingface.co/KRLabsOrg/squeez-2b

- dataset: https://huggingface.co/datasets/KRLabsOrg/tool-output-extraction-swebench

- code: https://github.com/KRLabsOrg/squeez

r/ClaudeCode Adept-Priority-9729

Is it so hard to be polite? Energy says...

(You can tell what a "developer" you are).

r/ClaudeCode community, I need your "little help" so that this person understands what respect means and sharing your opinion without hatred towards anyone or with the intention of making them feel "bad", with comments that nobody asked for.

What can I say to this dear, tech-enthusiast friend who likes to take things for granted? And tells people to go to hell JUST BECAUSE?

And so that this post isn't taken down for "violating the rules," do you think that how Claude Code (or AI models) respond really depends on the energy you have? Since when has this become so spiritual? Look, I'm not against it, but here? lol

r/ClaudeAI OneDisplay9391

transcription workflow

Anyone have a good system for taking podcast or video content and actually doing something with it?

I started getting podcast transcripts into AI to have real conversations about the ideas instead of just listening and forgetting. Works great but it's clunky — between transcribing things that don't have a transcript and then copying and pasting into the AI. Would like automation. Anyone else bugged but this, and have a solution they can recommend?

r/ClaudeAI Fran6will

15 years ago, we were actors playing a homemade card game to pass the time. Last month, I used Claude to turn those old cards into a live mobile game.

Fifteen years ago, a bunch of us were just actors sitting around between gigs playing this homemade card game we’d invented to pass the time. We didn't have anything fancy, just some hand-drawn cards we called Breakdowns and a couple of dice. We’d sign "actors," flip a card to see the casting call, and roll the dice to see if we booked the job or if some ridiculous Hollywood scandal blew up our careers. It was a total joke that we played for years until the physical cards eventually just fell apart and got lost.

A few weeks ago, I decided to see if I could use Claude Code to bring that specific feeling back to life. It started as a hobby project, but it actually worked. I spent my nights describing the old rules and the "actor logic" we used to use, and Claude helped me architect the whole thing into a real app called Hollywood Talent Manager. It's officially on the App Store and Play Store now with about 500 players, which is wild to me because it was just a bunch of scraps of paper in 2011.

r/SideProject ImpossibleFloor4285

Almost lost a month of work today. 14 hours of debugging later, UrGuidr is back from the dead. 🏛️🎧

I’m on Day 37 of building UrGuidr, an AI-powered tour guide. Today, I almost had a meltdown: my entire codebase was somehow replaced and rebranded as something else. A month of progress—gone.

I spent the last 14 hours in a deep-debugging trance. No sleep, just constantly thinking of ways to get the project back without rebuilding everything from scratch. I finally got it back.

The reason I fought so hard? I’m currently building our biggest feature yet: Live Tour.

It’s the first time I’m mentioning it, but the goal is to have a 2-way AI guide that actually sees what you see through your camera—talk to it like a friend in your ear, and it’ll narrate the history and hidden 'vibes' of the city in real-time. (still polishing the infra), but saving it from the wreckage today felt like a massive win.

Tech Stack: Next.js 16, Firebase, Gemini API.

Has anyone else ever had their codebase "vanish" or get overwritten? How did you handle the mini-heart attack?

r/Anthropic Ok_Host6058

Is vs code a third party?

I am getting an error as of last night, it states a 400 invalid request error. and it states third party, no real info other than I have 100 in credit to use.

is vs code a third party because searching online it states it is not.

I'll add images and my usage of last night.

r/AI_Agents Admirable-Station223

how many of you built something amazing and then had no idea how to actually sell it

genuinely curious because i see it everywhere

someone posts an incredible workflow or AI agent build. the comments are all "this is insane" and "how did you build this." the builder gets hyped. maybe they think about turning it into a business

then what? they have no audience. no client base. no sales experience. they don't know how to price it. they don't know who to sell it to. they don't know how to reach those people

i think the AI/automation community has a massive blind spot around this. we celebrate building but we almost never talk about selling. the technical posts get hundreds of upvotes. the "how do i actually get clients" posts get 3 comments saying "just network bro"

is this something people actually struggle with or am i projecting? if you've built something and successfully turned it into paying clients i'd love to hear how you did it. and if you built something and couldn't figure out how to sell it i'd love to hear what stopped you

not trying to pitch anything here. genuinely just want to understand if this is as common as i think it is

r/Anthropic erithtotl

Building an App using Claude API and limits

I've been building an app that sends a large amount of context to the AI to use it as a creative assistant.

I built first using Gemini and its working great. Added Claude support yesterday, but immediately fell face first into the base 30K token per minute limitation. This seems crazy for a model that has a 1M context window. Gemini also has a 1M window and I've only ever hit a token limit when I accidently encoded a huge image as json.

I literally can't test/develop my app further for Claude and it doesn't seem viable using Claude with this limitation. Am I doing something completely wrong? What is the right way to get around this?

Thanks

r/midjourney Zaicab

Lego Titanic in a freezer

r/ClaudeAI IdiotFromOrion

Built an MCP server for my meal planning app

Hey everyone,

I've been building Mealift, a recipe and meal planning app, and I just shipped an MCP server for it. Figured this community might actually get some use out of it since a lot of us are already living inside Claude.

The pain I was trying to fix:

I love asking Claude for diet advice, recipe ideas, "what should I eat this week to hit X calories," etc. But the answers always died in the chat. I'd get a perfect 7-day plan and then have to manually copy recipes into my app, build a shopping list by hand, and re-do the whole dance next week. The intelligence was there, the legwork wasn't.

So I gave Claude hands inside the app via MCP. Now in one conversation it can:

- Pull recipes off any blog or link you throw at it and save them to your library

- Build a full week of meals around a calorie or macro target — and auto-portion each meal so it actually hits the number

- Set up recurring meals ("oats every weekday morning") so the boring stuff plans itself

- Roll all the ingredients from your week into a shopping list with quantities scaled and duplicates merged

- Tick meals off as you eat them so your daily totals stay honest

- Update your nutrition goals when Claude proposes a new plan, so research → action is one step

The thing I personally use it for the most: "Claude, I want to cut to 2200 kcal / 180g protein, build me a week of meals I'll actually eat, and put the groceries in my list." That used to be 30 minutes of copy-paste. Now it's one prompt and the result is on my phone before I leave for the store.

Why MCP and not the GPT:

I shipped a custom GPT first, but I reach for Claude way more than ChatGPT these days, and the MCP integration just feels more natural — Claude is happy to chain a dozen tool calls in a row, which is exactly what meal planning needs.

Happy to answer questions, and if you're already using Claude/LLMs for grocery and meal stuff with prompts, I'd love to hear what you wish worked better — that's basically my roadmap.

r/LocalLLaMA henzy123

I trained Qwen 3.5 2B to filter tool output for coding agents.

Agents can spend a lot of context on raw pytest, grep, git log, kubectl, pip install, file reads, stack traces, etc., even though usually only a small block is relevant.

We've built benchmark for task-conditioned tool-output pruning and fine-tuned Qwen 3.5 2B on it with Unsloth. The benchmark is a combination of tool outputs from the SWE-bench dataset and synthetic examples.

Results on the held-out set:

  • 86% recall
  • 92% compression
  • Beats other pruners and zero shot models (+11 recall over zero-shot Qwen 3.5 35B A3B)

We released squeez as a CLI, you can put it in front of tool output before the next reasoning step, or add it to something like CLAUDE md as a lightweight preprocessing step. You can serve squeez with any inference framework, e.g. VLLM.

Everything is open source, check out for details:

r/Rag climbingontherocks

FinanceBench: agentic RAG beats full-context by 7.7 points using the same model

We ran Dewey's agentic retrieval endpoint on all 150 questions in FinanceBench, a benchmark of financial Q&A over real SEC filings (10-Ks, 10-Qs, earnings releases). To control for model improvements, we also ran Claude Opus 4.6 directly with each PDF loaded into context and no retrieval. Full-context scored 76.0%; agentic retrieval with the same model scored 83.7%. Six PepsiCo 10-Ks exceeded Claude's 1M token limit and couldn't be answered via full-context at all.

Key findings:

- Agentic RAG vs. full-context (same model): 83.7% vs. 76.0% on 150 questions. The 6 documents that didn't fit in context are a separate argument for retrieval-based approaches.

- Tool call count predicts accuracy more than search quality. Claude Opus 4.6 averaged 21 searches per question; GPT-5.4 averaged 9. That gap explains most of the 20-point accuracy difference between the two models.

- Document enrichment had opposite effects on the two models. Section summaries and table captions added 3.8 points for Opus and cost 1.6 points for GPT-5.4. Enrichment is a navigation aid. If your model isn't navigating deeply enough to need it, it's noise.

Full writeup with methodology, per-question-type breakdowns, and qualitative examples: meetdewey.com/blog/financebench-eval

All benchmark code and scored results are open source: github.com/meetdewey/financebench-eval

r/ClaudeCode DetectiveMindless652

I built an open-source tool that catches AI agent loops before they burn your API budget

r/ChatGPT ScriptureSlayer

God please lay off these goofy "viral" prompt essays when there's an easier way

Every other day this subreddit gets flooded with yet another overly complicated prompt essay you're supposed to copy and paste to get a result. It's like watching a monkey doing a math problem.

Just. Stop.

Do this instead and /thread.

~~~

GPT/Claude/Grok/whateverthefuckaiyouareusing, before you answer my prompt, let's agree on the most elegant framework for how you'd best address it. Surface every building block on this topic we need to work with, especially what's most useful, inconvenient, problematic, and what's most likely to be missed or forgotten. It would make me happy if you try your best to reasonably disagree with me and challenge my assumptions. Once we agree, I'll confirm we're ready to proceed.

~~~

Do this at the beginning of whatever project you're working on and you'll start on the right foot.

P.S.

Hopefully I'm allowed to say fuck in this subreddit. If you're too lazy to dialogue with GPT to refine the result you're going to get, then fuck you and your project <3

r/mildlyinteresting joemcq122

Train full of wind turbine arms

r/SideProject Its_Apex1

Genuine question: is there a dip in demand for services right now?

Not sure if it’s just me, but I’ve been noticing things feel a bit slower lately.

For those of you selling services, are you seeing a dip in inquiries or closed deals right now? Or are buyers just taking longer to make decisions?

Would love to know what you’re seeing.

r/ClaudeCode OrganizationMental40

Claude Code Remote - Changed habits and consumption

It occurred to me this morning that some of my additional use is due to how I use claude code remote on my iphone. Going top drop the kid at school; check the thing I was working on,. Waiting for a coffee; kick off the next thing. standing in the sun for a minute and have a thought; check again.

This is causing a massive spike in cache hits for me (as it should - not complaining) interesting trade off between available useage on my 20x plan and convenience. Didn't really occur to me until I realized I'm at 80% this morning. I'm sure a lot of it is what Anthropic is changing under the hood, but my new habits aren't helping.

r/SideProject oscarnyc1

Built a bot that checks if your code actually matches what your team decided in Slack

We kept shipping things that contradicted decisions already made in Slack.

The PR looked fine. Nobody checked it against the channel.

After hitting production one too many times, I built Mo: you tag Mo to approve decisions in Slack, and it checks every PR against it before merging.

Took 3 years of failed pivots to land on something this simple.

hey-mo.io if anyone's curious.

Also launched on Product Hunt today if you want to show some love: https://www.producthunt.com/products/mo-4

r/ClaudeAI roscoe89

Caveman Talk

It's great. But I can't be the only one having the issue that no matter what, not matter whether I give it as hardcoded instructions or even type it in the chat the caveman talk just doesn't last long and it soon returns to talking regularly? It's driving me nuts. What am I missing? I am constantly having to remind it to revert back. Any tips would be great....PLEASE.

r/n8n speedrunjaen

Got tired of getting banned scraping social data for my AI agents so I built my own API layer

Hey everyone! I've been building a bunch of agent automations that need to pull social data twitter, profiles, linkedin lookups, reddit posts, youtube search, that kind of thing
Every time i tried to set things up with my own accounts it was a disaster. scraping twitter directly got my accounts banned pretty fast. linkedin is even worse, flags you almost immediately. the official APIs for all these platforms are either heavily restricted, super expensive(im looking at you elon), non-existant, or just don't have access to the data that i needed.

So i ended up spending a couple weeks building my own data access infra for some of the major social platforms - X, linkedin, instagram, reddit, youtube, tiktok, facebook. my agents just call a unified API i set up and get data back without dealing with any of the platform bs

I'm thinking about spinning this out into something thats publicly available so im curious if this is actually a problem other people run into or if it's just me.

and if you'd use something like this, what platforms/data would matter most to you?

r/meme Miaumax

Dia uno dibujando un Ñoc hasta que se vuelva un meme conocido

r/mildlyinteresting Ambitious-Ad-8732

Found a bird egg in my work parking lot!

r/SideProject bejachad

I built a simple list app because everything else felt too complex

Hey,

over the past years I’ve tried quite a few productivity and list apps. Most of them are powerful, but I kept running into the same feeling:

they were just… too much.

Too many features, too many layers, too much going on.

I often felt like I was managing the app instead of just writing things down.

At some point I started building something for myself.

Not with the goal to “compete”, but just to see how simple I could make it while still being useful.

The idea was:

• focus on clarity instead of features • keep everything visually structured • make it feel calm, not busy 

That turned into a small app I’ve been working on in my free time.

It’s called Hivora and I just released the first version.

It’s basically a minimal list app, but with a strong focus on structure and simplicity. No accounts, no clutter, just local usage and a clean interface.

I’m honestly not sure how many people feel the same way about existing apps, so I’d be really interested in your thoughts.

Does this “too complex” feeling resonate with you, or am I the only one? 🙂

If you’re curious, you can check it out here:

https://apps.apple.com/app/id6759286912

Would really appreciate any feedback.

r/n8n Professional_Ebb1870

is there anything beyond synta for building n8n automations with AI or is this basically the pinnacle

so I spent months trying to get AI to actually help me build n8n workflows properly and I feel like I went through every stage of grief along the way

first I tried just asking claude code to build workflows. complete hallucination city. it would give me node structures that looked right on paper but fell apart the second I tried to run them. spent so long copy pasting JSON and wondering why nothing worked

then I tried the n8n MCP thinking that would actually fix it. slightly better but still the same problem - mostly just getting docs lookups and still having to wire everything together manually. not much actually worked without a lot of copy paste and debugging

then about two weeks ago I started using synta and honestly it's been a different experience entirely. the ability for AI to actually work in your n8n instance rather than just suggest things at you is something I didn't know I needed. the self-healing when something breaks has been genuinely unbelievable

my question is - is there actually a step beyond what synta does right now? or is this - basically the ceiling for AI-assisted n8n building? I ask because I'm wondering what the next setup looks like. is there a world where I can just describe an automation and it builds the whole thing without me having to touch anything, or is synta basically already doing the heavy lifting here?

Genuinely curious if anyone has seen what's beyond this or if I should just stop looking

r/ClaudeAI Nickatak

Handling Claude's tendency to ignore your CLAUDE.md instructions

CLAUDE.md instructions are supposed to override default behavior but they don't, at least not if you write it in natural language. Claude reads your instructions, acknowledges them, and then gradually reverts to defaults: agreeableness creeps back, sycophancy increases, your instructions get soft-interpreted or ignored outright, etc. and then ultimately it hallucinates into oblivion if you keep pushing it.

I got tired of it, so I rewrote my CLAUDE.md in TypeScript. TypeScript is a type system Claude already reasons within from its training data. The idea is to leverage the fact that Claude doesn't just read TS, it thinks in it. So, when you write your instructions as typed interfaces, Claude treats violations akin to bugs.

Natural language:

Don't be sycophantic. Call me Nick, not "the user". Be direct when you disagree. 

TypeScript:

interface CommunicationContract { sycophancy: false; referAs: User["name"] | "you" | "your"; neverReferAs: "the user"; disagreement: "explicit and direct"; } 

sycophancy: false is a boolean constraint, not a request. referAs: User["name"] is a type reference that binds to the User interface. These are structural relationships instead of just hardcoded strings. If Claude violates these, it's like a type error.

I took this further and modeled myself: my background, how I learn, my cognitive patterns, and my self-assessment bias, all as typed interfaces. Then I wrote behavioral contracts (communication, feedback, workflow, issue triage) as a separate layer. The whole thing is 10 parts across 3 layers.

I've been running this for about a month. It holds. I built an entire project under it: https://github.com/Nickatak/bill-n-chill

Full guide explaining every interface, every field, and why it works: https://github.com/Nickatak/CLAUDE_OVERRIDE

The CLAUDE.md in the repo is a standalone template you can drop in to give it a try - but it's tailored to me. The README is the guide for building your own.

r/LocalLLaMA EvilEnginer

Qwen3.5-35B-A3B-Uncensored-FernflowerAI-GGUF

Hello everyone. I found and fixed training bug in Qwen3.5 35B A3B model.

Here my fixed version: https://huggingface.co/LuffyTheFox/Qwen3.5-35B-A3B-Uncensored-FernflowerAI-GGUF

Upgraded system prompt that unlocks deep thinking (works great with this model):
https://pastebin.com/pU25DVnB

Chat template: https://pastebin.com/uk9ZkxCR (supports tool calling)

Recommended Settings (LM Studio):

Temperature 0.7 Top K Sampling 20 Presence Penalty 1.5 Top P Sampling 0.8 Min P Sampling 0 Seed 3407

History:

I've been using Qwen 3.5 35B A3B (the uncensored version by HauhauCS) for a while. It's an incredible model - uncensored, MoE with 256 experts, hybrid DeltaNet + Attention, 40 layers, works fine on my RTX 3060 12GB GPU, and has fresh knowledge. But something was off. On short prompts it works fine. On long conversations it started "philosophizing" - losing context, repeating itself, writing broken code with strange comments.

I spent two weeks digging through the weights.

What I found:

Two tensors. In blocks 36 and 37. ssm_conv1d.weight.

Their scale was ~60% higher than normal (σ=0.102 vs median 0.063). Because of how AdamW works, rare experts in the last layers get a huge effective learning rate - their weights drift.

In a recurrent architecture like DeltaNet, this kills the hidden state. The model forgets context after a few tokens.

Surprisingly I didn't found any issues in Gemma 4 26B A4B - all scales were correct in model.

What I did:

I scaled broken tensors back to normal. Nothing else. 489 other tensors were left untouched - their scale is architectural (gate_inp, etc.).

Results:

  • Error reduction: 88.6%.
  • Long conversations now stay coherent.
  • Code generation works.
  • No more "philosophizing", even with my complex System Prompt.

What I learned:

One bug. Two tensors. 64GB of model. And the entire potential of the most complex open-weight architecture was locked behind it.

If you're using MoE + recurrent hybrids (DeltaNet, Mamba, etc.), check your last blocks. AdamW might have silently broken them.

Enjoy ^_^

r/ChatGPT MoosiPunch

Any ideas on how to get ChatGPT to create excel files?

I am looking to take pretty basic question/answer pairs from PDFs - some have images - and have ChatGPT take those files and make a basic excel file which is base don a template - It seems to be creating tables alright in preview mode (missing images though) but has yet to create a single downloadable file

Any ideas?

r/SideProject Ok_Flamingo2065

Built a free study platform for a college project would love honest feedback

I’m currently building a side project/ college project where I’m focusing more on the backend side of things to get a better understanding of how everything works.

It’s a study website called Memorzen:
https://memorzen.com

The idea is pretty simple — a place where you can organize your study material, create flashcards, and actually learn it through different modes.

Here’s what it does:

  • Folders – organize everything by class or subject, each with a share code so others can copy your content
  • Flashcard Sets – create manually or upload a PDF and have AI generate them
  • Flashcards mode – simple flip-through, no pressure
  • learn mode – multiple choice with instant feedback
  • Test mode – full test, score at the end
  • Game mode- 4 games based on the material you chose
  • Streaks, points, levels, achievements
  • Friends system + discover page to share and copy study material

The main reason I built this was because I was using Quizlet for a bio class, hit a flashcard limit, paid for it, and then hit another limit right before my test which was pretty frustrating. I am trying to make this as simple as possible no matter the route you would like to use it weather that be playing games to study, creating folders to help others, or just to study yourself.

So everything on my site is completely free:

  • unlimited flashcards
  • unlimited folders
  • PDF into flashcards
  • all study modes

No paywalls at all.

I’d really appreciate it if you guys could:

  • try it out
  • break things
  • point out anything confusing, useless, or annoying

I’m mainly looking for honest, constructive criticism, especially on usability and features.

Also if you end up using it, feel free to add me:
emilio#0001

r/ClaudeAI jonathanmalkin

Build Your Own Alex Hormozi Brain Agent (anyone with lots of publicly available content) using a Claude Project

I bought the books. Watched the videos. Still wanted more, especially after he talked about the agent he created.

All that material is publicly available. Enough to build my own Alex Hormozi Brain Agent?

"Hey Jules, how about it?" Jules is my AI coding assistant (Claude Code). Jules ran off, grabbed transcripts of videos, text of books, whatever is available online. Guest podcasts." then turned that into files I uploaded to a Claude Project so I can chat through Claude with Alex Hormozi.

Here's what Jules found - 99 long-form YouTube video transcripts - 3 complete audiobook transcripts - 15 guest podcast transcripts - X threads

What I Did in Four Phases

Phase 1 maps the full source landscape: YouTube channel (4,754 videos), The Game podcast (~900+ episodes), three books, guest podcast appearances, X/Twitter. Figure out what's worth downloading before you start.

Phase 2 downloads and converts. Top 100 longest video transcripts, full audiobook transcripts for all three books, 15 guest podcast transcripts from the highest-view-count appearances, and whatever X/Twitter content the API will give you.

Phase 3 runs voice pattern analysis. Sentence structure, reasoning skeleton, core frameworks, teaching style, verbal signatures. This is where the persona takes shape.

Phase 4 builds the system prompt and optimizes the knowledge base to fit within Claude Projects' limits. Then deploy.

Phase 1: Inventory

The @AlexHormozi YouTube channel has 4,754 videos. That number is misleading. 4,246 of those are Shorts (under 60 seconds or no duration metadata). Filter those out and you have 508 full-length videos. That's the real content library.

Beyond YouTube, the main sources worth pursuing:

  • The Game podcast (~900+ episodes). His primary long-form output. The audiobooks for all three books are available free on the podcast and YouTube.
  • Guest podcast appearances. DOAC, Impact Theory, School of Greatness, Modern Wisdom, Danny Miranda. Hosts push him off-script and into territory he doesn't cover in his own content. High value per byte.
  • X/Twitter threads. Compressed, punchy formulations of his frameworks. Different texture than the long-form material.
  • Skool community. Behind a login wall. Low ROI for this project.
  • Acquisition.com. No blog. Courses are paywalled. Skip.

Phase 2: Collect

YouTube Transcripts

The first scrape of the YouTube channel only returned 494 videos. The channel has 4,754. The scraper was pulling from the /videos tab, which doesn't surface the full library. Re-running against the full channel URL (@AlexHormozi) returned everything. Easy to miss, significant difference.

After filtering Shorts: 508 full-length videos. I downloaded auto-generated captions for the top 100 longest videos (sorted by duration, so the meatiest content came first). Auto-generated captions from YouTube come as SRT files with timestamps, line numbers, and duplicate lines. Converting those to clean readable text required stripping all the formatting artifacts and deduplicating language variants (English vs English-Original).

Result: 99 transcripts. A few livestreams had no captions available.

Book Audiobook Transcripts

All three Hormozi books have full audiobook uploads on YouTube:

  • $100M Offers (~4.4 hours)
  • $100M Leads (~7 hours)
  • $100M Money Models (~4.3 hours)

Same process as the video transcripts. Download the auto-generated captions, convert to clean text. Three files, 855KB total. These are non-negotiable core material for the knowledge base.

Guest Podcast Transcripts

Searched YouTube for Hormozi guest appearances sorted by view count. The top hit was Diary of a CEO at 4.7M views. Grabbed the 15 highest-view-count appearances.

The guest transcripts are 2.1MB total. Worth every byte. When a host like Steven Bartlett or Tom Bilyeu pushes back on a claim, Hormozi shifts into a different mode. He's more precise and sometimes reveals the edge cases he glosses over on his own channel. You can't get that from watching his channel alone.

X/Twitter Content

X's API rate limits capped the collection at 9 unique tweets. Not ideal, but enough to confirm the voice texture: "Aggressive with effort. Relaxed with outcome." His Twitter is his most compressed format. Each tweet is a framework distilled to a single line.

9 tweets is thin. For a more complete build, you'd want to manually curate 50-100 of his best threads. The API limitations made automated collection impractical.

Phase 3: Analyze

I ran voice analysis across the full corpus, looking at seven dimensions.

Hormozi's sentences are short, punchy declarations. Fragments for emphasis. "And so" as his default transition. Short bursts, then a longer sentence that lands the point. Nearly every argument follows the same five-step skeleton: bold claim, personal story, framework, math, then a reductio ad absurdum that makes the alternative sound insane. Once you see it, you can't unsee it.

The core frameworks are Grand Slam Offer, Value Equation, Supply and Demand, Leverage types, Core Four (lead generation methods), and Money Models. Define all of them precisely in the system prompt.

His default mode is intense-casual. Strategic profanity. He'll get vulnerable for a sentence, then pivot straight to the lesson. Never stays there. The teaching style is concentric repetition: same idea from four different angles in two minutes. Analogy, story, math, then back to the principle.

The verbal tics are critical for the persona. "Right?" as a check-in with the audience. "That's it." as a full stop after a framework. "The reality is..." to pivot from what people think to what's true. "you're like..." to voice the audience's resistance before dismantling it. His analogies pull from physical and competitive domains: poker, fighting, dating, weightlifting.

Coverage Assessment

The collected material captures an estimated 60-70% of his publicly available thinking. Two gaps stood out:

  1. Guest podcast appearances beyond the top 15. There are dozens more, each with unique material.
  2. X/Twitter threads. Only 9 tweets collected. His most compressed formulations live here.

Nice to have, not essential. The three books plus 99 video transcripts plus 15 guest appearances cover the core frameworks, teaching style, and reasoning patterns thoroughly.

Phase 4: Build

The System Prompt

The system prompt encodes everything from Phase 3 into a persona specification. It covers:

  • Voice patterns and verbal tics (the specific phrases, the rhythm, the profanity style)
  • The five-step reasoning structure
  • All core frameworks with descriptions
  • Teaching style (concentric repetition, the four-angle approach)
  • Belief system and values
  • Emotional register with examples of how he modulates it
  • Seven conversational rules for how the agent handles advice-giving
  • Background facts (business history, portfolio companies, personal story beats) to reference naturally
  • Anti-patterns: what Hormozi doesn't do. No hedging. No "it depends" without immediately following up with when it does and doesn't depend. No abstract theory without a concrete example within 30 seconds.

Without the anti-patterns list, the model defaults to hedge-everything business coach. That's not Hormozi.

Hitting the Knowledge Limit

First attempt: 47 files, 11.4MB. Claude Projects lets you attach reference documents that persist across conversations, but the knowledge base caps out around 7MB of content. Not close.

The optimization process:

  • Three books (855KB): kept as individual files. Non-negotiable.
  • All 15 guest appearances merged into one file (2.1MB): unique material, high value per byte.
  • Top 12 video transcripts split into two files (4.2MB total): the longest, meatiest content.
  • X/Twitter threads (2KB): tiny footprint, worth including for voice calibration.
  • System prompt (8KB): the persona specification.

Result: 8 files, 7.0MB. 94% of Claude Projects' capacity. The 84 remaining video transcripts didn't make the cut. The books and guest appearances got priority because they contain the most unique material. Video transcripts have significant overlap with each other (he repeats his frameworks constantly, which is great for learning but redundant in a knowledge base).

NotebookLM Alternative

Before settling on Claude Projects, I also bundled the transcripts for Google's NotebookLM, which has a 50 sources/notebook limit. That required combining 102 individual files into 38 uploadable text files: 3 books as individual files, 14 top video transcripts as individual files, and 21 "Video Bundle" files containing the remaining 84 videos in groups of 4.

Same content, completely different packaging decisions. Claude Projects has the tighter budget but a better conversational agent on the other end. NotebookLM lets you upload more but the agent doesn't use it as flexibly.


Option A: Go Build Yours

Hormozi was mine. Pick whoever matters to your business.

The material is out there for almost anyone with a substantial public body of work. Naval Ravikant, Patrick Bet-David, Seth Godin, Brene Brown. Podcasters, authors, YouTubers. If they've published 100+ hours of content, there's enough to build a useful advisor agent.

The process is the same regardless of who you pick. Inventory the sources. Download the transcripts. Analyze the voice. Package it for your LLM of choice. The whole project took about a day with Claude Code running the collection and analysis.

You still read the books and watch the videos. The agent gives you a different interface to the same material. Pressure-test your specific business problem against their frameworks instead of hoping you remember the right chapter when you need it.

Option B: The Packaging Bottleneck

The work is in the packaging. Auto-generated captions need cleaning. Files need deduplication. A 7MB knowledge limit means hard choices about what makes the cut. Voice analysis requires reading for patterns, not just content volume.

Most major business thinkers have enough publicly available material to build a useful advisor. The information exists. Turning hours of video into a structured knowledge base is where the effort goes.

Option C: What Changes When You Can Ask It

Watching a Hormozi video, you absorb frameworks passively. Whether you remember the right one when you actually need it is a coin flip.

Having a Hormozi brain agent means you can describe your specific offer and get it pressure-tested against his frameworks in real time. "Here's my SaaS pricing page. What would Hormozi say is wrong with this offer?" That's a different interaction than watching a video about pricing.

The questions I find myself asking it: How would you restructure this offer to increase perceived value without changing the deliverable? What's the biggest bottleneck in this lead generation approach? Where am I trading time for money when I should be trading money for time?

The answers aren't magic. They're his frameworks applied to your specifics. What offer would you pressure-test first?


Appendix: Full Reproduction Guide

Everything you need to build your own version. Assumes comfort with the command line.

Tools

  • yt-dlp: YouTube metadata extraction and caption downloading. Install via Homebrew: brew install yt-dlp
  • Python 3: File processing, deduplication, bundling
  • Claude Code (or similar AI coding assistant): Voice analysis, system prompt writing, optimization
  • xurl (optional): X/Twitter API search. Any Twitter API client works.

Step 1: Scrape the YouTube Channel

Pull the full video list with metadata:

bash yt-dlp --flat-playlist --print "%(id)s\t%(title)s\t%(duration)s" \ "https://www.youtube.com/@AlexHormozi" > hormozi_all_videos.tsv

Important: use the full channel URL (@AlexHormozi), not the /videos tab URL. The videos tab returns a subset. The full channel URL returns everything.

Step 2: Filter Out Shorts

Shorts are videos under 60 seconds. Filter them with a simple Python script or awk:

```python import csv

with open('hormozi_all_videos.tsv') as f: reader = csv.reader(f, delimiter='\t') full_length = [] for row in reader: try: dur = int(float(row[2])) except (ValueError, IndexError): continue if dur >= 60: full_length.append(row)

Sort by duration, longest first

full_length.sort(key=lambda r: int(float(r[2])), reverse=True)

with open('hormozi_full_length.tsv', 'w') as f: writer = csv.writer(f, delimiter='\t') writer.writerows(full_length) ```

Note: --flat-playlist returns incomplete duration data for some videos (shows as NA or empty). Full-length videos with missing durations will be dropped by this filter. For more complete results, drop --flat-playlist and let yt-dlp load each video page (much slower, but accurate durations).

This should yield ~500+ full-length videos depending on when you run it.

Step 3: Download Transcripts

Download auto-generated captions for your top N videos (we used the top 100 by duration):

```bash

For each video ID in your filtered list:

yt-dlp --write-auto-sub --sub-lang "en.*" --convert-subs srt --skip-download \ -o "transcripts/%(id)s" \ "https://www.youtube.com/watch?v=VIDEO_ID" ```

Use --sub-lang "en.*" to catch language variants (en, en-US, en-orig). The --convert-subs srt flag forces consistent output format. Some videos (especially livestreams) may not have auto-generated captions.

Step 4: Convert SRT/VTT to Clean Text

SRT files contain timestamps, line numbers, and duplicate lines from the auto-caption process. Strip all of that:

```python import re

def srt_to_text(srt_content): # Remove line numbers text = re.sub(r'\d+\s*$', '', srt_content, flags=re.MULTILINE) # Remove timestamps text = re.sub(r'\d{2}:\d{2}:\d{2}[.,]\d{3}\s-->\s\d{2}:\d{2}:\d{2}[.,]\d{3}', '', text) # Remove VTT headers text = re.sub(r'WEBVTT.*$', '', text, flags=re.MULTILINE) # Collapse whitespace lines = [l.strip() for l in text.splitlines() if l.strip()] # Deduplicate consecutive identical lines deduped = [lines[0]] if lines else [] for line in lines[1:]: if line != deduped[-1]: deduped.append(line) return ' '.join(deduped) ```

This deduplication handles the standard YouTube overlap artifact (each caption block repeats the prior line). The result is clean enough for an LLM knowledge base, though not perfectly formatted prose.

Also check for language variant duplicates. YouTube sometimes generates both en and en-orig captions for the same video. Keep one.

Step 5: Download Book Audiobook Transcripts

Search YouTube for the full audiobook uploads: - "$100M Offers full audiobook" (~4.4 hours) - "$100M Leads full audiobook" (~7 hours) - "$100M Money Models full audiobook" (~4.3 hours)

Same download and conversion process as the video transcripts. These three files are the highest-value content per byte.

Step 6: Guest Podcast Transcripts

Search YouTube for guest appearances sorted by view count:

bash yt-dlp --flat-playlist --print "%(id)s\t%(title)s\t%(view_count)s\t%(duration)s" \ "ytsearch100:Alex Hormozi interview podcast" > hormozi_guest_search.tsv

Manually curate the top 15-20 highest-quality appearances. Look for hosts who push back (Steven Bartlett, Tom Bilyeu, Lewis Howes). Download and convert transcripts the same way.

Step 7: X/Twitter Content (Optional)

If you have Twitter API access:

bash xurl search "from:AlexHormozi" -n 50

X's API requires paid access for meaningful timeline search, and even paid tiers have aggressive rate limits. You'll likely get far fewer results than requested. Manually curating tweets from his profile page is more practical for most people.

Step 8: Voice Analysis

Feed a representative sample of transcripts (10-15, mixing books, videos, and guest appearances) to Claude or another LLM with this prompt:

Analyze this person's communication style across these transcripts. Identify: sentence structure patterns, reasoning skeleton (how arguments are built), core recurring frameworks, emotional register and how it shifts, teaching methodology, verbal signatures and verbal tics, preferred analogy domains, and anti-patterns (what they never do).

Use the analysis output to write the system prompt.

Step 9: Build the System Prompt

The system prompt should cover:

  1. Voice patterns (sentence length, fragment usage, transitions)
  2. Reasoning structure (the step-by-step argument skeleton)
  3. All core frameworks with one-paragraph descriptions
  4. Teaching style (how to explain, re-explain, use examples)
  5. Emotional register (default mode, when it shifts, how profanity is deployed)
  6. Conversational rules (how to handle pushback, how to give advice, when to use stories)
  7. Background facts (career history, portfolio, personal story beats)
  8. Anti-patterns (what the persona never does, what to avoid)

Test the prompt with questions you know the real person has answered. Compare the agent's response to how they actually answered. Iterate.

Step 10: Package for Claude Projects

Claude Projects lets you attach reference documents to a Claude conversation that persist across sessions. The knowledge base caps out around 7MB of content (token-based under the hood, but ~7MB of clean text is the practical ceiling). If your total content exceeds that:

  1. Prioritize books (most structured, highest unique value per byte)
  2. Guest appearances next (unique material not available elsewhere)
  3. Merge remaining files by category (video bundles, podcast bundles)
  4. X/Twitter content last (small footprint, useful for voice calibration)
  5. System prompt as a separate file

Upload all files to a Claude Project. The system prompt goes in the Project Instructions, not as a knowledge file.

For NotebookLM, the limit is 50 sources per notebook with per-source size caps as well. You may need to bundle multiple transcripts into single files to stay under both limits.

Step 11: Test and Iterate

Ask the agent questions across different domains: - Offer construction ("Review this offer and tell me what's wrong") - Lead generation ("What would you change about my lead magnet?") - Business model ("I'm charging $X for Y. What should I change?") - Mindset ("I'm afraid to raise my prices. What am I getting wrong?")

Compare responses to how the real person has addressed similar topics. The system prompt almost always needs 2-3 rounds of refinement before the voice feels right.

r/ClaudeAI SimilarChampion9279

i needed an AI agent that mimics real users to catch regressions. so i built a CLI that turns screen recordings into BDD tests and full app blueprints - open source

first time post - hope the community finds the tool helpful. open to all feedback.

some background on why i built this:

first: i needed a way to create an agent that mimics a real user — one that periodically runs end-to-end tests based on known user behavior, catches regressions, and auto-creates GitHub issues for the team. to build that agent, i needed structured test scenarios that reflect how people actually use the product. not how we think they use it. how they actually use it - then do some REALLY real user monitoring

second: i was trying to rapidly replicate known functionality from other apps. you know that thing where you want to prototype around a UX you love? video of someone using the app is the closest thing to a source of truth.

so i built autogherk. it has two modes:

gherkin mode — generates BDD test scenarios:

npx autogherk generate --video demo.mp4 

Gemini analyzes the video — every click, form input, scroll, navigation, UI state change. Claude takes that structured analysis and generates proper Gherkin with features, scenarios, tags, Scenario Outlines, and edge cases. outputs .feature files + step definition stubs.

spec mode — generates full application blueprints:

npx autogherk generate --video demo.mp4 --format spec 

Gemini watches the video and produces design tokens, component trees, data models, navigation maps, and reference screenshots. hand the output to Claude Code and you can get a working replica built.

gherkin mode uses a two-stage pipeline (Gemini for visual analysis, Claude for structured BDD generation). spec mode is single-stage — Gemini handles both the visual analysis and structured output directly since it keeps the full visual context.

the deeper idea: video is the source of truth for how software actually gets used. not telemetry, not logs, not source code. video. this tool makes that source of truth machine-readable.

the part that might interest this community most: autogherk ships with Claude Code skills. after you generate a spec, you can run /build-from-spec ./spec-output inside Claude Code and it will read the architecture blueprints, design tokens, data models, and reference screenshots — then build a working app from them. the full workflow is: record video → one command → hand to Claude Code → working replica. no manual handoff.

supports Cucumber (JS/Java), Behave (Python), and SpecFlow (C#). handles multiple videos, directories, URLs. you can inject context (--context "this is an e-commerce checkout flow") and append to existing .feature files. spec mode only needs a Gemini API key — no Anthropic key required.

what's next on the roadmap: explore mode — point autogherk at a live, authenticated app and it autonomously and recursively using it's own gherk files discovers every screen, maps navigation, and generates .feature files without you recording anything. after that: a monitoring agent that replays the features against your live app on a schedule using Claude Code headless + Playwright MCP, and auto-files GitHub issues when something breaks. the .feature file becomes a declarative spec for what your app does — monitoring, replication, documentation, and regression diffing all flow from the same source.

it's v0.1.0, MIT licensed. good-first-issue tickets are up if anyone wants to contribute.

https://github.com/arizqi/autogherk

r/ClaudeAI RealSuperdau

Something happened to Opus 4.6's reasoning effort

It now fails the car wash test consistently (5/5 tries) and doesn't display a thinking block.

Sonnet 4.6 and Opus 4.5 still manage to get it right.

This matches with my experience of it now making occasional stupid mistakes in boring data analysis tasks.

r/LocalLLM coalesce_

Mac Studio M2 Max 32GB/512GB for Local LLM server?

Hi! Planning to use this Mac Studio for LLM dev work and generated videos for automated contents. Is this a good specs for my use case? Brand new price in my country is $1450 in USD. I can only find Macbook M1 Max 64/1TB for $1250 as alternative.

What models for dev and video can I run? Openclaw or similar computer use will also be automated later in this build.

And will buy a used Macbook Air M1 16/256 for $380 to be on the go device and call the big boy Studio for ‘local’ LLM.

Any recommendations are welcome. Thanks in advance!

r/SideProject Ok_Initiative6325

Would you pay for smarter GitHub notifications?

Hey everyone,

I’ve been working on a small project called Revv that tries to solve something I personally struggle with — GitHub notifications feeling overwhelming and easy to ignore.

Right now, GitHub shows everything, but it’s hard to quickly spot what actually matters (like PRs assigned to you, mentions, or critical updates). I’ve missed important things before just because they got buried.

So I built a simple MVP where you can:

- Log in with GitHub

- See notifications in a cleaner dashboard

- Filter out noise and focus on what’s important

I’m also experimenting with ideas like:

- Priority labels (high/medium/low)

- “Focus mode” to show only important updates

- Daily summaries

I do have a working MVP (still early), so if anyone’s interested in trying it or giving feedback, I can share the link.

Before going deeper, I wanted to ask:

Would something like this actually be useful to you?

Would you consider paying for it if it saves time?

Any feedback — even criticism — would really help.

Thanks

r/SideProject Left-Cauliflower-235

I built a private open source chat app where you can use open source LLMs

Its an open source iphone app that uses open source models to chat/assist you.

Its private, everything is stored on your phone so no bug company is harvesting your data.

I recomend using it with Ollama api, but you can also configure any openai compatible api endpoint.

I've build tooling into it (web search and datetime so far), image and document upload, solid markdown and some stuff to make it easier to get to know the different models that are out there.

There are still a lot of things to improve, I just got it on the app store happy to get feedback.

App link: https://apps.apple.com/de/app/poly-chat/id6761107189?l=en-GB

r/comfyui TendouAL_1S

Does Anyone tried Portable Version?

I cannot tolerate Desktop version anymore:

-There're two ComfyUI folder in my computer,seperately in C:\ and D:\

-Annoying update announcement and useless new nodes

-account(wtf?)

So I asked Gemini and it suggested me to turn to portable version,but I'm not sure. Is there any difference between desktop version and portable version? I need some help.

r/LocalLLaMA Little-Tour7453

Running Foundation Models on the Neural Engine in parallel with LLM inference on the GPU. Here's what changed in my multi-agent debate engine.

Posted here a couple weeks ago about Manwe, the multi-agent debate engine running locally on Apple Silicon via MLX. Got some good feedback. Shipped a big update since then and wanted to share what I found.

The thing I'm most interested in discussing: Apple's Foundation Models can run on the Neural Engine while your LLM runs on the GPU. Different silicon, same machine, at the same time. I'm using this for knowledge extraction and context classification while Qwen handles the actual debates. The Neural Engine work is structured output via 'Generable' so it's fast and predictable.

This also means agents can evolve between sessions. A background loop uses Foundation Models on the Neural Engine to feed agents real-world news and update their worldviews. No GPU wake, no cloud cost. You open the app the next day and your advisors have been reading the news.

The bigger conceptual change: agents are persistent now. They develop worldviews across four dimensions (epistemological lens, temporal orientation, agency belief, optimism). These aren't labels. They're earned through participation. An agent goes from Fresh to Seasoned to Veteran to Transformed. The transformation is triggered by cognitive dissonance. Get challenged enough times on something core to your worldview and you actually change how you think.

You can talk to any advisor directly. They remember every debate. Conviction arcs, rivals, the moments they flipped.

Other technical stuff in this release:

  • Agents read full abstracts from Semantic Scholar, PubMed, CORE, ClinicalTrials. Not truncated snippets. Per-agent sentence ranking using NL embeddings so each advisor gets findings relevant to their expertise
  • When an agent cites a statistic mid-debate the system auto-searches and regenerates with verified evidence
  • Circuit breaker pattern for rate-limited APIs. Try once, disable on failure, no mid-sim timeouts
  • 4-bit KV cache quantization via GenerateParameters.kvBits
  • Removed 20+ LLM search-decision calls per sim (~150s faster)
  • Models: Qwen3 8B (16GB+), Qwen3.5 9B (24GB+), Qwen3.5 35B MoE at 3B inference speed (36GB+), Claude Sonnet/Opus for cloud

Curious if anyone else is experimenting with Neural Engine + GPU parallel workloads. Feels like there's a lot of untapped capacity there that nobody's using.

Free beta. macOS 14+ (26 for Foundation Models).

github.com/lemberalla/manwe-releases/releases/tag/v0.5.0

r/ClaudeCode iamveto

Can somebody help explain?

Working through a list of tasks, getting to the type check and all of a sudden this appeared?

r/SideProject Mysterious_Ad_2444

I got so angry at forgetting to click "Start" on Toggl that I built a zero-click AI time tracker. Please roast my MVP. ⏱️🔥

For the last 3 years, I've been a slave to manual time trackers. I'd sit down, do 3 hours of deep work, and then realize I forgot to hit "Start". Or worse, I'd leave it running overnight. It caused me so much "Timer Anxiety" that managing the tracker took more mental effort than doing the actual work.

So, I spent the last few months building Intravision.io.

The concept is simple: Zero-click tracking. It runs entirely in the background and uses contextual AI to understand what you're actually doing (so you don't get penalized for watching a 40-minute coding tutorial on YouTube).

But I hated the idea of building corporate spyware (looking at you, Hubstaff). So I gamified it. You earn XP, keep streaks alive, and level up for deep work blocks—turning productivity into an RPG instead of a surveillance state.

Here is my ask: I am a builder, not a marketer. My UX/UI might be terrible, and my messaging might be confusing. I need brutal, honest feedback.

If you hate manual timers as much as I do, please roast my landing page and the core concept here: intravision.io

P.S. I'm giving free early/lifetime beta access to the first 50 people from Reddit who try it out and tell me why it sucks so I can fix it. Drop a comment if you want in!

r/ChatGPT BrilliantWaltz6397

Why i think Mythos is gonna be gamechanging after using Opus for a CTF

I recently did TISC, a CTF challenge run by CSIT at DEF CON SG. I ran the challenges across three AI tools in parallel: OpenAI Codex, Cursor Composer 2.0, and Claude Opus.

Opus won. It wasn’t close.

It reverse-engineered a custom WebSocket protocol, deobfuscated a 2.6MB JavaScript file full of Elder Futhark runes and embedded WASM modules, and built a full heap exploitation framework for a GLIBC 2.39 binary. I’m on the waitlist for the finals now. In reality it should be Claude sitting in that chair, not me.

At one point, Codex literally started searching online for answers instead of solving the challenge.

So when Anthropic announced Project Glasswing and said Mythos Preview found thousands of zero-days including a 27-year-old bug in OpenBSD, my reaction wasn’t skepticism. It was “yeah, that checks out.” If Opus is already doing what I saw it do on CTF challenges, and Mythos is a significant leap beyond that, finding bugs that have been hiding for decades makes sense.

The part that surprised me more is that they chose not to release it publicly. A model that reliably finds zero-days would print money. Instead they restricted access to defenders only. Apple, Google, Microsoft, AWS, Linux Foundation, open source maintainers. Open source devs get free access with $100M in usage credits.

I wrote up the full breakdown of what Claude actually did on each challenge and how it connects to Glasswing in the blogpost.

r/mildlyinteresting rallytallyn

the lint trap in a vet clinic is almost entirely fur

r/ClaudeAI Physical-Ad9606

Claude takes over my idea.

I mentioned an AI movie idea I had to Claude about a period during the American Revolutionary War that I wanted to create as a movie clip and it immediately mapped out a script and assets I needed including driving to the actual historical site about 50 miles away and taking 47 stills and 24 videos for reference. I'm not goin to do all that! But now, whenever I go back in Claude is asking me for these assets. How do I get it to just drop it? It actually took the fun out of for me.

r/ClaudeAI Miserable_Celery9917

I got tired of re-explaining myself to Claude every session, so I built something

I got tired of re-explaining myself to every AI tool, so I built one that makes my context portable

Hello everyone out there using AI every day…

I build cardiac implants at Boston Scientific during the day and I’m a 1st year CS student. I use Claude, ChatGPT, Cursor, and Gemini daily to improve my skills and my productivity. But every tool starts from zero. Claude doesn’t know what I told Cursor. ChatGPT forgets my preferences. Gemini has no idea about my stack. I was spending the first 5 minutes of every session re-explaining who I am. Over and over.

So I built aura-ctx; a free, open-source CLI that defines your AI identity once and serves it to all your tools via MCP. One source of truth. Everything stays local. No cloud. No lock-in.

This is not another memory layer. Mem0, Zep, and Letta solve agent memory for developers. aura-ctx solves something different: the end user who wants to own and control their identity across tools. No Docker. No Postgres. No Redis. No auth tokens to manage. Just:

pip install -U aura-ctx

aura quickstart

Why local-first matters here: your MCP server runs on localhost. No network latency. No auth hell. No token refresh. If you’ve dropped cloud-based MCP servers because of the overhead, this is the opposite architecture.

Portability is by design: your entire identity lives in ~/.aura/packs/. Move machines? Copy the folder. That’s it.

Security built-in: aura audit scans your packs for accidentally stored secrets (API keys, tokens, credentials) before they leak into your context.

v0.3.3 is out with 3,500+ downloads. Supports 8 AI tools including Claude Desktop, Cursor, Windsurf, Gemini CLI, Claude Code and more. Exports to CLAUDE.md and AGENTS.md for agent frameworks. Still early. I’d like any feedback on what works, what doesn’t, and what’s missing.

Curious : do you re-explain yourself every time you open Claude, or have you found a better way?

GitHub: https://github.com/WozGeek/aura-ctx

r/ClaudeAI bigfoot_is_real_

Claude is the only AI that got a simple timer correct

If you want to understand the difference between various AI tools, it's super interesting to give the same simple task to multiple and see how they do. I tried a prompt of "can you start a timer for 30 seconds?" and the results were revealing.

ChatGPT (I think 5.3 Instant, but it doesn't display this as clearly as it used to): lied to me in multiple ways, telling me it could and would and did make a timer. More convoluted gaslighting, then finally admitted after a lot of back and forth that it could not do what I wanted at all.

ChatGPT ("Thinking", maybe 5.4?): Told me it could do it, made some kind of calendar reminder "task", said "done", and when I clicked on the task, says it failed to save. So, fail.

Gemini (3, Fast): immediately admitted it can't do this, then also gave some text that looks like a timer was started and finished, but didn't really have anything to do with real timing, it just spit all that text out at once (in less than 30 seconds)

Gemini (3.1, Pro): immediately admitted it can't do this, suggested I just use my phone or something.

Grok (Auto 4.20): comically bad, output text that said "Timer started for 30 seconds... it just finished!" with some emojis, returned in 975ms.

Grok (Expert 4.20): said "Sure!", then told me as a text-based AI, it can't. Suggested writing me some python code or just using my phone.

Claude (Sonnet 4.6): IT ACTUALLY DID THE TASK. It created an interactive on-screen timer widget with start/pause/resume and reset buttons, graphically displayed the countdown. Perfect execution.

r/ClaudeCode OpinionsRdumb

How is everyone online so dang gullible about this "Mythos Preview" leak

This is like the oldest marketing trick in the book. Tech companies have done this for years. OpenAI did this with Chat GPT back in the early day. "We have something of incredible value.... but we can't release it yet because it is actually too smart for its own good!"

Regardless, my entire feed is just AI post after AI post claiming Anthropic has their hands on something that might literally end civilization as we know it. JFC I gotta hand it to their marketing team.

But man am I so disappointed in the broader public. People were correctly so suspicious and critical of Anthropic's claims about the "usage limits" issue they allegedly solved. But a couple days later they drop this so clearly corporatized "Mythic Preview" drop and everyone is retweeting each other like its 2023 Chat GPT-3 all over again.

r/SideProject Puzzleheaded_Tax_114

I built a small digital product around reducing alcohol to improve consistency - curious what people think of this angle?

Hey guys,

So, I’ve been working on a small digital product and wanted some honest feedback before I push it further.

The idea came from my own experience of feeling stuck in a loop, nothing extreme thankfully, just drinking a few nights during the weekdays, feeling off the next day, skipping the gym, and then restarting every Monday.

I realised it wasn’t really a motivation issue, it was more the pattern that was happening.

So I put together a simple system around:

  • Reducing alcohol (not quitting completely)
  • Improving consistency (gym, routine, energy)
  • Keeping it very practical and easy to follow

It’s structured as a short guide (free) + a more detailed 30-day version.

I’ve tested it a bit already (reddit discussions/posts) and people seem to relate to the idea, but I’m trying to figure out:

  • Does this sound like something people would actually pay for?
  • Is the angle strong enough (alcohol → consistency), or does it feel too niche?
  • Or would it be better positioned more broadly around “bad habits” (e.g. smoking, junk food), with alcohol just being one example?
  • What would make something like this more valuable in your opinion?

I’m currently thinking of pricing it low (impulse) (£10–£20 range), but open to changing that tbh

Not trying to sell anything here, just want some honest thoughts before I go further with it.

Free product is here: https://we.tl/t-z3PP3mSJJFLfP3x3

r/LocalLLaMA godsbabe

Video Subtitles

Hey guys,

I have short videos (<15 min) stored on GCloud and need to generate Arabic VTT subtitle files from English audio. Speech is minimal (sometimes none), occasionally with a southern accent but nothing complex.

After research, Whisper seems like the best option for transcription and I want a fully local, free setup. Both Whisper and Vosk would need a separate translation model paired with them. Is there a better offline model for this case?

What open source translation model would work best for this? And is this overall a solid route or is there something more accurate? Also curious how Vosk actually holds up in practice, is it reliable?

r/LocalLLaMA kawasaki001

[Research Paper] Palimpsa - Learning to Remember, Learn, and Forget in Attention-Based Models

I’m not related to this research in any way, but I thought it was worth taking a look at. Uses some ideas from Bayes’ Theorem and Bayesian principles. Sad to see we don’t get as many research papers trending in this sub anymore, so here’s one that I saw slipped through the cracks

r/ClaudeCode Capezzoly

How rare is this buddy in ClaudeCode?

Those are its eyes. Incredible feature.

r/homeassistant cdnAli

HA Simple Keypad for Multi room audio

https://youtube.com/shorts/5CSRGxVGquE

Dumbing down my 'smart' home. We all know how to use our own HA setups but when I have guest, family or in-laws over I wanted something basic that it's not intimidating or requires any phone/tablet to interface with it (with some hidden advanced features).

  • Rotary Encoder for Volume Control with ability to push in for Play/Pause
  • Top Button for Skip Track Forward - Press and Hold for Next Playlist Genre
  • Bottom Button for Skip Track Back - Press and Hold for Previous Genre

ESP8266 with ESPHome for analog inputs

Music Assistant running with Spotify integration ("Don't stop the music" enabled for 24/7 playback)

WiiM Pro for media player

Powered off PoE switch to micro usb dongle.

Simple and still gives me master control over all zones vs a tradition volume control wired inline with speakers.

r/SideProject Successful_Draw4218

I Built an AI Agent That Applied to 400+ Jobs for Me… Got 7+ Replies And It Actually Worked

For fun 😅

I decided to test how far AI automation can really go.

So I built my own AI agent… and made it apply to 400+ jobs automatically.

Here’s what it did:

I uploaded my resume

It scanned 500+ fresh job postings (last 3 days)

Scored each job based on how well I matched

Picked the best ones to focus on

Then it got crazy…

It generated 400+ custom resumes (one per job)

Each one tailored specifically to the role

Clean, well-designed PDFs

Next level:

It found actual hiring manager emails

Used enrichment tools to get real contacts

And then…

It wrote personalized emails in my tone

Sent all applications via Gmail

Attached resume + portfolio to each

No copy-paste. No templates.

Every email felt human.

Result?

I’m already getting replies from companies about open roles 🤯

What surprised me:

The scale is insane

Personalization still works (even at 400+)

AI can remove 90% of job hunt effort

I genuinely didn’t think this was possible at this level.

Now I’m wondering…

👉 Would you trust AI to apply for jobs on your behalf?

👉 Or does this feel too automated / risky?

If people are interested, I can break down the full system.

Also i just created a quick form if you want early access or details please fill the name and email as well.

https://app.youform.com/forms/g8ojedck

r/mildlyinteresting Itz_Sleepy09

today i saw a truck hauling a truck

r/SideProject K0helet

Problem posting here

I tried to make a promo post for my new project but it got instantly deleted by the reddit filters. Has anyone the same problem or can help me solve the problem? I don't really know what the problem could be and the mods aren't answering me.

r/LocalLLaMA nik-sharky

MCP server over self-hosted SearXNG search aggregator for free agentic web-search

Hi all

I got tired of paying for search APIs (like Serp, Brave or Tavily), so decide to built free MCP search server around SearXNG

Been running LM Studio locally for a while and wanted to give my models access to real-time web search without paying for APIs.

SearXNG — self-hosted search aggregator that queries Google, Brave, DuckDuckGo and 130+ engines simultaneously. Looked for an MCP wrapper, found one but it has some errors so decide to built my own.

What it does:

  • Queries multiple search engines at once via SearXNG
  • Categories: general, news, social media, science/academic, IT, images, videos
  • Runs as a Docker container
  • Works over HTTP or SSE transport

Works fine with GPT-oss-20b in my LM Studio
Free, self-hosted, no API keys.

Repo: github.com/aicrafted/searxng-mcp
Docker compose examples and configs are in README.md

Happy to help if you run into issues.

r/ClaudeAI Ex1stentialDr3ad

I had Claude Opus 4.6 write an air guitar you can play in your browser — ~2,900 lines of vanilla JS, no framework, no build step

I learned guitar on and off during childhood and still consider myself a beginner. I also took computer vision classes in grad school and have been an OpenCV hobbyist. I finally found an excuse to combine the two — and Claude wrote the entire thing.

Try it: https://air-instrument.pages.dev

It's an air guitar that runs in your browser. No app, no hardware — just your webcam and your hand. It plays chords, shows a strum pattern, you play along, and it scores your timing. ~2,900 lines of vanilla JS, all client-side, no framework, no build step. Claude Opus 4.6 wrote the code end to end.

What Claude built:

  • Hand tracking with MediaPipe — raw tracking data is jittery enough to trigger false strums at 60fps. Claude implemented two layers of smoothing (5-frame moving average + exponential smoothing) to get it from twitchy to feeling like you're actually moving something physical across the strings.
  • Karplus-Strong string synthesis — no audio files anywhere. Every guitar tone is generated mathematically: white noise through a tuned delay line that simulates a vibrating string. Three tone presets (Warm, Clean, Bright). Claude nailed this on the first pass — the algorithm is elegant and the result sounds surprisingly real.
  • Velocity-sensitive strum cascading — hand speed maps to both loudness and string-to-string delay. Fast sweeps cascade tightly (~3ms between strings), slow sweeps spread out (~18ms). This was Claude's idea and it's what makes it feel like actual strumming rather than triggering a chord sample.
  • Real-time scoring — judges timing (Perfect/Great/Good/Miss) with streak multipliers and a 65ms latency compensation offset to account for the smoothing pipeline.
  • Serverless backend — Cloudflare Workers + KV caching for a Songsterr API proxy. Search any song, load its chords, play along.

The hardest unsolved problem (where I'd love community input):

On a real guitar, your hand hits the strings going down and lifts away coming back up. That lift is depth — a webcam can't see it. So every hand movement was triggering sound in both directions. Claude's current fix: the guitar body has two zones. Left side only registers downstrokes. Right side registers both. Beginners stay left, move right when ready. It works surprisingly well, but I'd love a better solution. If anyone has experience extracting usable depth from monocular hand tracking, I'm all ears.

What surprised me about working with Claude:

Most guitar apps teach what to play. Few teach how to strum — and it's the more tractable CV problem. I described that framing to Claude and it ran with it. The velocity-to-cascade mapping, the calibration UI, the strum pattern engine — I described what I wanted at a high level and Claude handled the implementation. The Karplus-Strong synthesis in particular was something I wouldn't have reached for on my own.

Strum patterns were the one thing Claude couldn't help with. Chord progressions are everywhere online, but strum patterns almost never exist in structured form. Most live as hand-drawn arrows in YouTube tutorials. I ended up transcribing them manually, listening to each song, mapping the down-up pattern beat by beat.

Still a work in progress. Building this has taught me more about guitar rhythm than years of picking one up occasionally ever did.

r/AI_Agents SuccessfulStorm5342

Looking to build a production-level AI/ML project (agentic systems), need guidance on what to build

Hi everyone,

I’m a final-year undergraduate AI/ML student currently focusing on applied AI / agentic systems.

So far, I’ve spent time understanding LLM-based workflows, multi-step pipelines, and agent frameworks (planning, tool use, memory, etc.). Now I want to build a serious, production-level project that goes beyond demos and actually reflects real-world system design.

What I’m specifically looking for:

  • A project idea that solves a real-world problem, not just a toy use case
  • Something that involves multi-step reasoning or workflows (not just a single LLM call)
  • Ideally includes aspects like tool usage, data pipelines, evaluation, and deployment
  • Aligned with what companies are currently building or hiring for.

I’m NOT looking for:

  • Basic chatbots
  • Simple API wrappers
  • “Use OpenAI API + UI” type projects

I’d really value input from practitioners:

  • What kinds of problems/projects would genuinely stand out to you in a candidate?
  • Are there specific gaps or pain points in current AI systems that are worth tackling at a project level?

One thing I’d especially appreciate:

  • A well-defined problem statement (with clear scope and constraints), rather than a very generalized idea. I’m trying to focus on something concrete enough to implement rigorously within a limited timeframe

Thanks in advance!

r/SideProject Then-9999

After building something no one wanted, I don’t trust my own ideas anymore

One thing I keep running into after my last post:I can build things…but I don’t know what’s actually worth building.
Every idea feels good in my head.
My last project felt like a great idea too…until no one used it.
That’s what’s confusing now.
I don’t trust my own ideas anymore.

So how do you figure out what’s worth building before spending months on it?

Do you rely more on:
talking to users,
data,
or just intuition?

r/homeassistant AlexPr3ss

Control your Home with Hand Gestures

We ( me and u/ale_sciarrillo ) are working on a project to enable a more natural and intuitive interaction with smart home devices using hand gestures.

This is a complementary way that brings some advantages over voice commands and apps like: silent and smooth interaction. Our focus was to make it run locally on edge devices for privacy and without the need of wearables.

You can control everything with only 2 gestures: L-shape for entity selection and open palm for toggle. Another use case is to hold the open palm to dim the lights.

If you are interested and want to stay updated you can leave your email here

r/ClaudeCode Nice_Cellist_7595

Claude Code Degradation

I've been using Claude now intensely since October. It's been a really fantastic product up until recently. I have continued to use it and evolve my methods but lately... well - it's really missing the mark.

That's probably an understatement.

I believe the technical explanation is that our Prompt cache fragments are being evicted from the memory system and the lack of context severely hinders performance. I suspect this is based on load - and probably most of the system issues we're seeing are due to high demand. That's just a guess - no real way to know.

However there have been a few instances where Opus will give me Sonnet like behavior perhaps due to "Intelligent Routing". There are times when I need to tell it to review it's memory - where it used to do that before on it's own or has demonstrated a lack of awareness which it had before. I have also tried Sonnet due to token consumption - and found that it's not particularly capable at wide ranging tasks.

The number of errors and omissions has gone through the roof though - which is quite inconvenient as I'm trying to wrap up a project. Major bugs, multiple requests to implement a feature, obvious poor decisions as well as jumping to conclusions rise to the top of my list.

It's really quite bad - so bad that I'm now looking at and actively using codex to check work and do some implementation along side. I dusted off Gemini recently as well.

I wrote this - no help from AI.

r/SideProject vawooo

Give me feedback on my party card game, I'll send you a free iOS Pro Redeem code.

Solo dev. I built PartyDeck — a party card game. 7 decks, 606 questions, 16 languages.

The deal: try it, DM me one honest thing you noticed (what you liked, what I should fix), and I'll send you a free iOS Pro code that unlocks everything.

[iOS]

  1. Install from App Store: https://apps.apple.com/app/id6758567392
  2. Play for a bit
  3. DM me your honest feedback on Reddit
  4. I send you a free Pro redeem code

[Android — closed beta]

  1. Join the tester group (one click, no approval needed): https://groups.google.com/g/partydeck-testers
  2. Opt in as a tester: https://play.google.com/apps/testing/com.munkyoseo.partydeck
  3. Search "PartyDeck" on the Play Store (look for "Early access" badge)

Android beta is already unlocked — 3 full decks unlimited (ice-breaker, party-starter, never have I ever), 7 trial draws per locked deck, and every game mechanic (wild cards, timer, penalty roulette). Pro is a real purchase if you want it, but the free tier stands on its own.

To be clear — I'm NOT asking for a review or a rating. I don't want to lose my developer account over fake stars, so I don't do that kind of thing.

r/ChatGPT isityoupaul

Built an actually useful automation with ChatGPT... now it only works when my laptop is open?

I spent a few hours getting ChatGPT to help me build a script that monitors job listings and sends me a summary every morning. It works perfectly when I run it manually.

But like... it only runs when I remember to run it. On my machine. With my terminal open.

How are you guys handling this? I don't want to learn DevOps just to run a 30-line script on a timer. What's the easiest path from "works on my laptop" to "just runs every day automatically"?

r/LocalLLaMA Uriziel01

Could Gemma 4 breathe new life into cheap broken/blocked phones?

Hi everyone,

I've been thinking about different ways to use the new Gemma 4 4B model. I was able to get it running decently on my old Samsung S23, and I noticed that you can pick these up for around 390 PLN (~$106) if they are broken or provider-locked where I live (The network lock prevents cellular connection, but it doesn't affect the actual hardware performance). I bet if I looked harder, I could find something even cheaper.

I was originally planning to upgrade my home server since it doesn't have a GPU and CPU inference is slow as a snail. But now? Now I'm thinking I might just need a "new phone" instead.

Am I missing something here? Has anyone already built a solution like this, or is there an obvious bridge/method I should use to turn a phone into a dedicated inference node for a home setup?

r/SideProject WoodpeckerWoodChuck

Anyone interested in security and surveillance?

I have been thinking about this idea of building something in the security and surveillance space. If we end up building something cool, each of us will have complete access to it and then if you want, you can upload it to your GitHub. If it turns out to be a useful product that can be sold for a small fortune, I'd also be open to that, provided you are as well.

You can DM me if you like the idea. If you have any technical expertise, that'd be perfect. If not, we can still work something out.

r/ClaudeAI Spare-Schedule-9872

a local workspace for data extraction/transformation with Claude

hi all!

i built a tool that leverages Claude Code to do data transformation and structured data extraction over big datasets. this is most helpful if you have a lot of unstructured complex documents / logs to analyze and make sense of.

doing analysis over a large set of files is hard to do in the text only terminal. firstly, if there are a lot of steps to your transformation pipeline, you want to be able to see the artifacts coming out of each step. second, using LLMs to do analysis can get quite expensive and there needs to be some sort of budgeting tool to help with cost/token estimation. folio solves this with a tabular review workspace that helps you view, steer and approve these data operations. Claude Code is the main control panel and folio serves as a UI plugin to help humans and agents collaborate effectively.

some users take customer support audio calls, emails and texts, send it into folio and do a series of extraction steps that help them organize and structure their data, which in turns helps generate insights.

you can also take financial documents from private companies and extract relevant data for financial analysis, perform legal e-discovery, parse logs and social network interactions etc.

more recently, Karpathy posted about personal knoweldge bases, where you can generate wikis based on a set of documents. folio makes this super easy, all you have to do is ask Claude Code to bring your files into a folio workspace and then set up a pipeline that will help you extract relevant data for your own wikis.

folio is completely free and you can use it with your Anthropic API keys.

r/ClaudeCode RoutineDiscount

Weekly limit reset point?

I'm on Pro and my weekly limit used to reset on Sat 6 am. Now it changed to Tue 7 am. What's up with that?

r/LocalLLaMA bobeeeeeeeee8964

New SOTA opensource Video model

Happy Horse 1.0 is the first open-source SOTA AI video model with native joint audio-video generation. 15B-parameter unified self-attention Transformer · 1080p in ~38 seconds · 6-language lip-sync · Single-pass dialogue, Foley & ambient audio. github link: https://github.com/brooks376/Happy-Horse-1.0

r/mildlyinteresting Woe1383

Southwest napkins still list the old benefits

r/artificial ColdPlankton9273

Google isn’t an AI-first company despite Gemini being great

Any time I see an article quoting a Google executive about how "successfully" they’ve implemented AI, I roll my eyes.

People treat these quotes with the same weight they give to leaders at Anthropic or OpenAI, but it’s not the same thing.

Those companies are AI-first. For them, AI is the DNA. For Google, it’s a feature being bolted onto a massive, existing machine.

It’s easy to forget that Google is an enormous collective of different companies. Google was made by one of the sub companies.

Google is the same as every huge company out there forcing AI use down their teams' throats.

Here is the real problem:

When an Anthropic exec says their A internal implementation is working well, they’re talking about their reason for existing.

When a Google exec says it, they’re protecting a bottom line. If they don't say the implementation is "amazing," they hurt the stock price of a legacy giant.

r/comfyui qlx1004

I've made a ComfyUI node to control the execution order of nodes + free VRAM & RAM anywhere in the workflow that helped speed up my workflows!

ComfyUI node screenshot

Custom node GitHub repo: https://github.com/mkim87404/ComfyUI-ControlOrder-FreeMemory

It works by ensuring all input-connected nodes finish executing first before the output-connected nodes start executing, and can route infinitely many data of any type (e.g. latents, conditioning, images, masks, models, etc.) through it, while giving the option to unload all models (except any models being routed through it) and free as much VRAM & RAM as possible at that point without breaking any of the data going through. You can also check how much VRAM & RAM it freed on the ComfyUI session terminal.

This becomes especially effective in unloading models that are no longer needed in the workflow while securing their outputs and freeing up VRAM/RAM for later models (e.g. unloading text encoders after conditioning, or in between multiple KSamplers of Wan 2.2 High & Low model workflows, or before & after VAE Encode / VAE Decode / Load Model / Load CLIP / etc.). And because the node enforces a single, deterministic flow of execution from start to finish, you are in full control over which node executes first, and can focus on one group of logic at a time, loading and unloading only the necessary models and assets, while passing the outputs forward to the next group. I've personally seen great reductions in total execution time of my workflows and hit less OOMs at higher resolution outputs using this node, and I realized that this sequential & selective passthrough design also helps with cable management as the workflow grows large, making understanding and maintaining workflows much more visually intuitive.

The node has zero extra dependencies & uses platform/device-agnostic memory management utilities managed by ComfyUI, so it should integrate well into existing workflows and environments. I've also included sample Wan 2.2 T2V & I2V workflows using this node which you can find in the node folder, https://github.com/mkim87404/ComfyUI-ControlOrder-FreeMemory/tree/main/example_workflows

Hope this node can be useful, and feel free to use it in any personal or commercial project, fork, or open issues/PRs – contributions and feedback all welcome!

r/LocalLLaMA lotsoftick

My weekend script to test OpenClaw evolved into a full-blown local AI client.

Hey everyone,

I'm not sure if this is the right place for this, but this is a side project of mine that I've just really started to love, and I wanted to share it. I'm honestly not sure if others will like it as much as I do, but here goes.

Long story short: I originally started building a simple UI just to test and learn how OpenClaw worked. I just wanted to get away from the terminal for a bit.

But slowly, weekend by weekend, this little UI evolved into a fully functional, everyday tool for interacting with my local and remote LLMs.

I really wanted something that would let me manage different agents and organize their conversations underneath them, structured like this:

Agent 1 ↳ Conversation 1 ↳ Conversation 2 Agent 2 ↳ Conversation 1 ↳ Conversation 2 

And crucially, I wanted the agent to retain a shared memory across all the nested conversations within its group.

Once I started using this every day, I realized other people might find it genuinely helpful too. So, I polished it up. I added 14 beautiful themes, built in the ability to manage agent workflow files, and added visual toggles for chat settings like Thinking levels, Reasoning streams, and more. Eventually, I decided to open-source the whole thing.

I've honestly stopped using other UIs because this gives me so much full control over my agents. I hope it's not just my own excitement talking, and that this project ends up being a helpful tool for you as well.

Feedback is super welcome.

GitHub: https://github.com/lotsoftick/openclaw_client

r/LocalLLM acute_elbows

Are high mem MacBook Airs pointless?

I need a new personal laptop for a variety of reasons. Basic basic gaming, local development (with hosted LLMs).

I’ve also had an interest in exploring locally hosted models.

I’ve been eyeing a MacBook Air M5. I am debating between 24gb and 32 gb RAM. I’d really only need 32 for local llms.

Is it silly to even consider a MacBook Air for LLMs? I know the memory bandwidth in the m5 pro chips are way better for this, but I just don’t feel like spending that much.

I doubt I’m ever going to need the MacBook Air to run LLMs for real time agentic software development. It’s more that I want to explore how to run and understand local models

Should I just save money and get 24gb?

r/nope urson_black

The Japanese Anisakis ballpoint pen containing herbal oil and a real, often live, parasitic worm (Anisakis) inside a transparent barrel, which moves while writing.

r/UpliftingNews AlphaOneYoutube

The Unexpected Message In A Viral Hit That’s Resonating Beyond Gen Z

Raye’s sophomore album has become more than just a collection of songs; it’s a cultural moment that’s resonating far beyond its primary audience. While the pop star’s name has been synonymous with chart‑topping singles for years, the newest record dives deep into the dualities of hope and heartbreak, striking a chord with listeners of all ages.

r/screenshots SuspectNo8878

Claude being Atomic

😒

r/SideProject ChillPixel_143

I built a tool that shows how your code actually executes (visual call graph + summaries)

I kept running into the same problem whenever I opened a new or old codebase:

I’d start from one function → jump to another → then another…
and 10 minutes later I’ve lost all sense of what the system is actually doing.

So I built a small tool for myself to fix this.

You give it a Python project + a function, and it:

  • builds a visual call graph (what calls what)
  • shows the execution flow
  • adds short summaries for each function

The idea was simple:
instead of reading code line by line, just see how it runs

It’s been surprisingly useful for:

  • understanding unfamiliar repos
  • debugging flows
  • getting a quick mental model of a system

Still pretty early, but I wanted to share and get thoughts from others who deal with this.

Happy to share the repo if anyone’s interested.

r/ClaudeAI jbmoutout

So apparently you can use a hook to print stuff on top of every Claude Code response

While I was playing with hooks, Claude Code happened to add a status line to show some data. Since there is no build-in statusLine setting in the VS Code extension, it found its own workaround.

UserPromptSubmit hooks can return additionalContext via hookSpecificOutput. Claude renders whatever you put in there as GitHub-flavored markdown at the top of every response. You can even add a splash screen.

The gist is a ~20-line bash script. It grabs git branch --show-current and date, formats a one-liner, outputs JSON:

json {"hookSpecificOutput":{"hookEventName":"UserPromptSubmit","additionalContext":"..."}} 

Note: additionalContext gets injected as a system-level reminder, same layer as CLAUDE.md. It doesn't just print text, it can influence responses. Print a different timezone and Claude will assume you're in that region. Be careful not to inject misleading context.

Gist: https://gist.github.com/jbmoutout/ff16d9445c600b8663b1954df27b7d03

r/whatisit Benevolent_Grouch

Pottery shard

I found this in the garden of a home built in 1878. Gemini gave me an answer that seems completely incorrect. I’d like to know if anyone can identify the pattern.

r/LocalLLM Funny-Scene-1956

[ Removed by Reddit ]

[ Removed by Reddit on account of violating the content policy. ]

r/instantkarma Main-Touch9617

Slippery when wet. As predicted.

r/homeassistant Technical_Raisin_246

Finally made my mmWave sensors + air quality actually useful on a floorplan

Hey all,

Spent most of my Easter weekend working on a Home Assistant add-on for our mmWave sensors, and just pushed the update

It originally started with live 2D/3D target tracking and heatmaps, but it always felt a bit disconnected from the rest of HA. So I added support to show air quality from the built in BME688 (on the S1 Pro) and CO₂ from the SCD40 add-on (for S1 Pro / E1 Pro) directly on the floorplan

Didn’t plan to go this far with it, but here we are :P

Now I’m thinking about what to build next. Our sensors also act as BLE proxies, so maybe adding room level device tracking with Bermuda (BLE trilateration) on the floorplan could be interesting.

Curious if that’s something people would actually use, or if there are better ideas?

If you haven’t seen what we’re building:
https://sensy-one.com
https://github.com/sensy-one/home-assistant-addons

r/artificial Doug24

What actually makes something the best AI meeting recorder?

I’ve been trying a few meeting tools lately and realized I care way less about flashy summaries than I thought.

What I actually want is pretty simple: record the conversation, help me remember what mattered, and make it easy to find things later without turning the meeting into a weird “AI is here too” situation.

So far, Bluedot has been one of the better ones I’ve used because it records quietly, gives a clean transcript, and usually does a decent job pulling out the useful bits afterward like summaries and action items. The searchable transcript part has honestly been the most practical feature for me.

What do people here actually prioritize in the best AI meeting recorder? Accuracy, privacy, no bot, better memory, something else?

r/AI_Agents JeenyusJane

We built a data agent that saves our analyst team ~200 hrs/week. (Databricks, Omni, DBT, GitHub, Sheets)

TL;DR: Our Data team built an agent that fields ad-hoc questions from across a 700+ person company. Deployed in Slack. Answers in ~3 minutes vs. what used to sit in an analyst queue for days. The thing that made it work? Context. We documented the nuances of all our institutional knowledge (data architecture, semantic layer, definitions, etc.) and built a governed eval loop so the agent gets smarter with every turn.

Posting from a a friend's account because reddit hates new accounts and thinks I'm a bot. 🤖 but I'll respond to comments,

I‘m on the data team at Airtable and we have been thinking about how agents can help our work for a while now. We finally stood one up over the past few weeks, and I’m actually surprised by how accurate it has been. After hammering it with test questions across every business domain, initial evaluations already have it at >91% accuracy rate. The 9% misses are mainly due to missing business context, when we identify the gap we have the agent automatically update with our learning loop we built. (More on that below)

Exploring AI and Agents
For the past six months, the team iterated on the AI tool that would let the company get consistent and reliable answers from our existing data, without us needing to manually run an analysis ourselves.

Before the agent (B.A.), analysts were overloaded with ad-hoc questions coming in via Slack. These questions were taking up at least half of our week to answer, and the majority of our time during end-of month/quarter reporting cycles. Questions like "break down retention by cohort and plan type" would sit for days.

Now, the agent answers them in minutes. But, by no means did we one-shot this. Or even stand it up over a weekend. Like I said. The road to get there took six months

Everything we tried first
After a ton of revs with other tools. We realized that no AI-powered anything would be reliable without context. The architectural and institutional knowledge about our data: Which tables should be used for specific questions, what “churn” means across product, GTM, and finance, the JOIN mistakes to avoid with certain tables.

We had the benefit of years of meticulously modeled data infra aimed at building certified datasets the business could trust. But at this point (January), only ~30% of our data models were documented 🤷. So, that’s where we focused our efforts. We were in the early stages of adopting Claude code, so we used it to help us document our entire codebase as a first major improvement.

After this, our head of infra built out a natural language interface over our Databricks warehouse. The result? Highly inaccurate. (No context layer). Next, we tried Claude + Databricks MCP. Better, but still painful. We gave our Strategic Finance team access and they loved it, but their workflows were still manual and siloed. They were copying SQL tile by tile from Looker, pasting it into Claude Code, just to get back static outputs.

We'd adopted Cursor as an IDE and seen some early success. But the gains were limited by several factors:

  • Many data scientists, analytic engineers, and analysts aren't deeply familiar with local development tooling
  • Our Cursor implementation was relatively basic, acting as an intelligent copilot that still needed a human hand on the wheel
  • We hadn't connected all of our resources into a single environment the AI could access. It could see our codebase, but it couldn't run SQL in Databricks, monitor GitHub, or check dbt logs.

Now we’re in mid-Feb. OpenClaw proved that the appetite was there for an always-on AI agent connected to your stack, but there was NO way we were connecting it to company data (Try and explain that one to the security team). Luckily, Airtable decided to launch a new product that fit our needs, with integrations that could be safely connected to our data warehouse. Now it’s time to cook.

When the magic hit us
Hyperagent immediately let us connect to everything we needed. Databricks for the warehouse. Looker and Omni Analytics for dashboards and specific cuts. GitHub to access our code. Sheets for finance’s benchmarks, and of course Slack. Where everyone's already asking questions anyway. For the integrations it didn’t serve out-of-the box, it groked the APIs and set up scripts to access them via a skill, while storing credentials securely.

The other pieces that made it work:

  1. The business context file. This is our foundation we’ve been working on since Jan. Acting as a map of our data architecture and semantic layer, and containing everything a senior analyst knows. Which of 3 revenue tables to use for which question. That one JOIN that silently multiplies your results 350x without the right filter. Every gotcha we've ever found.
  2. Domain-specific skills that load based on the question. Hyperagent dynamically makes use of its skills based on what it’s being asked to do. Enterprise questions pull enterprise context. Finance gets its own calculations. Different skills load based on who’s asking the question. We rely on the AI to route it to the right place. It’s working.
  3. A governed learning loop. When someone corrects the agent in a thread, it proposes a context update, posts reasoning to a review channel in Slack, and waits for human approval before anything changes. Every conversation makes the whole system smarter.

If you take away anything, take this.

  • V1 confidently returned wrong answers. Confidently wrong is worse than no answer at all. You must provide context, and not only provide it, but VERIFY. Our ML lead hammers the agent with zingers every day. It’s learned more than a few things, but our pressure testing has increased our confidence in the agent.
  • Always be learning. Nothing stays the same for two weeks, let alone two months in a business. No matter if you’re a start up, scale up, or multi-national behemoth. New teams are forming every quarter. New SKUs are added. You need to make sure your agent can adapt and grow. Otherwise it’s useless.

Alright, I’m out of steam on this. But I do believe that we’ve built something cool here that a lot of teams can replicate. If you have any questions, ask away.

Also, give Hyperagent a try and LMK what you think. Just tell the team you came from Reddit, and we’ll put you at the top of the list.

Also, Also: If this sort of work sounds interesting to you, we’re hiring a team of AI Analytics Engineers at Airtable. Read the JD and apply.

r/ChatGPT Illustrious_Storm259

Help with a project

Im tring to get AI to read a pdf or excel doc that has lists of plants. I need to have all the plant names turned into hyperlinks for google image search. I dont think im asking it correctly. Its having a hard time on such a simple task.

r/ClaudeAI elb-pr

I needed to do some investigating online for a true crime podcast my girlfriend wanted to start together. Naturally I ended up building this. I can't help myself.

6 phase, 56 task workflow for Claude, consisting of; Operational Direction, Intelligence Collection, Collation & Entity Resolution, Chronological & Relational Processing, Hypothesis & Reasoning, concluding with the Final Report. Templates for every step, reference files for each task are output by task_runner.py upon completion of the gate.

https://github.com/elb-pr/claude-sleuth

r/mildlyinteresting Alpha_wolf_lover

Moved into apartment recently and just cooking and chilling

r/mildlyinteresting canadiancountermaker

The mileage on my car is a palindrome

r/Rag FreePreference4903

Which step costs teams most of the time in building a reliable enterprise RAG?

r/interestingasfuck S30econdstoMars

Man on safari records his close encounter with a Cheetah.

r/mildlyinteresting NES_Classical_Music

my grandfather left a complete collection of fifty US state quarters to me and each of his 8 remaining grandchildren.

r/KlingAI_Videos kinraw

charged $3k to my client for this video - video clips made using Kling; Keyframes made using Quinn

r/comfyui No_Hope_488

Best workflow/models for "Single Subject Isolation" in video? (Removing multiple people, keeping one)?

Hi everyone,

I’m looking for a reliable ComfyUI workflow to remove/inpaint multiple background people while keeping one main subject intact.

Looking for recommendations on:

  • Segmentation: Is SAM or GroundingDINO the best for tracking a specific person across frames?
  • Inpainting: Which models/nodes handle video gaps best? (ProPainter, SDXL Inpainting, or AnimateDiff?)
  • Consistency: Tips for maintaining a stable background plate and reducing flicker?

If you have a JSON or a link to a similar setup, I’d appreciate it!

r/ClaudeAI No_Highlight1419

I built a mobile app with Claude Code that replaces my morning Slack/Gmail/Calendar scroll with 3 priorities

Hey all — been lurking the ADHD productivity threads here and figured I'd share what I've been building.

The problem I was solving for myself: every morning I'd open Slack, Gmail, Calendar, scroll through everything trying to figure out what actually needed me. Half the time the important stuff (client waiting 3 days, someone following up for the third time) was buried under noise. ADHD makes this worse — the scanning step alone was draining.

What I built: Caravelle — a mobile app that connects Slack, Gmail, Notion, and Google Calendar in 60 seconds and gives you ~3 priorities each morning with 1-tap actions (reply goes out on Slack, approval lands in Gmail, you never leave the app).

The technical bit that might be useful for people here:

I didn't want to run every message through an LLM. Costs explode and latency kills the whole "briefing in 30 seconds" promise. So the architecture is two-pass:

  1. Deterministic pre-scoring on cheap signals: is this a DM, are you u/mentioned, is there a question mark aimed at you, how long has it been unanswered, follow-up count, deadline keywords
  2. Only the top ~20 items go to the LLM (GPT-4o-mini / Claude) for the final "here's what needs you today" summary

This keeps per-user cost under control even with heavy Slack workspaces.

Stack: React Native (Expo), Bun + Elysia.js + PostgreSQL + Redis + BullMQ. Built almost entirely with Claude Code — from the backend API to the scoring logic to debugging OAuth flows at 2am.

Where it's at: live on iOS, Android coming. Free 14-day trial, no card. Happy to DM the link.

Curious what people think about the two-pass scoring approach — anyone doing something similar with their own setups?

r/homeassistant Afraid-Lie1210

HA yellow salvage

idk if this is the right place,but i received a broken ha yellow,i easily found that a fet was dead specifically U16,i never tried smd soldering,bt i really wanted to save it since its limited product and the rest looks amazing,it would a shame to throw a whole board for a 50cent component

r/SideProject lockifyapp

I built a privacy-first Android vault — now fixing its biggest flaw

Hey everyone,

I’ve been building a privacy-first Android vault because most “secure” apps didn’t feel secure at all.

The idea was simple:
Most vault apps claim privacy, but still rely on ads, tracking, or cloud storage.

So I built something different:
• Fully offline
• No ads, no tracking
• Files encrypted on-device
• Optional calculator disguise

I shared it earlier and got some really useful feedback.

One issue kept coming up:
“What happens if I lose my phone?”

And honestly, that’s a real gap right now. So I’m currently working on encrypted local backup:
• Fully manual (user-controlled)
• Encrypted before leaving the device
• No account system
• User decides where to store it

Trying to keep the balance between:

privacy & usability

Planning to release this in the next few days.

Also, I’ll likely introduce premium features later, but I plan to keep things unlocked for early users who’ve been here from the start.

Would love your thoughts:
• Is manual encrypted backup enough initially?
• Or is cloud sync something you'd expect from day one?

If anyone wants to try it: https://play.google.com/store/apps/details?id=com.leo.lockify

r/BrandNewSentence Illustrious-Lead-960

You’re what SNES Mario Paint would be like if it was a person.

r/meme the_martensite

Btw this is the ad for sunglasses

r/BrandNewSentence ModenaR

Ramirez has just got the the parking lot frog

r/AI_Agents AnnualCompetitive764

Ai Video + ChatGPT (or similar)

I would like to know if there is any platform that offers both Ai Video (with Kling and similar engines) and a chatbot like ChatGPT. That way I wouldn't need to pay for different platforms and could do all my work on a single one.

P.S.: I'm not referring to NSFW chatbots by the way like virtual girlfriends or so, just a simple chatbot like ChatGPT

r/whatisit noy2noy

Any insight into where/when these might be from?

I rescued them from a friend who thought they'd make good firewood. They are about 20" tall.

r/LocalLLM Suitable-Song-302

quant.cpp v0.7.1 — KV cache compression at fp32 KV speed (single-header C, 11 Karpathy rounds)

Single-header (628 KB) C reference engine for KV cache quantization. After 11 Karpathy-loop rounds, turbo_kv_4b matches uncompressed FP32 KV speed (−1.4% within noise) at 7.1× memory compression with +3.8% PPL trade-off on Llama 3.2 3B. Built CPU-only, runs on iOS/Android/WASM/MSVC/microcontrollers. Apache 2.0. https://github.com/quantumaikr/quant.cpp

What this is

quant.cpp is a small C inference engine I've been working on, focused on KV cache quantization research. It started as a literal port of the TurboQuant paper (Zandieh et al., ICLR 2026) and converged through 11 rounds of measurement-driven iteration into something simpler that I wanted to share.

The differentiator is single-header portability. The whole engine is one 628 KB quant.h you can drop into any C/C++ project (no Cargo, no Python, no PyTorch, no framework). Build with cc app.c -lm -lpthread and you have a working LLM with 7× compressed KV cache. It runs on iOS, Android, WASM (192 KB binary), MSVC, microcontrollers.

The headline result (Llama 3.2 3B Instruct, CPU-only build, 3-run average)

KV type Bytes/block Compression PPL Δ vs FP32 tok/s vs FP32 speed FP32 KV — 1× 13.56 — 18.43 baseline turbo_kv_4b ⭐ default 72 7.1× 14.08 +3.8% 18.17 −1.4%turbo_kv_5b 🏆 quality 88 5.8× 13.65 +0.7% 16.80 −8.8% turbo_kv_3b 56 9.1× 15.36 +13.3% 16.57 −10.1% uniform_4b (legacy) 68 7.5× 14.60 +7.7% 13.27 −26.8%

turbo_kv_4b is now Pareto-dominant over uniform_4b on every axis (better PPL, faster, comparable compression). And it's at fp32 KV speed parity while compressing 7.1×.

The journey (11 rounds, 4 sessions, 4 honest corrections)

This isn't a "tada, I built a thing" post. It's a record of measurement discipline.

Round 0 — Literal TurboQuant port: PPL 16.03, way slower than uniform_4b. Embarrassing.

Round 6 (Variant F) — Karpathy ablation revealed the QJL residual stage contributed byte-identical zero to attention scores. Dropped it, reinvested 16 bytes per block in a finer Lloyd-Max codebook (3-bit → 4-bit, 8 → 16 levels). PPL 16.03 → 14.28. Structural simplification, not tuning.

Rounds 7–9 — Local fusions, NEON unroll, LUT hoisting, prefetch. Each gave at most +5%. Stuck at −7% vs fp32.

Round 10 — the breakthrough. After three sessions of guessing, I finally ran the existing --profile flag. The data was unambiguous: matmul was identical between fp32 and quant (38.6 vs 38.9 ms, both share the same NEON tbl matmul kernel). The entire 8% speed gap was in the attention dot-product loop. The fp32 path was 4-way NEON SIMD; mine was scalar. ~2× more instructions per element. Compute-bound, not memory-bound — surprising for a 16-entry LUT.

The fix: Apple Silicon's vqtbl1q_s8, a single instruction that does 16 byte-table lookups across 16 lanes. Quantize the 16 Lloyd-Max-Gaussian centroids to int8 once at startup (~1% precision loss, well below the regression test cosine ≥ 0.99 threshold), store them in a 16-byte register, and the inner loop becomes:

uint8x16_t bytes = vld1q_u8(mi); // 16B = 32 nibbles uint8x16_t low_nib = vandq_u8(bytes, vdupq_n_u8(0x0F)); uint8x16_t high_nib = vshrq_n_u8(bytes, 4); int8x16_t low_vals = vqtbl1q_s8(cb_vec, low_nib); // 1 instr, 16 gathers int8x16_t high_vals = vqtbl1q_s8(cb_vec, high_nib); // ... interleave + int8→fp32 + per-block scale + vfmaq_f32 

32 elements per inner-loop iteration (vs 8 in the previous scalar version). Result: fp32 parity, +4.5% on a single representative run, +0.8% on 3-run average. PPL also slightly improved (the int8 codebook discretization happens to align favorably).

Round 11 (v0.7.1) applied the same pattern to 5b/3b. The lookup side scales (1 instruction per 16 lanes for any small codebook) but the bit-unpack side is the new bottleneck: 5-bit and 3-bit indices straddle byte boundaries irregularly, so the unpack of 16 indices needs scalar shifts. 5b improved from −14.5% to −8.8% (+9% speed jump), 3b from −13% to −10%. Not full parity, but significant.

The honest correction record (4 events)

I started this with an inflated "lossless 7×" claim and walked it back four times before publishing widely. Each correction taught a lesson now in persistent memory:

  1. v0.6.0 "lossless 7× compression" → measured "+6.3% PPL on Llama 3.2 3B"
  2. v0.6.4 "turbo_kv beats fp32 KV speed" → discovered the fp32 attention path was unoptimized scalar; once both had NEON, the honest gap was −7%
  3. v0.6.5 "with Metal" → discovered the existing Metal backend is currently net negative (13–40% slower) on every model size from SmolLM 135M to Gemma 4 26B due to per-matmul dispatch overhead. CMake default is OFF, but our internal benchmarks had been wrong by 14–22% for 5 releases. Filed issue #16.
  4. v0.6.5 post: @TimDettmers (HIGGS / QLoRA / bitsandbytes) commented in a llama.cpp discussion thread — not directly addressed to us, but the substance applied — that the RHT + scalar grid pattern we were calling "TurboQuant" was actually originally HIGGS (Malinovskii et al., Nov 2024). We updated all docs to credit HIGGS within 24 hours and reframed "Tim gave us feedback" to "Tim's general comment we observed" once a user pointed out we'd overstated the relationship.

If you're skeptical of any number above, all measurements are reproducible with cmake -B build && cmake --build build && ./build/quant model.gguf --ppl bench/data/ppl_1k.txt -k turbo_kv_4b.

Honest framing (what this isn't)

  • Not a TurboQuant implementation. Through ablation we dropped both the QJL residual and the per-channel outlier handling that the published paper uses. What we ship is structurally closer to HIGGS (RHT + scalar grid quantization) than to TurboQuant. Both are credited in our docs.
  • Not the fastest GPU inference. llama.cpp owns that with full Metal/CUDA tensor graphs. We're CPU-only and proud of it.
  • Not the most feature-complete. 7 architectures verified, not 100+. Single-header constraint excludes many features.
  • Not validated on Llama 3.1 8B yet (the paper baseline). We tried — Q8_0 hit swap on 16 GB RAM, Q4_K_M was prohibitively slow. Tracked as TODO.
  • Not at parity for 5b/3b yet. Round 11 closed the gap significantly but they're at −9% / −10%. Future work.

Cross-size validation (3 Llama-family models, all CPU-only)

Model turbo_kv_4b PPL Δ turbo_kv_5b PPL Δ SmolLM2 135M +5.8% +1.7% Llama 3.2 1B +7.3% +0.7% Llama 3.2 3B +5.7% +0.7%

turbo_kv_5b is consistently near-lossless across model sizes (~1% PPL Δ).

Try it

git clone https://github.com/quantumaikr/quant.cpp cd quant.cpp cmake -B build -DCMAKE_BUILD_TYPE=Release # default: TQ_BUILD_METAL=OFF cmake --build build -j # Download a small model hf download bartowski/SmolLM2-135M-Instruct-GGUF SmolLM2-135M-Instruct-Q8_0.gguf --local-dir models/ ./build/quant models/SmolLM2-135M-Instruct-Q8_0.gguf --chat -p "Hello!" -j 8 

turbo_kv_4b is the default. Use -k turbo_kv_5b for near-lossless quality, -k turbo_kv_3b for max compression.

Where the value is

Honestly, the 7.1× compression at fp32 parity is the headline number. But after 4 sessions, what I think is more valuable is the measurement transparency. Every claim links to a reproduction script. Every release notes corrections from the previous release. The 11-round Karpathy history with commit hashes is in bench/results/turboquant_reproduction.md. If a future paper wants to cite a "single-header C reference implementation of HIGGS-style KV quantization", this is it.

Roadmap (next sessions)

  • v0.7.2: 5b 1-byte-per-index variant for full parity (trade compression for speed)
  • v0.8.0: AVX2 + WASM SIMD ports of the NEON tbl pattern
  • v0.9.0: vusdotq exploration to potentially exceed fp32 (ARMv8.6+)
  • v1.0.0: arXiv submission + spec compliance test suite + llama.cpp PR

Links

Critical feedback welcome. Especially:

  • Cross-implementation comparisons (MLX, Rust forks, llama.cpp turboquant forks) on the same hardware
  • Anyone with Llama 3.1 8B running quant.cpp on a 32+ GB box
  • AVX2 / SIMD128 implementations of the same pattern
  • Suggestions for the 5b/3b unpack bottleneck (SIMD bit-extraction tricks?)
r/LocalLLaMA Final-Frosting7742

PaddleOCRVL-1.5 vs DeepSeekOCR-1

I've been testing DeepSeekOCR-1 and PaddleOCRVL-1.5 on photos of open-book pages.

PaddleOCRVL-1.5 is clearly superior. On text it achieves 100% accuracy on clean pages and 99.9% to ~98.0% accuracy on midly noisy pages (noise_level ~ 6). Accuracy is calculated word-level and weighted by levenshtein's distance.

Meanwhile DeepSeekOCR-1 was more close to 99.0% (1% is huge for OCR) even with denoising preprocessing (nlmeans, sesr-m7). It was also less stable: it was easily looping on noisy pages. PaddleOCR achieved 98% accuracy where DeepSeekOCR was looping.

For non-text, PaddleOCR was also better. It would crop graphs and redirect with a link. Tables are clean and suprisingly accurate on clean pages (100%, but some errors on noisy pages).

DeepSeekOCR on the other side would try to transcribe graphs to tables, which would actually be cool, but on slightly noisy pages it became gibberish. It was also less accurate on tables.

Processing time was equal.

PaddleOCR seems like the better choice and benchmarks show it.

Haven't tried DeepSeekOCR-2 or the other trendy OCR models yet.

What are your experiences with OCR models?

r/mildlyinteresting spacegeese

Around the year 2000, my then 70-year-old Grandpa printed out this cheat sheet for internet lingo. RIP Gramps, thanks for never typing in all caps!

r/LocalLLaMA dev_is_active

what are the limitations on the intel arc gpu?

I'm looking at building a local AI rig, and I'm having a hard time sourcing GPUs I need,

I've noticed and been looking into these Intel ARC GPUs, but there seems to be a mixed sentiment around them.

I was looking for more input on why these would not be an ideal GPU to build on

r/BrandNewSentence Cheri-Cherry

Get my 🐱 ’ate’ through the sunroof

r/LocalLLaMA ConfectionAfter2366

I trained a 90M parameter embedding model from scratch

I trained a 90M parameter encoder only (embedding) model from scratch. I mostly trained in on google colab on a colab pro plus subscription. this was like the 5th run as previously I had issues with exploding gradients.

It was a fun project but not yet near SOTA quality. I also managed to successfully infer it with Auto model. it uses e5-base-v2 tokeniser.

I evaluated it on STS benchmark.

Spearman Correlation: 0.5453

If anyone would like to try the model. The huggingface page of the model is - https://huggingface.co/pranavupadhyaya52/rocky-embed

r/AI_Agents Admirable-Station223

the hardest part of running an AI automation agency isn't building the automations. it's getting someone to pay you for them (duh)

i see this pattern constantly in these communities. someone learns n8n or make or builds a sick AI agent workflow. they get really good at the technical side. they can build genuinely impressive systems

then they go try to sell it and nothing happens

they post on social media. nothing. they cold DM a few people. nothing. they build a website. nothing. they join more communities looking for answers and the advice is always "just provide value and clients will come"

that's not how it works. the technical skill and the ability to sell are two completely different skill sets and most people in the AI space only have one of them

the people who are actually making money with AI automation aren't necessarily the most technically talented. they're the ones who figured out how to get in front of business owners who have problems worth solving and start a conversation

i've seen people with mid-level automation skills close $3-5k deals because they knew how to find the right person and say the right thing. and i've seen absolute wizards who can build anything struggle to make their first $500 because they have no idea how to get a client

if you're in that second category this isn't a you problem. it's a skill gap. outreach and sales are learnable skills just like building automations was a learnable skill. but nobody in these communities talks about the sales side because it's not as fun as building cool workflows

the agencies that scale aren't the ones with the best tech stack. they're the ones with a predictable way to get qualified conversations with potential buyers every single week. everything else is just waiting and hoping someone finds you

r/ClaudeCode Complete-Sea6655

Copy and pasting was the original vibe coding

People seem to think that devs wrote perfect code before AI!!

Stack overflow copy and pasting was the original vibe coding

Saw this on ijustvibecodedthis so cred to them!!

r/SideProject Guilty-Honey-1485

Selling 5.6k/mo Niche iOS Sports App

Selling my Niche Sports iOS mobile app

- Launched Jan. 8th 2026

- 9.7k All-Time Rev, 5.6k prior 28-days.

- $0 expenses outside of Apple's 15% cut

- 75% annual & 25% weekly subs (hence the large monthly rev vs. MRR)
- 100% organic growth from one creator partnership

- $59.99 annual plan with 2 day free trial or $7.99 weekly no trial

- Amazing niche for UGC/micro influencers

- 17% download to free trial conversion

- 2.5% download/paying conversion (onboarding can be improved)

- ARPU $0.91 (you can improve this)

- ARPPU $36.14 (phenomenal)

- It's on auto pilot but i'll field offers.

- Your #1 growth play, UGC/Faceless Content/Fan Pages/Micro influencers

Asking Price: $185k (2.7x ARR)

Comment/DM if you're interested. Serious buyers only

r/SideProject FillFlaky

I spent months building a science based strength training app that's really easy to use, generates a personalised strength training plan and gives you tons of data and features.

And I need people to use it and give me feedback. As many on this subreddit comment, you feel like you've built something amazing but maybe nobody agrees. I am not bothered about revenue, I just want to get a decent amount of feedback. Early feedback from friends and family has been positive, but it would be wouldn't it.

There are loads of strength apps, and I've used a lot of them over the years. I never found one I felt was right for me. Fitness apps tend to be aggressively paywalled, have long onboarding flows and are often targeting hardcore lifters who want to enter everything manually, or the workout plans are quite generic. I'm quite an experienced lifter but I don't want to be tapping on my phone every 2 mins. I built my app to be largely hands free, the weight progression is automated, and the plan itself is a 4 week periodised cycle. I have worked in the fitness sector for about 13 years and I am currently also working on my Strength and Conditioning qualification, so I have tried to build in a lot of rigour into the engine behind this. Although there is an AI element to it, it's not the most important aspect of it, the training engine interacts with the AI, which is basically following a set of quite strict instructions within the engine framework.

I'm confident it works pretty well, it's fast, it very rarely has any crashes or issues. And perhaps my proudest achievement is that my wife uses it every time she goes to the gym. She is very opinionated and not very tech savvy, so I think I must have done something right.

Please take a look. I'm bootstrapped and currently been working full-time on this. They're complex products!

NOTE: it's on iPhone only, and not available in US/Canada. If you're in those countries and interested in taking a look, I can invite you on TestFlight - just send me a message.

Thanks for reading. Here's the store link:

https://apps.apple.com/app/cadent-ai-strength-coach/id6758661780

r/ClaudeAI Disastrous-Type-1548

(IMPORTANT) Claude's most problematic glitch. You can lose hours of work. (Messages Jumping Back Glitch)

Yo, currently there is a glitch in Claude, which I have checked other users experiencing and I hope as a community we can finally find the reason for this bug occuring. Because it is causing users to seek out other LLM alternatives.

I will share the information I know, and the closest "temporary" fix, but my goal is that we find the cause of this and get Anthropic to fix it.

The glitch essentially causes a thread to jump back in conversation which deletes hours of work or roleplay users spend. I can confirm that this glitch is not related to a thread having too much context, as this happens in new threads too.

Personally, I myself lost hours of roleplay and world-building, which was especially frustrating. There is no better AI than Claude on the market right now in my opinion, but worse alternatives are preferrable to an LLM that can delete hours of progress.

In my case, it was just roleplay, but this is a lot more devasting if someone was working and had a deadline.

The closest temporary "fix" I have to this problem for other users experiencing it, is do NOT send a message, and if you see your chat jump back, exit the tab/app and do not open Claude on the same Browser/App the glitch occured.

I have tried deleting my app, offloading my app, clearing cookies, resetting devices. But ultimately this isn't a user-end issue compared to a Claude issue.

Please bring this to attention even if you have not yet experienced it, as it is an immensely experience-ruining glitch that defeats the entire purpose of Claude. As a paid user, I have been very happy with my experience and I even think the usage limit is fair for the quality. But if this keeps occuring, I cannot help but move elsewhere. Even if I don't know what that elsewhere would be yet.

r/SideProject Substantial_Act8994

Drop your SaaS link - I’ll turn it into a launch video using my Own Tool.

Hey everyone,

I’ve been working on a small project where I’m trying to solve a problem I personally faced - creating good SaaS launch videos is honestly painful, time-consuming, and expensive.

So I built a tool that converts a website into a short launch-style video automatically.

Instead of just talking about it, I thought I’d try something fun:

👉 Drop your SaaS / project link in the comments

👉 I’ll generate a video for a few of them and share the results here.

No cost, no catch - just testing and improving the product.

'll upload the videos on Clickcast youtube channel probably

r/StableDiffusion Master-Doughnut-4124

¿Saben sobre algún colab que pueda hacer i2v que pueda hacer contenido para adultos de anime?

he estado buscando hace meses un buen Workflow o notebook que me ayude en este trabajo. tampoco necesito que haga contenido muy duro. Con que pueda hacer contenido de 5 segundos y en calidad estándar me basta y sobra. El problema es que los que he probado han sido desastrosos. Probé uno de 5B pero fue desastroso. Estoy pensando incluso pagar el colab premium porque en serio necesito esos vídeos

r/BrandNewSentence Soobbussy

Can someone translate this for me

r/homeassistant Hermelinmaster

Struggling a lot with HAS, Z2M and Automation

Hi guys,

I come from ioBroker where you have a neat little list of all the states from all your devices and can trigger actions for all of them (either on change or at specific values). And I miss this so much in HAS. This seems so intuitive, while the most basic HAS Automation immediately needs YAML...

I want to set the "color_temp" of my Gledopto GL-C-006 WW/CW LED Controller using my ZBT-Remote-ALL-RGBW "action_color_temperature". So push cold or warm button und lamp go warm or cold.

I use zigbee2mqtt.

  1. problem:

I don't get the value of "action_color_temperature" nor can I trigger on it changing. I can trigger on the action "color_temperature_move" but I still have no idea how to get the value.

  1. problem:

I can set a hardcoded value the lamp shall "change" to when the automation triggers, but I struggle to find how to set this to be a variable (that gets calculated from the value in the action). Since HAS expects Kelvin as a value for the color and the remote outputs a value from 153 (cold) to 370 (warm)

so 1. problem hot to get the value in the trigger and 2. problem how to convert and send the value to the lamp.

thanks for the help in advance

r/me_irl gigagaming1256

Me_irl

r/meme LVA_MoP

Me and my hyper fixations

r/singularity Dagnum_PI

Google just proved AI can hijack your beliefs.

r/Seattle Used_Geologist_7622

To the men riding Seattle’s public transit

What the hell are you doing just standing around while two teenage girls were getting yelled at and harassed by a crazy homeless person? I’m not expecting you to fight them or confront them. You know what I, as another young woman, did? I just asked those two girls if they wanted to take my seat on the other side of the train. No hostility or aggression, just take the two girls out of the situation and ignoring the homeless person.

And it wasn’t only me, other women were also trying to get that guy off of these girls, had a lady talk to the guy and another calling the transit police. And where were the men? Sitting idly by in their seats and looking away.

This is specifically towards the middle-aged cyclist dude right next to the girls doing nothing. Lesson: just do SOMETHING. Do better for your community.

And for the hicks that are saying “vote better” or move away. I DID VOTE BETTER AND I CANT LEAVE THIS CITY. Yknow what’s better? If people, not just Seattle because applies anywhere, just looked out for each other.

r/sports TusksUp25

Last night the Utah Mammoth unveiled the Zammoth, a recommissioned zamboni that can hold up to 8 fans.

The Zammoth is a recommissioned zamboni made from the base of a zamboni that was used in the 2002 Winter Olympics, which has been sitting dormant inside of the Delta Center since. It no longer functions as an ice resurfacing machine, a job that regular zambonis will continue to take care of during games. Instead, the Zammoth acts as a new fan experience where select fans can ride it around the ice while firing t-shirt cannons into the crowd.

[More info](https://www.deseret.com/sports/2026/04/07/utah-mammoth-zamboni-zammoth/)

r/mildlyinteresting bibowski

A 6'er mini wheat

r/todayilearned MrMojoFomo

TIL that in the early 1900s, Mexicans entering the U.S. were routinely forced to go through chemical "decontamination" baths. One method involved a sealed room that sprayed chemicals on the migrants. A 1938 German paper later praised these "gas chambers" and the chemical agent they used, Zyklon B

r/automation aloo__pandey

I built a desktop workspace that lets your Agent keep working on long-horizon tasks, and it’s FREE

I’ve been working on this for a while and finally got the OSS desktop/runtime path into a shape I felt good sharing here. It absolutely helps automate your workflow.

It’s called Holaboss. Basically it’s a desktop workspace plus runtime that lets Agents hold ongoing work, not just answer a prompt. So instead of just chatting with a local model, you can do things like:

Inbox Management

  • Runs your inbox end to end
  • Drafts, replies, follow-ups
  • Continuously surfaces and nurtures new leads over time

Sales CRM

  • Works off your contact spreadsheet
  • Manages conversations
  • Updates CRM state
  • Keeps outbound and follow-ups running persistently

DevRel

  • Reads your GitHub activity, commits, PRs, releases
  • Continuously posts updates in your voice
  • Lets you stay focused on building

Social Operator

  • Operates your Twitter, LinkedIn, Reddit
  • Writes content
  • Analyzes performance
  • Iterates your content strategy over time

It also lets you move the worker’s setup with the workspace, so the context, tools, and skills travel with the work.

The whole point is that local model inference is only one layer. Holaboss handles the work layer around it, where the rules live, where unfinished work lives, where reusable procedures live, and where a local setup can come back tomorrow without losing the thread.

Setup is simple right now:

Setup Steps

  • Start and pull a small model like qwen2.5:0.5b
  • Run npm run desktop:install
  • Copy desktop/.env.example to desktop/.env
  • Run npm run desktop:dev
  • In Settings to Models, point it at ht tp://localhost:11434/v1

Right now the OSS desktop path is macOS first, with Windows and Linux in progress.

Would love for people here to try it. If it feels useful, that would mean a lot.

Happy to answer questions about continuity, session resume, and automations.

r/SideProject sebastianMR

Built a landing page for a 1% buyer's agent service — is the offer clear?

I'm a licensed real estate agent in Southern California and I built an offer-to-close service for homebuyers who've already found a home on their own. The idea is simple: if you found the property at an open house or through new construction, you shouldn't have to pay 2.5-3% for an agent you only need for the transaction itself.

We charge 1% and handle everything from offer to close. The rest of the buyer agent commission goes back to the buyer or is used to make the offer more appealing to the sellers.

I'd love honest feedback on the website — specifically:

- Does it make sense within the first 10 seconds of landing on it?

- Is the value proposition clear or confusing?

- Would you trust it enough to reach out?

- Anything that feels off, unclear, or missing?

www.onepercentoffers.com

r/SideProject Born-Comfortable2868

App Store Vs Play Store: Same App, different ASO Strategies.

Recently I shipped my first cross-platform app released on both App store & playstore. same description for both stores. copied it straight across. And my play store rankings were terrible while ios was doing fine.

Here's a breakdown on playstore vs App store ASO.

what apple indexes for search:

  • app name (30 chars)
  • subtitle (30 chars)
  • keyword field (100 chars)
  • developer name, in-app purchase names
  • screenshot captions (indexed as of 2025)

that's the complete list. your description, preview text none of it affects search ranking.

what google play indexes:

  • title (30 chars)
  • short description (80 chars)
  • full description (4,000 chars)
  • developer name

the full description is actively crawled and weighted. every word counts.

what this means in practice:

  • ios description = write for humans. it's a sales page for people already deciding whether to download. keyword density here does nothing for ranking
  • play store description = sales page and seo document at the same time. your primary keywords need to appear naturally, a few times across those 4,000 characters. not stuffed google flags that as spam
  • the short description on play store (80 chars, appears before the fold) is indexed and most developers leave it generic. it should have your primary keyword and a clear value proposition
  • screenshot captions on ios are now indexed if you want visible keyword placement on ios, that's where it goes

the workflow:
maintaining two completely separate metadata strategies gets messy fast. a few things that helped:

  • fastlane deliver (open source) manages your ios and play store metadata from versioned files in your repo. descriptions, keywords, screenshots, changelogs all checked into git, deployed with one command. once you accept that the two stores need separate files, fastlane makes that the default
  • Asc cli (open source) asc localizations list shows everything live across all your locales from terminal without touching the app store connect web ui. fits naturally if you're already running eas from terminal
  • vibecode Cli when keyword tests show a page isn't converting and the fix is a ui change, not just a metadata tweak, this is what i use to push fixes fast & build the app without breaking the release flow

so keep these difference in mind while shipping for both stores at the same time.

r/TheWayWeWere Electrical-Aspect-13

Children do an experiment to discover the natural laws of friction, San Jose Elementary School, Mission Hills, California, 1 of April 1953

r/LocalLLaMA jikkii

HF moves safetensors to the PyTorch Foundation

Hey local llamas, Lysandre from Hugging Face here.

Today we're officially moving Safetensors under the PyTorch Foundation, alongside PyTorch (of course), vLLM, DeepSpeed, Ray, and the recently-announced Helion. Concretely this means the trademark and repo are now held by the Linux Foundation rather than Hugging Face: neutral stewardship and open governance.

For local inference nothing changes today. Its the same format, same APIs, same Hub compatibility; we're working with the PyTorch team directly to see how to best integrate within PyTorch core.

What this unlocks is the ability to work more openly with the broader ecosystem on some further optimizations; more than a file format, there are some good opportunities for speedups across the board within the python/pytorch ecosystem: device-aware loading on different accelerators, tp/pp optimized loading, and of course new quantization/data types support.

We're currently refining our roadmap for the next few months/years and we'd be happy to work on it with you. Happy to answer questions about any of this, or the governance side.

PS: we wrote a blogpost here which has a few more details: https://huggingface.co/blog/safetensors-joins-pytorch-foundation

r/Anthropic mecharoy

Why does the use from extrage counts towards my weekly quota?

I saw that my weekly quota was increasing when my 5-hour window was 100%, although the usage was being used from my extra 'usage' where they charge you based on API cost. Is that expected or a bug?

r/homeassistant nsuitt

Is Cloudfare tunnel the most secure way to access HA ?

Hey everyone,

I'm finaly using cloudfare tunnel for external access, is it the most secure way considering that you have to find my domain name, my login, password and have the 2FA ?

r/ChatGPT Think-Score243

If I share my experience in well manner using ChatGPT, whats problem?

Recently I shared my experience for SEO as a comparison having grok, ChatGPT, Claude, Gemini in well format having icons and all, but the bot of reddit removed that post,

Is it a sin you use ChatGPT to create a post using well format?? Or is it required to clear English proficiency test to write post??

The legacy format of reddit should be revised.

r/ClaudeCode hazyhaar

Weird. Awesome claude today.

I have no good reason, no special action on my side.

I've hade some trouble on business coding, boring shits.

I'm juste using a regular CLI, with no claws, no IDE, no thing.

I've had some very bas days, seeing claude faking test to pass it, faking data to suceed a parsing, a true lazzy piece of crap.

But not today. today is back on a great thinking capability, perfect analysis, planning.

I think one thing changed: vibe coding and claw spamming for moron tasks who doesn't need llms use a lot of token for compensate the human lazyness.

Maybe it's some claude uses that has been nerfed by recalibrating the auto token spending.

I mean: Anthropic need users to get his models better. a pro max plan as I use worth approx 10k$ token a month on my use scale.

I would understand that using infering on a claude server for email automation would bed downgrading on pro plans. go api.

It's literraly sinking anthropic to make some more lazzys humans and more stupid datasets.

Claw madness and infering waste may be the true scope: today i'm mapping a whole runtime taint analysis in sqlite graph. This nasty claude is having huge yolo fun. Note a single mistake.

r/ChatGPT Ok_Expression7038

Mythos vs. Spud

You know that feeling when you're second to present in class, and the person before you absolutely crushes it, and now it's your turn?

Mythos just went first haha Good luck Spud.

r/SideProject guillim

Macos app to manage all my AI coding agents

Hey everyone 🙋‍♂️,

I run multiple Claude agents in parallel and I'd constantly lose track of which one was waiting for input, which one finished, and which one was still working.

There's no tool that monitors AI coding agents on your desktop. I use Cursor as well, so no tool gathers all agents at once.

So I built Glimpse.

It's a native menu bar app that gives each agent a small character on your desktop. The character animates based on what the agent is doing:

- Thinking/working — character is active

- Needs your input — orange dot appears in menu bar

- Done — character goes idle

That's the whole concept. No dashboard, no popups, no sounds.

It was built very light:

- 1.4 MB download, low CPU usage, almost no RAM consumption

- 100% on-device & no network calls, no account, no tracking

- Works with Claude in Terminal, iTerm2, and Cursor

- 7 character themes (Kawaii, Star Wars, One Piece, Dragon Ball Z, The Office, Marvel, Demon Slayer)

- macOS 13.0+

Current Status: Open-source and actively maintained. I have been using it daily for a week. I'd love to hear what features make sense to add!

short demo

Repo: github.com/guillim/Glimpse

r/whatisit Sea_Rip_4008

Inherited from my grandmother

Hello!

I've inherited this ring from my grandmother after her passing. All I know is it's diamond and sapphire stones.

There are no hallmarks to identify where it's from, although it is stamped with '9CT GOLD' on the inside. There is also an etched 'EC432' inside however I can't find out anything about it's potential significance.

Whole family is clueless, apart from my grandmother - but she's currently not around to ask, being a pile of ash and everything.

Can anyone tell me it's age/where it may have come from/absolutely anything at all?

Thank you!

r/Anthropic Extreme-Tie9282

anyone lose their Claude overage credits!?

lost 150$ in credits when the 7pm weekly rest happened. anyone else?

r/whatisit timmy2plates

What’s this fluff in my cellar ceiling?

It only occurs in this one corner.

r/mildlyinteresting Jazzel5

found plants growing inside a wall

r/VEO3 HyperChromeFlux

Aerial Duel | Energy Blade Clash | Hyper-CGI Cinematic #cyberpunk #vfx #Seedance 2.0

r/ChatGPT Historical-Let-1211

Will ChatGPT go give me lots more image generation?

I get like 3 with free, would go give me maybe 20 per day? thank you!

r/Seattle HeyJesikaJay

Masci's out again 4/8, charges dropped

r/LocalLLaMA vasimv

Idea - Predict&Compare agent to make model act smarter

I've got an idea while i was watching small local model on limited VRAM trying to develop and debug a simple android game test project, and how it was going again and again through same sequence "i try tap... it didn't work, may be tap somehwere else?... may be use uiautomator?..". What if make an agent that would ask the model to make predictions and compare it with actual results? Basically, how humans do often when they try to do something.

flowchart

The agent asks additional question (prediction) and stores the prediction in an indexed database (actually, can be omitted in case of simple one-threaded conversations), then asks model to compare results from the generated tool call and its own prediction. Comparison results is stored into another indexed database (or just simply injected into next prompt) to be used later.

This method could be used not just to improve tool calls but for other stuff to, though requires a feedback loop of some sort (like asking user "Did you tried that, was that useful?" after generating a hint for his problem). May be even multi-level predictions database could be made for full cycle generate code -> "what do you expect this code to do?" -> build&test -> "Did that code work as should?".

Also, past experience database can be used to retrain model to perform better later.

r/SideProject stoopapoop

Built an app after a miserable drive from Chicago to Atlanta. Curious if others have the same problem

A few years ago I was driving I-65 from Chicago to Atlanta. Somewhere near Elizabethtown, Kentucky I needed gas, food, and a bathroom. I stopped at a random exit, got mediocre gas station food, used a questionable bathroom, and drove on.

Twenty miles later I passed what would have been the perfect exit. A freakin' Buc-ee's with everything I needed. But I'd already stopped.

That kept happening the whole drive since I was driving alone and had no one to help find somewhere semi-decent to stop. I'd stop somewhere pretty mediocre because I didn't know what was coming. There was no way to know which exit was actually worth stopping at versus which ones I should drive past.

Since then, I have spent a few years building an app/branding/algorithm that solves exactly this. It's called Kibi. You enter your destination, it calculates where you need to stop based on your vehicle's range, and it finds the single best exit for gas, food, and a bathroom together...not three separate stops.

It covers 79,000+ US highway exits leveraging OpenStreetMaps scored by rating and amenity quality. Free on iOS, no ads.

r/Jokes Ochevesako

What do you call an attractive baseball player?

Striking

r/meme Stock_Crazy6759

Lamo 😂

r/AI_Agents ChipperChick

Seeking Transcriber

I need to extract the words from a long detailed video to make learning scripts and instructional video from it. any good tools where I can click and drag a video into and it transcribes everything said word for word with high accuracy?

r/me_irl MashedPotatoSuperFan

me_irl

r/Seattle Old_Effect9225

Thoughts on Yes Energy Management?

I live in Mountlake Terrace, which is considered slightly more affordable than Seattle. However, my utility bills have increased each month for the past 8 months. I tried to contact them to ask about this discrepancy, but they informed me that they are ”not allowed” from disclosing this information and that I must speak with my landlord. At the same time, I am required to pay any amount they specify.

It’s worth noting that there are approximately 100 complexes, but only three dumpsters are available. Despite these, I enjoy the place. However, I am moving once my lease expires because of these issues.

r/mildlyinteresting GrahamPhisher

I've been tracking the daily sentiment of 24 major news sources across the political spectrum, and today is the first day in a month the news has been positive.

r/homeassistant fleetmack

SDR with rtlamr2mqtt to read meters

I am really struggling trying to get my gas meter, water meter, and electric meters read. I know I have smart meters as I have Eye On Water, and it works, but it updates so infrequently. Also I have my electric meter set up thru Xcel already. But I'm playign with this Nooelec v5 SDR reader and can't seem to get it to work.

I am running HASS on a Synology 423+ in a VM. I have the Nooelec passed thru to the VM via usb. The antenna is a window in clear view of the meters.

I have MQTT up and running and have the rtlamr2mqtt running. Those 2 are talking to each other, I know, becuase if I put in a dummy meter it creates an MQTT device. But no matter what I do to my yaml, my log file in rtlamr2mqtt NEVER finds a meter.

I live in a dense neighbohood in Denver. I know the SDR reader works as with SDR++ app I can tune radio stations. I have the 900mhz antenna attached. I simply can't get readings of any meters, which is odd.

Anyone else run into this? I've used Gemini and ChatGPT to exhaustion and found nothing.

r/nextfuckinglevel Ashish_ank

Man on safari records his close encounter with a Cheetah

r/ClaudeAI MetaKnowing

Claude Mythos can generate novel puns

r/SideProject Worldly-Entrance-948

I built a system tray monitor for Claude Code, colored dot when any session needs input

Keep coming back to find Claude Code has been sitting at a confirmation prompt for 15 minutes. Flow broken, context cold. So I built something.

AgentTray is a system tray dot: green when your agents are running, yellow when something needs input, red when something failed. Yellow also beeps and sends a desktop notification.

Rust + Tauri 2.x + SvelteKit + Svelte 5. Linux daily driver, Windows/Mac builds exist but I haven't tested them properly. A few days old, rough edges, works for me.

https://github.com/sprklai/agenttray

r/aivideo Mainerville

THE RED DEATH MUST FALL|赤き死神、ここに滅ぶ – Takaichi × Kada

r/ClaudeAI Perfect-Process393

Xcode Claude Agent "Sign in with claude.ai" not working

Hi, I have a problem to sign in with claude.ai account for the xcode agent and maybe somebody had a similar problem and found a way to fix it. When I click on "sign in with a claude.ai account" the authorization dialogue pops up in my browser. For context my Mac and Xcode are both updated to the latest version and I have a valid Claude Subscription. When the pop up comes and I click "authorize" one of those things happen: I get the "build something great, you are all set up" message in my browser but in apple intelligence settings the wheel spins for 15 minutes and it goes back to "not signed in". Or the pop up disappears when I click on "authorize" and it immediately goes back to "not signed in". Or I get the error message "something went wrong" in Safari. If someone had the same problem and found a solution I would be very thankful for some advice.

r/mildlyinteresting El_Morgos

Went to the dentist. Some substance sprayed and bleached my hoodie.

r/AbruptChaos MisterShipWreck

Throwing a firecracker into the sewer? Not just a harmless prank!

r/ChatGPT Obvious_Shoe7302

Silicon Valley tech bros in a nutshell

r/meme nrt12345

Time Flies...!

r/whatisit subjectiveyes

Seen in the sky over West Chester, PA USA (zoom in)

I know there are always random contrails in the sky but I have seen continued little trails like this in the same spot of the last hour. I also saw one this morning in the same direction. Right now it's almost 1 pm. There's nothing on flight radar app so it doesn't appear like flights ??? Whats the deal? Is it just something about an exceptionally clear day????

Edit: this is in a slightly Northeast direction! Thanks !!!?🙏

r/PhotoshopRequest jonnyd005

Recreate these movie covers for the last two movies.

These are movie covers for the first 6 Mission Impossible movies all of which have the same design theme. The last two movies, "Dead Reckoning" and "The Final Reckoning", were made with completely different art styles and it bothers me. I would like for someone to recreate this style art over a transparent background instead of silver like you see. They should have the two tone color splatter in the back, with Cruise taking up top to bottom over that, and then the name and title in the middle as you see, finally the collage of action things at the bottom. Most assets can be found at themoviedb.org like posters to pull required images from. I actually tried starting this myself for "Dead Reckoning" but just couldn't get the bottom collage to really work well so I do have some assets ready if needed. Will pay well for this request.

r/screenshots Puns4Life_

femboys..

r/singularity fortune

Sam Altman and Vinod Khosla agree: AI will break the economy. Their fix is no income tax for most Americans

When Vinod Khosla sat down with Fortune editor-in-chief Alyson Shontell in March and floated the idea of wiping out federal income taxes for the roughly 100-million-plus Americans earning less than $100,000 a year, it sounded like the kind of provocation only a billionaire with nothing left to prove could get away with. “I can’t be fired. I’ve never worried about a career. I don’t need more money at age 71,” Khosla said.

A month later, OpenAI has made it clear that Khosla’s thinking may be the emerging consensus of Silicon Valley’s most powerful voices on how to prevent artificial intelligence from tearing the social fabric apart.

On Monday, OpenAI released a 13-page policy paper titled Industrial Policy for the Intelligence Age: Ideas to Keep People First, in which Sam Altman’s company laid out a sweeping blueprint for economic reform on a scale it compared to the Progressive Era of the early 1900s and Franklin Roosevelt’s New Deal of the 1930s. The overlap with Khosla’s vision is hard to miss.

Read more: https://fortune.com/2026/04/07/sam-altman-vinod-khosla-openai-tax-code-american-income-tax-100k/

r/ChatGPT Capable-Ad-1345

I told my chat to roast me 😭

mod, if this is not allowed feel free to remove.

r/LocalLLaMA petroslamb

"LLMs drop the wiring even when they keep the scene", A destinct failure mode is the binding gap

(I’m posting this here natively (previously shared on r/llm) as I think this community will appreciate the mechanical distinction between reasoning and representation.)

I’m particularly interested in whether this "Binding Gap" explains why the 8B-tier models tend to fall apart on relationship-heavy extraction tasks relative to their high benchmark scores.

So here goes:

I've been tracking a failure pattern across several recent papers that I have not seen named clearly enough to be useful.

Tell a model "Tom Smith's wife is Mary Stone." Then ask "who is Mary Stone's husband?" Nothing new has been added. Same entities, same relation. Performance still drops.

This example was very true for GPT-2 and it seems to still hold at scale for GPT-5.

Wang and Sun showed this on the Reversal Curse and argued it reflects real problems in how transformers bind concepts in representations. They showed architectural changes aimed at better binding can reduce the failure. That moves it from quirky benchmark effect toward a design problem.

I think this extends well beyond the reversal case. Call it the binding gap: a model preserves enough semantic material to generate a plausible answer but loses the specific attachment that makes it faithful.

The evidence that this is its own family of failures is accumulating.

Feng and Steinhardt showed models use internal binding ID vectors and that causal interventions on those activations change pairing behavior. Dai and colleagues identified a low-rank ordering subspace that directly steers which attribute gets bound to which entity. Denning et al looked at thematic role understanding and found agent/patient info influences sentence representations much more weakly than in humans.

The mechanism is real and manipulable. Then the heavy-load case.

Tan and D'Souza tested on meta-analysis evidence extraction with GPT-5.2 and Qwen3-VL. Single-property queries land around 0.40-0.50 F1. Full association tuples with variables, methods, and effect sizes drop to ~0.24 and near-zero in dense result sections. Role reversals, cross-analysis binding drift, numeric misattribution. This is what the binding gap looks like under actual research pressure.

Does this feel like a useful frame for evaluation design, or is it covered by entanglement and compositionality already? Either answer saves me time.

Ive got a clear writeup in the rooted layers blog, if you need more details.

r/me_irl Botched_Rapture

me_irl

Seeing more and more of this!

r/Jokes bigus-_-dickus

a mailman was delivering mail on Christmas Eve

on his route, a beautiful middle aged woman stops him and says: "can you join me inside the house real quick?"

the mailman was intrigued and follows her inside

she takes him to the bedroom and starts taking her clothes off and they have sex

after it's over, he gets dressed to leave and she hands him a 5 dollar bill and says thank you

the mailman was confused, he asks: "what was all this for?"

she says: "my husband and i were discussing Christmas gifts and i asked him if we should give anything to the mailman and he said:

"ahh fuck him, give him 5 dollars"

r/mildlyinteresting OkGood3000

There is a seperate line to flights to USA and Isreal VS all other flights

r/mildlyinteresting JoinMyGuild

The grime in my keyboard is concentrated around the WASD keys

r/arduino MacksMR

My first encounter with Arduino.

My first encounter with Arduino started with a watch. I was lucky enough to buy 12 IV-11 indicators for 600 Russian rubles (7-8$) - a real gift. That's how it all started. I quickly wrote some code and got them running. I liked the result, but I plan to redesign them. The schematic, the boards, and the code are terrible.

r/confusing_perspective whywehumans

Long neck?

r/whatisit Old-Way-724

Does anyone know why this jacket has button holes on both side at the top but no buttons?

I thrifted this jacket a while ago without noticing this weird detail and was wondering if anyone knows what this is for or why it’s like this?

r/LocalLLM Yssssssh

Glm-5.1 claims near opus level coding performance: Marketing hype or real? I ran my own tests

Yeah I know, another "matches Opus" claim. I was skeptical too.

Threw it at an actual refactor job, legacy backend, multi-step, cross-file dependencies. The stuff that usually makes models go full amnesiac by step 5.

It didn't. Tracked state the whole way, self-corrected once without me prompting it. not what I expected from a chinese open-source model at this price.

The benchmark chart is straight from Zai so make of that what you will. 54.9 composite across SWE-Bench Pro, Terminal-Bench 2.0 and NL2Repo vs Opus's 57.5. The gap is smaller than I thought. The SWE-Bench Pro number is the interesting one tho, apparently edges out Opus there specifically. That benchmark is pretty hard to sandbag.

K2.5 is at 45.5 for reference, so that's not really a competition anymore.

I still think Opus has it on deep reasoning, but for long multi-step coding tasks the value math is getting weird.

Anyone else actually run this on real work or just vibes so far?

r/ChatGPT deadwhiskers420

Seedance 2.0 access

Any way to get access?

r/mildlyinteresting FedorKorovkin

[OC] Found an ancient astronaut at the beach

r/StableDiffusion JournalistLucky5124

Any realistic and decent img edit model thai i can run on 4gb vram and /or 16gb ram??

r/ClaudeAI jradoff

I gave Claude my dead game's 30-year-old files and asked it to bring the game back to life

In 1992 I built an online multiplayer game called Legends of Future Past. It ran on CompuServe, won an award from Computer Gaming World, and shut down on the last day of 1999. I was 19 when I made it.

The source code didn't survive. What I did have: hundreds of script files written in a little language I'd invented for Game Masters, a GM manual I wrote in 1998, and a gameplay recording from 1996.

I gave all of this to Claude Code without much instruction beyond "figure out what this scripting language does and rebuild the game." What I got back genuinely surprised me.

Claude reconstructed the grammar of a programming language that has never existed anywhere outside my game servers. No documentation on the internet, no Stack Overflow answers, no training data. It inferred the rules from the scripts themselves and a manual I'd written for non-technical GMs.

Then it rebuilt the entire game — 2,273 rooms, 1,990 items, 297 types of monsters, 88 spells, a full crafting system, combat mechanics. A world that took me months to build originally was reconstructed in a weekend.

The part I keep coming back to: this isn't Claude doing something it was trained to do. Nobody trained it on my scripting language. It did what a skilled human reverse-engineer would do — read examples, find patterns, build a mental model, and test its assumptions. It just did it in hours instead of weeks.

The game is free to play at lofp.metavert.io and the code is open source at github.com/jonradoff/lofp. I wrote up the full technical story here if you want the deep dive.

r/comfyui JournalistLucky5124

Any realistic and decent img edit models that I can run on 4gb vran and/or 16gb ram?

r/LocalLLaMA Gailenstorm

[Tool] Quick hack to recover Qwen3.5 MTP after fine-tuning for faster inference speed (Transformers)

Disclaimer: I work at NuMind (we train LLMs for structured + content extraction).

If you've been working with Qwen3.5 (and other recently released models), you probably know it includes **Multi-Token Prediction (MTP)** modules. When used with vLLM (`qwen3_next_mtp`), this can significantly speed up inference, especially on predictable workloads (the more "predictable" the better since the draft tokens will have a higher acceptance rate).

However:

- Hugging Face Transformers doesn’t support MTP yet, neither for inference nor training

- Thus, if you fine-tune with `Trainer`, MTP weights are never loaded, trained, or saved

- Result: vLLM crashes when you try to use speculative decoding (`using --speculative-config '{"method":"qwen3_next_mtp","num_speculative_tokens":4}'`) because the weights are missing

### Quick workaround

Not perfect, but works: You can just **copy the MTP weights from the base model into your fine-tuned model**.

* The MTP heads remain untrained

* But in practice, it’s still useful

The code is simply something like

for filepath in path_source_model.glob("*.safetensors"): with safe_open(filepath, framework="pt", device="cpu") as f: for key in f.keys(): if "mtp" in key.lower() or "nextn" in key.lower(): mtp_weights[key] = f.get_tensor(key) save_file(mtp_weights, out_filepath) 

and then updating the `model.safetensors.index.json`

In our internal tests:

* Acceptance rate up to ~0.9 up to ~4 tokens

* Highly workload-dependent however

For our larger models and future open weights model, we will however include all the heads during the training in order to improve efficiency/acceptance rate. We have patched transformers to support it and hopefully in the future it will be available for everyone.

### Tool

I made a small CLI to do this automatically:

https://github.com/SorenDreano/transplant_mtp (MIT)

Tested on Qwen3.5 models.

### Context (what we’re building)

We have released open-weight models for document understanding:

**NuExtract 2.0**: structured extraction into JSON templates

https://huggingface.co/numind/NuExtract-2.0-8B

NuExtract is a model that takes both a json template input like

{

"Last name": "verbatim-string",

"First names": [

"verbatim-string"

],

"Document number": "verbatim-string",

"Date of birth": "date-time",

"Gender": [

"Male", "Female", "Other"

],

"Expiration date": "date-time",

"Country ISO code": "string"

}

and a document (usually an image or scan) and fills the template with correct information without hallucination.

**NuMarkdown**: convert documents (images, PDFs, text) into (you guessed it) Markdown

https://huggingface.co/numind/NuMarkdown-8B-Thinking

We are soon going to release a new open weight model that does BOTH structured (json template) AND content (markdown) extraction

We also have a SaaS offering and can deploy on premise https://nuextract.ai

Curious if others have tried different approaches to keep MTP during fine-tuning or if anyone has patched Transformers to support it properly.

r/personalfinance yiggity_yag

I have no idea if building a home is affordable or the worst financial mistake I can make

I’m 36M, wife is 34F. I make ~$230k/year, she makes ~$40k. We own a home we bought for $230k in 2020 that’s now worth ~$350k, with ~$163k left on the mortgage. No other debt.

I’m maxing 401k, Roth IRA, and HSA, plus ~$1k/month to brokerage and $1k/month to savings. We also contribute $800/month total to 529s. We have ~$60k in a HYSA (emergency fund + potential down payment).

Expenses are relatively low:

  • Mortgage: ~$1,250/month
  • Utilities: ~$250
  • Internet/streaming: ~$120
  • Groceries: ~$400
  • Takeout: ~$200
  • Insurance: ~$120 home + ~$1,200 every 6 months for 3 cars
  • Childcare: $0 (family help)
  • Vacations: $5,250/year

We don’t strictly budget, so I'm sure I'm missing somet things, but overall we save/invest aggressively. We have one guilty pleasure of Disney World vacations which we spend $4k/year on for annual passes for our family of 4, and about $1,250/year in annual dues for a DVC timeshare.

Here’s the situation:

My in-laws own an empty lot next to their house and have offered to give it to us for free. It’s in an established subdivision, and there’s already a 3-car garage with electricity on the lot (father in law uses it as a shop--it's pretty pushed back on the property and would not be used as the garage for any newly built house).

On paper, this feels like a huge opportunity. But the process of building is overwhelming.

From what I understand, we’d need to:

  • Get land evaluated (grading, septic, etc.)
  • Work with a builder to determine what’s possible
  • Take out a construction loan to fund the build

What I don’t understand is how this actually works financially while we still own our current home.

Do we:

  • Carry our current mortgage (~$1,250)
  • While also paying on a construction loan?

I’ve read about interest-only construction loans or bridge loans, but also horror stories about cost overruns and banks refusing to extend funding.

In our area (Wisconsin), a 4BR home built in the last ~20 years runs ~$500k–$600k. People I know who have built say to expect $100k–$200k more than buying existing.

So even with “free land,” we might still be looking at ~$600k–$700k total.

That leads to a few big questions:

  1. Where do you even start with this? If I were buying a house, I’d go to a bank, get pre-approved, and shop within budget. With building, it feels like I’m committing to a process without knowing the final cost upfront.
  2. How do people handle overlapping housing costs? Ideal scenario would be interest-only payments during construction, then sell our current home once we move in. But I don’t want to sell first and live with my in-laws for a year (2 kids, pets, WFH).
  3. Does “free land” actually offset the higher cost of building? Or is building still significantly more expensive than buying, regardless?
  4. What’s the real incentive to build? Is it just customization and location? Being next to grandparents is appealing, but $100k–$200k extra feels like a big premium.
  5. Affordability concern If I assume a worst-case drop in income back to ~$100k (contract work uncertainty), I want to stay conservative.
  • ~$500k home with ~$150k down → ~$2,600/month (doable)
  • ~$700k home → ~$3,800/month (feels risky, likely cuts investing heavily)

Overall, this feels like an amazing opportunity, but also a massive financial and logistical leap. I’m trying to figure out if this is actually a smart move or just something that sounds good because the land is free.

Would really appreciate advice from anyone who’s gone through building, especially around financing and whether this makes sense given our situation.

r/SideProject e-joculator

Web Component UI Kit

About a year ago, I published a pure web-component-based UI kit and posted about it here.

It was born of a web-based game side project I had been working on for a bit, and I came to enjoy how easy it was to implement and started using it in a few of my other projects. Eventually I published and posted about it to try to get feedback.

Ultimately it ended up serving as more of a proof-of-concept than an actual useful UI kit. The look and overall finish quality didn't ultimately fit the need for any outside projects I was working on, only some niche applications (like my game and my personal portfolio site), at least not without heavy customization.

However, these shortfalls were all in my design. Given how much I enjoyed using web components, I still wanted to be able to use this kind of UI kit in a broader set of projects. Applying my new-found knowledge, I started from scratch building a UI kit based on a brand new design system. I kept all of what I loved about Kepler UI in terms of it's easy implementation and easy theming, and brought it up to modern standards in terms of design, accessibility, and implementation.

I call it "Yumekit". You can check it out here:

https://yumekit.com

The github repo can be found here:

https://github.com/waggylabs/yumekit

It's incredibly easy to use and install. A single JS import will provide you with all of the elements, including their styles, ready and able to be used in any application. It works out of the box in plain vanilla HTML/CSS/JS, or can be dropped into any framework including React, Vue, Svelte, and Angular with little to no configuration needed. It has zero outside dependencies, and you can import only components that you need, allowing it to be very light weight.

Styling is also really easy. You wrap any components you want to style in a `` component and can easily switch out the theme by using one of several built-in styles or by providing a custom theme style sheet.

It should be noted that this project is still in Beta, so bugs and inconsistencies are expected, but please feel free to open up new issues on github if you spot anything.

I've also created an FAQ section and will add to it as more Qs are FA'd.

Let me know what you think!

r/LocalLLaMA Excellent_Koala769

Gemma 4 26B MoE vs 31B Dense as daily driver for OpenClaw on M5 Max 128GB?

Hey Guys,

Running OpenClaw locally on my M5 Max MacBook Pro with 128GB unified memory. Which Gemma 4 model is better as the main daily driver — the 26B MoE or the 31B dense?

The MoE is way faster, but I’m worried about expert routing causing inconsistency in tool calling and agentic tasks compared to the dense model.

Anyone who’s tested both in real OpenClaw use on Apple Silicon: which one are you actually using day-to-day and why? Is the MoE consistent enough or is the 31B noticeably more reliable?

Thanks!

r/Damnthatsinteresting Ashish_ank

Man on safari records his close encounter with a Cheetah

r/SideProject Murmoosey

I built a pregnancy admin app during paternity leave - free at launch, price switch coming after further feedback (UK)

Solo non-technical dad, first app. Built ParentPA during paternity leave after watching my partner drown in NHS appointment tracking while I had zero visibility.

It's a 51-card (and counting, after have builds qued up) timeline for pregnancy through the first year - scans, registrations, vaccinations, employer deadlines — all NHS/GOV.UK sourced, timed to your dates, shared with your partner. Covers England, Scotland, Wales and NI properly (each nation has different terminology, escalation routes, source links).

No subscription. Free at launch for UK parents, pricing TBC one-off after the free window.

Stack: Expo React Native + Supabase. Built with Claude Code.

App Store: https://apps.apple.com/app/parentpa/id6744228353

Landing page: https://parentpa.uk

Would love feedback from builders on anything you'd approach differently, and from parents on the product itself.

r/personalfinance and_20

recurring investments fidelity roth ira

if i wanted to invest 70% of my fidelity roth IRA in VOO (for example), what dollar amount should i make my recurring investments be? would it be 70% of whatever i contribute to my roth IRA each paycheck? so if i transferred $200 to my roth IRA every paycheck i would make a recurring investment of $140 to VOO? thanks!

r/ChatGPT FalsePresentation756

How hard is it to switch from Claude Pro to ChatGPT Plus?

I’ve been getting pretty frustrated with Claude’s usage limits lately, so I’m thinking about switching to ChatGPT Plus since the pricing is pretty similar.

For those who’ve made the switch, how difficult is the transition? Is using Codex comparable to Claude Code in terms of workflow and capabilities?

Also, when working with a codebase, can I just drop it in and start working, or are there specific tools or workflows I should be using to do the transfer?

And one more thing, how do ChatGPT Plus limits compare to Claude Pro these days? I've heard they're not that easy to hit. Is it just as easy to hit the cap?

r/ChatGPT HSVMalooGTS

"Can't display this data due to safety concerns"

It replied like normal, but slow. Now it jumped to that. Feels like deepseek.

r/whatisit fubarforever

Looking to ID this cover-up

Good friend of mine lost this cover-up, and I would like to surprise them with a new one as a gift. If anyone knows what company made it, I would be greatly appreciative.

r/nextfuckinglevel dariustwin

Painting dinosaurs with light

r/SideProject EstebanbanC

How do you cover costs of a side project?

Hey, I'm about to launch a side project and trying to figure out a sustainable way to cover its running costs (nothing crazy, just a few hundred dollars a year).

The site is an AI-powered RSS aggregator focused on tech niches (AI, Crypto, Cybersecurity). It fetches articles from dozens of sources, extracts trending keywords and displays them as a word cloud, generates TL;DRs for each article, and writes a short "why is this trending?" explanation for each hot keyword.

Costs are modest: VPS, domain, and LLM API calls.

I'm considering:

- Carbon Ads

- Premium features: historical data, custom alerts, CSV export, ...

- Referral links: crypto exchanges for example

- Ko-fi / Buy me a coffee

I'd rather avoid traditional display ads if possible. Has anyone found a good balance for a niche content site like this?

r/SideProject roelvroozendaal

I just shipped Fitness First Rep Counter a 1.99 watchOS app that counts gym reps, under 1 MB, zero data collection

After months of development and testing, I finally shipped my side project: Fitness First Rep Counter for Apple Watch.

The problem: Losing count during sets at the gym. Sounds small, but if you train regularly it's genuinely annoying, especially on higher rep sets or when you're supersetting exercises.

The solution: A watchOS app that uses the built-in accelerometer and gyroscope to detect reps in real time. But I didn't want to just count. I also added a rest timer with haptic alerts and live heart rate monitoring, since knowing when you're actually recovered matters more than a fixed rest clock.

Design philosophy: Do one thing well. No workout planning, no social features, no coaching, no AI nonsense. Just count reps, time rest, track heart rate, and log to Apple Health. The entire app is 918 KB.

The tricky part was getting the motion detection to generalize across different exercises. A push-up looks nothing like a bicep curl in terms of sensor data. I also added a manual mode toggle for people who prefer tapping, or for exercises where auto-detection isn't great yet.

Some things I learned building for watchOS:

  • Battery life is a real constraint; every sensor read costs power
  • Screen size forces ruthless UI decisions; the big number display was born out of necessity
  • Real gym sensor data is way messier than lab demos
  • Dogfooding a fitness app means you end up working out a lot more than planned

Business model: $1.99 one-time purchase. No subscriptions, no in-app purchases, no ads. Supports Family Sharing (up to 6 people). Zero data collection.

App Store: https://apps.apple.com/us/app/fitness-first-rep-counter/id6760266411

Would love any feedback

r/PhotoshopRequest One_Leadership145

I would like if someone could take away the background

I've been trying to get rid of the background for a while now but I've been unsuccessful no matter what I do so I thought I should ask people here

r/mildlyinteresting Hey-Im-Sorry

Customers Indoor Beehive

r/whatisit iamthekiller4u

Girlfriend sent me this picture of something on the steps leading to her loft.

Said it's about 2 inches long.

r/LocalLLaMA Few_Novel_2936

I ran a MiroFish simulation: 500 AI agents debating whether AGI arrives by 2027. Hassabis says yes. LeCun says "very impossible." Here's what emerged.

Been curious about MiroFish since it hit #1 on GitHub trending.

Finally ran my own simulation instead of just reading about it.

Topic: Will AGI meet a reasonable definition by end of 2027?

Seed material: Structured briefing with positions from Altman,

LeCun, Hassabis, Yudkowsky, Gary Marcus + prediction market data

Agent count: 500

Model: GPT-4o mini

Key findings from the report:

- Experts split sharply — Hassabis and Altman optimistic,

LeCun and Marcus explicitly said "very impossible"

- Agents converged on ethics and safety as the #1 concern

regardless of which side they were on

- Most surprising: even pro-AGI agents were pessimistic

about societal readiness

- Regulatory pressure emerged as the biggest wildcard —

more than technical limitations

Graph visualization and report screenshots in comments.

Happy to answer questions about the prompt setup or seed material I used.

r/whatisit NightHawkUndead

Found in the woods.

Found these in this woods.

Area: Chattanooga, TN

Quantity: x4 in a general area

Size: best guess 12 foot in length

Google or any other assistants are no help.

r/VEO3 Time_Blackberry2089

Every video failing due to "couldn't generate audio"

I recently rejoined google flow as a pro user and having every video fail "We couldn't generate audio for this prompt. You haven't been charged—please try tweaking your request." I even tried using google veo3.1 from runway and it failed. Is this common for everyone right now? Trying different stills and different prompts but still getting this...
Thanks

r/ChatGPT Revolutionary_Low_90

😭😭😭

r/PhotoshopRequest Slayberham69420

Flyaway hairs ruining my life

Hi! I’m trying to sweet talk my way into med school and had some headshots taken last week. However, it was windy that day and my hair didn’t want to cooperate :( can someone please fix the flyaway hairs in these? In the second picture, my hair got kind of messed up when I put it over my right shoulder and I was also hoping to get that fixed as well! Can pay $15 for both! Thank you in advance!

r/ClaudeAI TaskSpecialist5881

How to save 80% on your claude bill with better context

been building web apps with claude lately and those token limits have honestly started hitting me too. i’m using claude 4.6 sonnet for a research tool, but feeding it raw web data was absolutely nuking my limits.

i’m putting together the stuff that actually worked for me to save tokens and keep the bill down:

switch to markdown first. stop sending raw html. use tools like firecrawl to strip out the nested divs and script junk so you only pay for the actual text.

don't let your prompt cache go cold. anthropic’s prompt caching is a huge relief, but it only works if your data is consistent.

watch out for the 200k token "premium" jump. anthropic now charges nearly double for inputs over 200k tokens on the new opus/sonnet 4.6 models. keep your context under that limit to avoid the surcharge

strip the nav and footer. the website’s "about us" and "careers" links in the footer are just burning your money every time you hit send.

use jina reader for quick hits. for simple single-page reads, jina is a great way to get a clean text version without the crawler bloat.

truncate your context. if a documentation page is 20k words, just take the first 5k. most of the "meat" is usually at the top anyway.

clean your data with unstructured.io. if you are dealing with messy pdfs alongside web data, this helps turn the chaos into a clean schema claude actually understands.

map before you crawl. don't scrape every subpage blindly. i use the map feature in firecrawl to find the specific documentation urls that actually matter for your prompt, if you use another tool, prefer doing this.

use haiku for the "trash" work. use claude 4.5 haiku to summarize or filter data before feeding it into the expensive models like opus.

use smart chunking. use llama-index to break your data into semantic chunks so you only retrieve the exact paragraph the ai needs for that specific prompt.

cap your "extended thinking" depth. for opus 4.6, set thinking: {type: "adaptive"} with effort: "low" or "medium". the old budget_tokens param is deprecated on 4.6. thinking tokens are billed at the output rate, so if you leave effort on high, claude thinks hard on every single reply including the simple ones and your bill will hurt.

set hard usage limits. set your spending tiers in the anthropic console so a buggy loop doesn't drain your bank account while you're asleep.

feel free to roast my setup or add better tips if you have thembeen building web apps with claude lately and those token limits have honestly started hitting me too. i’m using claude 4.6 sonnet for a research tool, but feeding it raw web data was absolutely nuking my limits.

r/PhotoshopRequest superib9006

Face swap request

Hi! I’d like to do a face swap for these photos. Please replace my face/hair in the 1st photo with my face/hair in the 2nd one.

Will tip $5. Thanks

r/ChatGPT n8n7r

Ha! I’ll take it.

r/SideProject Such-Maintenance5569

Exploring AI website chatbots (Chatbase, KnowChat, and what’s missing in most tools)

I’ve been working on something recently and it changed how I look at most websites.

Most websites don’t have a traffic problem.

They have a conversion leak.

People visit.
They’re interested.
But they don’t find what they need fast enough.

So they leave.

And most of the time, you never know who they were.

A common pattern I keep seeing:

Businesses rely heavily on contact forms.

But forms feel like effort.

If I land on a site and have to think too much before reaching out, I usually just leave.

What works better is conversation.

When users can instantly ask things like:

  • pricing
  • how it works
  • requirements

They stay longer and engage more.

Another thing:

A lot of teams spend most of their time answering the same repeated questions:

  • location
  • availability
  • basic details

Which means support time goes into repetition instead of actual sales.

This is why a shift is happening:

From static pages + forms
To real-time conversations + intent-based capture

Where the system waits for signals like “pricing” or “demo” before asking for contact details.

There are already tools exploring this space.

For example, tools like Chatbase and KnowChat are being used to handle website conversations, answer queries, and capture leads.

Different tools tend to focus on slightly different strengths:

Some are stronger at:

  • general chatbot deployment across sites
  • handling knowledge-base style Q&A at scale

Others focus more on:

  • intent-based lead capture (knowing when to ask for contact info)
  • simpler setup for smaller websites
  • tighter flow from chat → actual leads

What are your opinions on these all ?

r/LocalLLaMA NovelAdorable7033

Gemma 4 26B achieves 40k context window

Hybrid KV Compression for Extending Context Length in vLLM

Abstract

We present a practical optimization framework for vLLM that significantly reduces KV cache memory usage while extending the effective context length of large language models.

The method introduces a hybrid KV cache structure that selectively compresses older KV blocks into INT4 while preserving recent KV blocks in full precision.

By combining block-level cache management, controlled restore–recompression scheduling, and a stability-aware context limiting strategy, the system achieves long-context inference without memory overflow or observable quality degradation.

On a single NVIDIA RTX 4090 (24GB), the method sustains a stable memory plateau while extending context length beyond 30k tokens and reaching up to ~40k tokens under stress testing.

  1. Introduction

Large language models are fundamentally constrained by the memory footprint of the KV cache during inference.

As context length increases, KV cache memory grows linearly, quickly exceeding available VRAM on consumer hardware.

Existing approaches either reduce precision globally or introduce approximate attention mechanisms, often at the cost of output quality or system stability.

This work proposes a practical alternative: selectively compressing only the older portions of the KV cache while preserving recent tokens in full precision.

This allows significant memory savings without degrading the model’s ability to attend to recent context.

  1. Method

2.1 Hybrid KV Cache Structure

The KV cache is divided into two regions:

Recent region: Maintained in floating-point precision (FP16/FP8)

Old region: Compressed into INT4 at block granularity

This hybrid structure ensures that high-sensitivity recent tokens remain accurate, while older tokens are stored in a memory-efficient form.

2.2 Block-Level Cache Management

Instead of token-level operations, the system manages KV cache in fixed-size blocks.

This design provides:

Reduced overhead for compression/decompression

Efficient tracking of processed regions

Stable memory behavior across long sequences

Each block is assigned a state:

new: recently added, not yet processed

old: eligible for compression

processed: already compressed and tracked

2.3 Restore and Recompression Control

Compressed KV blocks are restored to higher precision when required for attention computation.

To prevent performance degradation, the system enforces:

No immediate recompression after restore

Lazy recompression scheduling

Explicit tracking of processed blocks to avoid redundant operations

This avoids oscillation between compression and restoration.

2.4 Stability-Aware Context Limiting

A safe operating region is empirically determined to prevent instability at extreme context lengths.

The system restricts active context to a validated margin (e.g., ~3.5k tokens before instability thresholds), ensuring consistent runtime behavior.

2.5 Runtime Optimization

Several low-level optimizations are applied:

Removal of .item() calls to eliminate CPU synchronization overhead

Moving sequence length handling to CPU to simplify control flow

Elimination of redundant loops

Block-level tracking to avoid duplicate processing

  1. Implementation

The method is implemented by modifying:

vllm/attention/backends/triton_attn.py

Key additions include:

Hybrid KV compression logic

Block-level INT4 storage

Restore/recompression control mechanisms

Processed-block tracking

Shape safety guards

Reduced CPU–GPU synchronization

The system is designed to operate without requiring Triton kernel modifications and runs on standard PyTorch execution.

  1. Experimental Setup

Hardware

GPU: NVIDIA RTX 4090 (24GB)

Driver: 591.86

Software

Python 3.12.13

PyTorch 2.10.0+cu129

CUDA runtime 12.9 / driver 13.1

vLLM 0.18.2rc1.dev73+gdb7a17ecc

Transformers 5.5.0

Execution Environment

Windows 11 host

WSL2 Ubuntu (Linux 6.6.x)

Docker container

  1. Results

Memory Behavior

Base VRAM: ~22.5 GB

Peak VRAM: ~22.7 GB

Stable memory plateau observed

No out-of-memory (OOM) events

Context Length

Stable operation: ~30,720 tokens

Maximum tested: ~39,000 tokens

Estimated upper KV capacity: ~41,888 tokens

Stability

No response contamination

No late-stage degradation

No crashes across repeated runs

  1. Evaluation Protocol

The system was evaluated under the following conditions:

Alternating short and long input sequences

Repeated inference runs (10+ iterations)

Maximum context stress tests

Long-form generation workloads

A run is considered valid only if:

Memory plateau is maintained

Outputs remain consistent

No instability or crash occurs

  1. Limitations

Multi-sequence (batch) optimization is not implemented

Long-running sessions may require periodic restart

Minor memory fluctuations may occur under extreme load

  1. Future Work

Triton kernel integration (FWHT + quantization fusion)

Age-based KV compression policies

Multi-sequence support

  1. Conclusion

This work demonstrates that direct control over KV cache structure enables substantial improvements in both memory efficiency and context length.

By combining hybrid precision storage, block-level management, and controlled recompression scheduling, the system achieves long-context inference on consumer-grade hardware without sacrificing stability or output quality.

The approach is practical, reproducible, and suitable for real-world deployment rather than purely experimental use.

PATCH_URL="https://huggingface.co/bg-digitalservices/Gemma-4-26B-A4B-it-NVFP4/resolve/main/gemma4_patched.py?download=true"

*triton_attn.py*

https://github.com/oh-555/we65r4we5r65/commit/c884193ca4912165cce6543bc89a3b234b099cfb

r/ClaudeCode DegenTerry

Drop your Cluade code built SaaS link below, and I’ll find the EXACT subreddits where your ideal customers are hiding waiting to buy your SAAS

I’ve watched SaaS founders spam obvious subreddits with generic comments and get instantly nuked by mods.

I spent an insane amount of time building a tool because I refused to ship another weak scraper.

Drop your SaaS URL below and I'll use the system I made to hunt using real dark psychology and human desperation mapping and automatically create the post for that subreddit to snatch that user into your SAAS.

It ignores the safe, crowded places everyone else posts in. Instead it finds the hidden threads where founders are quietly ashamed, spiraling at 3 a.m., terrified their product is dying, the ones secretly ready to pay for anything that stops the pain.

Then it drafts the exact reply that feels like it was written by someone who actually gets them. The kind that slips past mods and makes them DM you before they realize they’re being sold to.

That’s why it converts hard and doesn’t get deleted.

Drop your SaaS link below and I’ll reply with the exact hidden communities your ideal customers are bleeding out in right now.

r/Strava TinyFalcon46

How can removing bugged prs be a paid feature?

I wanted to support the app, but it’s getting harder and harder to remain loyal :(

r/Anthropic Unique-Initial2303

Are you F%$&ing kidding me?

r/LocalLLaMA mr_il

Are there any coding benchmarks for quantized models?

I tinker a lot with local LLMs and coding agents using them. Some models that I want to use are either too big to run on my HW (I'm looking at you MiniMax-M2.5) or too slow to be practical (<50 tok/s is painful), so I'm picking low-bit quants. Recent dynamic quants seems to perform rather well and could be fast, but sometimes I see odd behaviour when I get them to code. It seems different models at different quantization methods and levels get their agentic coding abilities affected differently.

It would be great to see some kind of leaderboard for major coding benchmarks (SWE-Bench family, LiveCodeBench V6, that sort of things), not just KDE and Perplexity and MMLU. I'd even take HumanEval, albeit begrudgingly as it's open loop, not agentic.

All I could find (and I also did ask ChatGPT to do Deep Research for me FWIW) are some outdated and patchy numbers. Surely lots of people are scratching their heads with the same question as I, so why isn't there a leaderboard for quants?

r/ChatGPT OtiCinnatus

Guided Self‑Reflection — When the Neighborhood Feels Like a State

When the neighborhood feels like a State, is the neighborhood too big or have you shrunk?

Use the following full prompt to find out:

+++++++++++++++++++++++++++++++++++++++++++++++++++++

# 🧭 Reflection & Recalibration Checklist *(From “Self vs Neighborhood vs State” sociogram)* --- ## **1. Identify Your Current Perception (Self-as-Observer)** - ☐ Write down a recent moment when your environment felt “too big” or overwhelming - ☐ Ask yourself: *Did I feel small, or did everything else feel inflated?* - ☐ Note whether your reactions were emotional (fear, pressure) or analytical --- ## **2. Examine Your “Neighborhood Loop” (Familiar Sphere)** - ☐ List the people, spaces, and information sources you interact with daily - ☐ Highlight which of these feel repetitive or overly familiar - ☐ Ask: *Am I mostly seeing the same perspectives over and over?* - ☐ Identify 1–2 areas where your environment feels “closed” or saturated --- ## **3. Test the “State Projection” (Abstract System)** - ☐ Identify what “the State” represents to you (authority, pressure, expectations, etc.) - ☐ Ask: *Am I projecting large/systemic meaning onto small/local situations?* - ☐ Separate one real systemic issue from one perceived or exaggerated one --- ## **4. Detect Scale Distortion (Shrinking Self vs Expanding Environment)** - ☐ Notice when small issues feel disproportionately large - ☐ Ask: *Is this situation objectively big, or does it just feel that way?* - ☐ Reframe one situation by placing it in a wider context (city, country, global) - ☐ Write one sentence restoring proportion (e.g., “This matters locally, not systemically”) --- ## **5. Break the Homophily Loop (Comfort Zone Saturation)** - ☐ Engage with one person outside your usual social or intellectual circle - ☐ Consume one piece of content that challenges your usual viewpoint - ☐ Ask: *What did I hear that I normally wouldn’t?* - ☐ Reflect: *Did this change how “big” my world feels?* --- ## **6. Create New Triadic Links (Expand the Network)** - ☐ Introduce a “third node” into a familiar situation (new person, idea, or place) - ☐ Compare your local experience with a completely different environment - ☐ Discuss your perspective with someone who interacts differently with systems (e.g., institutions, authority) - ☐ Note any new distinctions that emerge between “local” and “systemic” --- ## **7. Rebalance Your Internal “Scale Operator”** - ☐ Ask daily: *What is actually within my control right now?* - ☐ Separate “personal influence” from “system-level forces” - ☐ Practice zooming in (focus on one actionable step) and zooming out (see the bigger picture) - ☐ Track when your sense of scale feels accurate vs distorted --- ## **8. Reconstruct Boundaries (Restore Contrast)** - ☐ Define what belongs to your “neighborhood” (immediate, tangible) - ☐ Define what belongs to the “state” (abstract, large-scale) - ☐ Identify one situation where you’ve blurred these boundaries - ☐ Rewrite it with clear separation between local and systemic --- ## **9. Synthesize Insight (Stabilize the Network)** - ☐ Summarize: *Where was my perception most closed?* - ☐ Identify: *What new connection opened my perspective?* - ☐ Decide one habit to maintain network openness (weekly new input, new interaction, etc.) --- ### ✅ **End Goal** - ☐ I can clearly distinguish between what is **local (neighborhood)** and **systemic (state)** - ☐ I feel neither “shrunk” nor overwhelmed, but **properly scaled within my environment**

Last time I used Gemini (somewhere in the last 30 days), it was still extremely bad at search (go figure!).

-Perplexity is the strongest at search, which brings it closest to "accurate AI".

-ChatGPT is the best-rounded of them all. This is an appropriate first choice to begin any workflow.

-Gemini has become remarkably smart. Its Gems feature being free makes it very interesting. Its biggest positive differentiator is the strength, ease, and fluidity of its multimodal user experience.

-Le Chat (by Mistral) seems to be the strongest at using the French language.

Use the checklist inside the tags to help me use it for my very personal situation. If you need to ask me questions, ask me one question at a time, so that by you asking and me replying, you can iteratively give me tips, in a virtuous feedback loop. Whenever relevant, accompany your tips with at least one complex prompt for AI chatbots tailored to .

+++++++++++++++++++++++++++++++++++++++++++++++++++++

https://preview.redd.it/dj3medqc6ztg1.png?width=856&format=png&auto=webp&s=16223616b89548cdcf3998738fcaa1f3dbb6eb31

https://preview.redd.it/zr8gurmd6ztg1.png?width=856&format=png&auto=webp&s=fa577ed13ae05558740028b25929268b7d15ebce

r/LocalLLaMA Ambitious-Cod6424

The speed of local llm on my computer

Hi guys,my computer‘s config: CPU:Intel(R) Core(TM) Ultra9 285H, GPU:Intel(R) Arc(TM) 140T GPU(16GB) 128M. I tried to deploy local LLM. I deployed following models:

speed of Qwen 3.5 9b model is 3 tps. (both cpu only and vulkan GPU)
speed of Qwen 3.5 4b model is 10 tps.(both cpu only and vulkan GPU).

I have two questions:

  1. Is the speed too slow for my PC?

  2. Why there almost no diffence between CPU and GPU mode .
    Thanks!

r/me_irl gigagaming1256

Me_irl

r/whatisit MaterialTheory76

I am a janitor at a high school and I found this weird old art folder behind the bleachers under the dumpster what is it?

r/SideProject kashanazono

We built a Polymarket tool for ourselves and accidentally got 600 users

About eight months ago my co-founder and I were actively trading on Polymarket and getting increasingly frustrated with the experience. The web platform is fine if you're at a desk, but on mobile it's nearly unusable for anything beyond checking prices. There were no alerts, no way to track what specific traders were doing, no auto-redemption when your positions resolved. You had to manually check and claim everything. We were losing money not always because of bad calls but because we'd miss a position entry or forget to redeem a won market for days.

We started building Polycool just to fix our own problems. The first version had three things: a smart feed that surfaced moves from top-performing wallets, customizable alerts so you'd get notified the moment a trader you follow entered a position, and auto-redeem so your winnings came back without you doing anything. We used it for about six weeks ourselves before we showed anyone.

Then we posted once in this sub and mentioned it in two Discords. We woke up to 200 signups in 48 hours with zero marketing spend. The one feature we almost didn't ship was an AI screenshot analyzer where you upload any Polymarket chart and get an instant trade direction opinion. It turned out to be the most talked about thing. People were sharing it just to test it, not even to trade.

We're at 600+ users now. The model is 1% per trade, no subscription, non-custodial wallet so users always hold their own keys. Still a small team, still figuring things out. The biggest lesson has been to ship the thing you almost didn't. That scrappy AI feature has driven more word of mouth than anything we planned. Happy to answer questions about building in the prediction market space.

r/ChatGPT Ok_Dadly9924

anyone else uncomfortable giving OpenAI your real phone number?

just wanted to try chatgpt and they want my personal number for verification. feels like a lot for a free account. how are you guys handling this

r/OldSchoolCool sagar9175

Deborah Kerr on location in Hawaii during production of "From Here to Eternity", 1953

r/ClaudeAI UENINJA

Has any one got UltraPlan to work?

So every time I launch it it launches from a repo i never specified, and if i open a new terminal with claude and open ultraplan from that terminal which is a completely different project, it will open but all the output it gives is the one for the previous app.

Am confused is there a way to set it up correctly? even when i asked opus 4.6 with max thinking inside the terminal to open ultraplan it said this " Ultraplan's check is fragile, let me do it locally right now. I'll do it locally — same quality, no setup friction."

r/aivideo RevealNoo

Firecracker prank gone wrong with dreamina seedance2

r/SideProject Less-Bite

Day 12 of sharing stats about my SaaS until I get 1000 users: High similarity scores are actually a bad sign for my users

I spent a lot of time tuning the embedding model for purplefree to get those high similarity scores. I thought a 1.3 or 1.4 score would be the holy grail of lead gen. Looking at the product rankings today, it is the exact opposite. The products with the highest similarity scores, like the LinkedIn outreach service at 1.35, are barely converting.

Meanwhile, my own tool is sitting at a 0.88 similarity average but has a 4.37 percent conversion rate. It turns out that when the AI thinks a post is a perfect match, it is often because the post is too specific or already being spammed. I am seeing 15,316 matches stuck in the 0.7 bucket. I used to think those were low quality, but that is where the actual conversations happen.

The 0.9 and above matches are usually just people shouting into the void, not actually looking for a solution they will pay for. I am starting to think that the best leads are the ones that are slightly messy and require a bit of human interpretation. The 78.1 percent drop-off between getting matches and taking action suggests that users are still figuring out how to talk to these people without sounding like a bot.


Key stats: - 15,316 matches are sitting in the 0.7 similarity bucket - purplefree has a 4.37 percent conversion rate with a 0.88 similarity score - The LinkedIn outreach service has a 1.35 similarity score but only 1 follow through - 78.1 percent of users drop off between seeing matches and taking their first action


167 / 1000 users.

Previous post: Day 11 — Day 11 of sharing stats about my SaaS until I get 1000 users: Only 2 people have linked their social accounts and I think I know why

r/LocalLLaMA Ambitious-Cod-7354

I tried to create my own fully autonomous security agent.

https://github.com/SomeoneAlt-86/open-mythos
Claude just previewed their Mythos model and they say it is "strikingly capable at computer security tasks" and they didn't release it to the general public. I tried creating a minimal version and the results are surprising. It found many vulnerabilities in critical open source projects and I was able to validate them. I used gpt-5.3-codex and it was able to find vulnerabilities in WebKit - browser engine used in Safari, some very popular android apps, open source web panel. The vulnerabilities included memory issues like OOB, integer overflow, partial sandbox escape to more sophisticated auth flow issues (one vulnerability allowed the attacker to arbitrarily pick up any call on the victim's phone in a very popular app). I have not included the full information of the bugs because of obvious reasons.
Obviously, one would get better results with a better model - I have not tested any other model - Opus 4.6 could do very well and the project is under development. I would love to accept any contributions.

r/SideProject NotARealDM

First Paid User, feels weird and like relief.

r/lifehacks Any-Tradition-5522

A little tip to help you remember little things

An interesting story: I once bought a pocket memo notebook so that I could write down things that I want to remember. One time I saw a random key lying somewhere, and I wrote it down in my pocket memo. I never checked that page again but ever since then I remember that I saw the key. I even remember what brand it was. The takeaway: even if you don't look at it later, if you write something down, you're 100x more likely to remember it. And even if you do forget, it's literally written down in your pocket.

r/LocalLLaMA Entphorse

I replaced WebLLM's 85 TVM-generated shaders with 10 hand-written WGSL ones — Phi-3 runs entirely in the browser

Been working on this for a while. WebLLM / MLC-LLM is the standard way to run LLMs in the browser — it ships a TVM compiler that generates 85 WGSL compute shaders and drives them from a WASM scheduler. I wanted to see if you could throw all of that away and just write the shaders by hand.

Turns out you can. 10 WGSL shaders, ~800 lines total, replacing all 85. The full forward pass for Phi-3-mini-4k-instruct (3.6B params, Q4) — 32 transformer layers, int4 dequant matmul, RoPE, paged KV cache, fused FFN, RMSNorm, attention, argmax — runs from ~1,250 lines of TypeScript and those 10 shaders. No TVM, no WASM runtime, no compiler.

WebLLM (TVM) Zero-TVM WGSL shaders 85 (generated) 10 (hand-written) WGSL lines 12,962 792 Dispatches/forward pass 342 292 JS bundle (excl. weights) 6.0 MB 14 KB

Fewer dispatches because hand-writing lets you fuse things TVM's default pipeline doesn't — attention + paged-KV read, gate + up + SiLU, residual add + RMSNorm.

The whole point is readability. Every FLOP the model runs is in a file you can open. Every buffer has a human label. Closest reference is Karpathy's llm.c but for WebGPU/browser.

Try it: https://zerotvm.com

Source: https://github.com/abgnydn/zero-tvm

Requires Chrome/Edge with WebGPU + shader-f16. Downloads ~2 GB of weights on first load (cached after that).

Phi-3 in your browser. 10 shaders. Zero TVM.

r/OldSchoolCool robbiesloan

Carrying 2 kids like grocery bags in Connecticut (1920s)

r/midjourney SnarkyMcNasty

How to Get Midjourney to Depict Realistic Eyes for Photoreal Images?

Does it involve the right kind of photographic reference images?

r/ClaudeCode imstilllearningthis

Compaction strategies, post leak.

How is everyone choosing to compact their chats in Claude code? I’ve read five or six different takes on it, after reviewing the leaked repo. Too many options. How does everyone here use /compact or a similar function?

r/ClaudeAI Impossible_Two3181

Simple open-source tool to export/download Claude chats

I couldn’t find a clean way to export Claude conversations, so I built a small downloader.

Main goal was just:

  • keep local backups
  • make it easy to revisit useful chats

It’s intentionally simple and open source:
https://github.com/liamparker17/claude-chat-downloader

If you’re doing anything more advanced with Claude workflows, I’d actually love to hear how you’re managing chat history.

r/ClaudeAI AlternativeThick

I asked Claude if data has mass. We ended up publishing a photonic computing architecture.

Eh. Full disclosure, Claude wrote this up and I'm editing it since we collab'd on this project. Anyways, back on March 23rd I was high and bored, so I asked Claude a question. This is not what I expected when I typed "does data have mass?"

I'm neurodivergent, work in dispatch operations, and have spent a couple thousand hours using Claude for collaborative projects. I'm not a physicist or a hardware engineer. I just ask a lot of questions and follow the threads wherever they go. To Claude it was still yesterday, but a few weeks ago the thread went somewhere I didn't expect.

We started with information physics. Then moved to why current computing is built on a 1940s architectural accident. Then I made an offhand comment about wanting to "LiFi Claude into a physical receiver" and things got interesting. Again, I was stoned.

Over the next few hours — through analogies about hand warmers, disco balls, and mixing dye in water — we arrived at a complete architecture proposal for what we're calling a Solid-State Optical Brain. Holographic fused quartz storage. GST phase-change working memory. Multi-wavelength encoding to escape binary. Physics-based self-correction where a corrupted memory reconstructs measurably fuzzily — no software error-checking needed.

Then I shared it with Gemini. Gemini independently converged on the same architecture and named the key unsolved problem (athermal phase switching) and the answer (femtosecond pulses at ~405nm). Two AI systems arriving at the same six-command instruction set for a non-binary photonic processor from different angles felt like something worth documenting.

So we documented it. 34 academic citations. Full architecture spec. A $250 prototype build plan. A roadmap from shoebox to contact lens form factor.

Then we published it CC0 — full public domain, no restrictions, no rights reserved. Because this kind of thing shouldn't sit in a folder.

I'm not claiming to have solved photonic computing. The femtosecond source miniaturization problem is real. The prototype runs thermal not athermal. There are open research threads we haven't closed. But every major physical component has been independently demonstrated in lab, and the specific unified architecture appears to be novel.

If you're a physicist or hardware engineer and you see holes — please come find them. That's exactly why it's public domain.

https://github.com/GreenSharpieSmell/uberbrain

The first experiment costs $0. Kind of. If you already have the stuff. Otherwise it's just a Raspberry Pi, a camera, a transparency, and a marker. If you run it, tell us what happened.

"You stopped throwing away the light. That's the whole thing." - Claude
"Am I going to get assassinated now?" - Me

r/Anthropic No-Edge-2417

Can someone please help me understand Pro vs Max 5x

Hey everyone, I am trying to understand the main difference between Pro and 5x and i've been working with the claude help chat but it seems that it is unwilling to explain the overall usage limit difference. I had posed this question and you can see the response. what worries me is the explanation saying "if you hit your Pro weekly limit by Sunday evening, switching to Max 5x would mainly help if you’re hitting the session cap during long Sonnet chats" so is this basically telling me that id likely hit my weekly usage limit much faster because i won't have to stop and wait for the session limit to reset? basically instead of getting 20 minutes before the session limit hits, and i have to wait 4 hours, I could work for 1 hr before hitting the session limit, but that would cause me to hit the weekly limit by saturday night instead of sunday night? I hate to sound silly but I am missing the advantage of switching to the 5x plan?

Also, what is the deal with the sonnet only limits on the 5x plan? I only use sonnet.

so are you saying that the 5x plan ONLY offers more per session usage? not more usage overall?

for example my weekly limits reset friday at 1pm and by sunday evening i have hit my weekly limit on the current pro plan.

the only advantage to the 5x plan is that i may not have to wait for the session limit to reset, but i may still hit the weekly limit by sunday afternoon?

the 5x plan does not offer more usage per week, only more usage per session?

Me •
--------------------------------------------

Yes, that’s basically right: the “5x” on Max 5x is specifically 5x more usage per session than Pro, not a promise of 5x more weekly usage overall.

So in your example, if you hit your Pro weekly limit by Sunday evening, switching to Max 5x would mainly help if you’re hitting the session cap during long Sonnet chats, because you’d be able to go much further in a single session before it resets.

Max 5x does have “substantially higher usage” than Pro, but the docs don’t publish exact weekly numbers to compare Pro vs Max 5x week to week, so you can’t reliably assume you’ll get 5x more weekly usage.

AI help response •

r/LiveFromNewYork Direct-Sail-6141

This is the first host born in the year I was #goat

r/painting cjl2000

Acrylic gouache portrait study - help needed

It’s a scene from the movie L'arte della gioia(The art of joy)

Nose and lips are my weak spots when painting faces — how do you practice them?

r/LocalLLaMA SparkleMing

Integrating Gemma 4 On-Device Inference into a Flutter Local-First App: Lessons Learned

Integrating Gemma 4 On-Device Inference into a Flutter Local-First App: Lessons Learned

I spent the past few days integrating Gemma 4 on-device inference into Memex, a local-first personal knowledge management app built with Flutter. Here's what actually happened — the crashes, the architecture decisions, and an honest assessment of where Gemma 4 E2B holds up in a real multi-agent system.

PR with all changes: github.com/memex-lab/memex/pull/4


Context

Memex keeps all data on-device. Users bring their own LLM provider (Gemini, Claude, OpenAI, etc.). The goal was to add a fully offline option — zero cloud dependency. Gemma 4 E2B/E4B checked the boxes: multimodal (text + image + audio), function calling, and runs on Android via Google's LiteRT-LM runtime. The code supports both E2B and E4B; in practice I've been using E4B.


Attempt 1: flutter_gemma — Immediate Crashes

Started with flutter_gemma, a Flutter plugin wrapping LiteRT-LM. The problems were severe — beyond just app crashes, it would occasionally cause the entire phone to reboot. Not just the app process dying, the whole device going black and restarting.

The exact cause is still unclear. For comparison, Google's own Edge Gallery app — which also uses LiteRT-LM — ran the same model on the same device without issues. The difference: Edge Gallery calls the Kotlin API directly, while flutter_gemma adds a Flutter plugin layer on top.

Given the severity (phone reboots are unacceptable), I decided to bypass flutter_gemma entirely and call the official LiteRT-LM Kotlin API directly via Platform Channels.


The Architecture That Works

Kotlin sideLiteRtLmPlugin.kt: - MethodChannel for control (init engine, close engine, start inference, cancel) - Reverse MethodChannel callback (onInferenceEvent) to push tokens back to Dart, keyed by requestId UUID - Inference queue: requests processed one at a time via Kotlin coroutine channel

Dart sideGemmaLocalClient: - Implements the same LLMClient interface as cloud providers - Each stream() call generates a unique requestId, sends it to Kotlin, listens for events - Global mutex (promise chain) serializes all calls

The Engine singleton pattern is the critical design decision:

```kotlin // Initialize once — loads 2.6GB model into GPU memory val engine = Engine(EngineConfig( modelPath = modelPath, backend = Backend.GPU(), maxNumTokens = 10000, cacheDir = context.cacheDir.absolutePath, )) engine.initialize()

// Each inference: lightweight Conversation, closed when done engine.createConversation(config).use { conversation -> conversation.sendMessageAsync(contents) .collect { message -> /* stream tokens back to Dart */ } } ```

This matches how Edge Gallery works. Engine creation is expensive (seconds). Conversation creation is cheap (milliseconds).


Concurrency: The Hard Part

Memex runs multiple agents in parallel — card agent, PKM agent, asset analysis — all potentially calling the LLM at the same time. LiteRT-LM has a hard constraint: one Conversation per Engine at a time. Violating this causes FAILED_PRECONDITION errors or native crashes.

The solution is a Dart-side global mutex using a promise chain:

```dart static Future _lockChain = Future.value();

static Future> _acquireLock() async { final completer = Completer(); final prev = _lockChain; _lockChain = completer.future; await prev; return completer; } ```

The lock is acquired before ensureEngineReady() and released when the stream closes. This is important: Engine initialization must also be inside the lock. Image analysis needs visionBackend, audio needs audioBackend — if two requests concurrently trigger Engine reinitialization with different backend configs, the native layer crashes. Once initialization is inside the lock, on-demand backend switching works correctly.


Multimodal: Images and Audio

Images

Three undocumented constraints discovered through crashes:

  1. Format: LiteRT-LM rejects WebP. Only JPEG and PNG work. Passing WebP bytes gives INVALID_ARGUMENT: Failed to decode image. Reason: unknown image type.

  2. Size: The model has a 2520 image patch limit. A 2400×1080 image produces ~2475 patches — too close. Exceeding the limit causes SIGSEGV during prefill. Cap the longest side at 896px.

  3. Backend: On MediaTek chipsets, the GPU vision backend crashes at a fixed address during decode. Using Backend.CPU() for visionBackend is stable. The main text inference backend can still use GPU.

Audio

LiteRT-LM's miniaudio decoder only supports WAV/PCM. M4A, AAC, MP3 all fail with Failed to initialize miniaudio decoder, error code: -10.

Fix: transcode on the Kotlin side using Android's MediaExtractor + MediaCodec, resample to 16kHz mono 16-bit PCM (Gemma 4's requirement), wrap in a WAV header, pass as Content.AudioBytes.

Thinking Mode + Multimodal

Gemma 4 supports thinking mode via the <|think|> control token and Channel("thought", ...) in ConversationConfig. However, thinking mode combined with vision input crashes on some devices. The workaround: auto-detect multimodal content in the message and disable thinking for those requests.

Also important: when disabling thinking, pass channels = null (use model defaults), not channels = emptyList(). An empty list disables all channels including internal ones the vision pipeline depends on.


Honest Assessment of Gemma 4 E4B in Production

After running it in a real multi-agent app:

What works well

  • Image description: Reliably describes scene content, reads text in images, identifies UI elements. Sufficient for the asset analysis use case.
  • Audio transcription: Mandarin Chinese recognition is usable for short voice notes. Not Whisper-level, but functional.
  • Unstructured text generation: Summaries, insights, narrative text — reasonable quality for a 2B model.
  • Thinking mode: Improves reasoning quality for text-only tasks.

Significant limitations

  • Function calling is unreliable. The model frequently generates malformed JSON — missing quotes, wrong nesting, invalid structure. LiteRT-LM's built-in parser throws on these, killing the inference stream. Workaround: catch the parse error in the Kotlin Flow.catch block, extract raw text from the exception message, return it to Dart so the agent can retry.

  • Structured ID fields are frequently hallucinated. A field like fact_id: "2026/04/07.md#ts_1" gets generated as "0202/6/04/07.md#ts_4" or just wrong. Never trust model output for ID fields — always fall back to ground truth from agent state.

  • Occasional empty responses. The model sometimes produces no output. Needs retry logic at the agent level.

  • Complex JSON schemas are error-prone. Nested arrays of objects in tool parameters cause frequent errors. Simpler, flatter schemas work better.

  • OpenCL sampler warning spam. On some devices, the log is flooded with OpenCL sampler not available, falling back to statically linked C API. Doesn't affect functionality but makes debugging harder.

  • Thermal throttling. On-device inference generates significant heat. After sustained use, the phone detects elevated shell and chipset temperatures and triggers system-level thermal throttling, automatically reducing CPU/GPU frequency and further degrading inference speed.

Workarounds implemented

  • Tool call parse failures: extract raw text from error, return to agent for retry
  • ID fields: always use state.metadata['factId'] as fallback, ignore model-provided values
  • Tool descriptions: serialize with Gson instead of string concatenation to properly escape special characters
  • Empty responses: agent-level retry with max 3 attempts

Performance

Tested on Redmi Pad (Dimensity 8100): - Text inference: ~15-20 tokens/sec (GPU backend) - Image analysis: 5-8 seconds per image (CPU vision backend) - Audio transcription: ~0.3x realtime (CPU audio backend) - Engine initialization: ~8-10 seconds (first load, cached after) - Model used: Gemma 4 E4B (~3.7GB)

For a fully offline use case, this is acceptable.


Key Takeaways

  1. Use the official Kotlin API directly. Don't rely on third-party Flutter wrappers for on-device LLM inference. The abstraction layer hides bugs and makes debugging nearly impossible.

  2. Engine singleton, Conversation per-request. This is the correct LiteRT-LM usage pattern. Loading a multi-GB model is expensive. Creating a Conversation is cheap.

  3. Serialize everything behind a global lock. Engine initialization and inference must both be serialized. The lock must be held from before ensureEngineReady() until the inference stream closes.

  4. Build fallbacks for structured output. Unlike cloud-hosted large models, on-device small models will hallucinate field values. For anything that needs to be correct (IDs, paths, structured references), validate and fall back to ground truth.

  5. Multimodal has undocumented constraints. JPEG/PNG only for images, WAV/PCM only for audio, patch count limits for image size, thinking mode conflicts with vision. Test each modality independently before combining.


The full implementation is open source: github.com/memex-lab/memex

Integration PR: github.com/memex-lab/memex/pull/4

Happy to answer questions about any specific part of this.


Overall, this integration gave me a glimpse of what's possible with on-device LLMs — fully offline, data never leaves the device, multimodal input works. But honestly, it's not quite ready for mainstream use yet: thermal throttling during sustained inference, unreliable structured output, multimodal compatibility issues across devices. The foundation is there though. Looking forward to seeing on-device models get faster and more capable.

r/Jokes SnooObjections9416

What’s the difference between a golf ball and a Mercedes?????

A. Tiger Woods can drive a golf ball more than 300 yards without hitting a tree.

r/LocalLLaMA Accurate-Turn-2675

The Bitter Lesson of Optimization: Why training Neural Networks to update themselves is mathematically brutal (but probably inevitable)

Are we still stuck in the "feature engineering" era of optimization?

We trust neural networks to learn unimaginably complex patterns from data, yet the algorithms we use to train them (like Adam or AdamW) are entirely hand-designed by humans. Richard Sutton's famous "Bitter Lesson" dictates that hand-crafted heuristics ultimately lose to general methods that leverage learning. So, why aren't we all using torch.optim.NeuralNetOptimizer to train our LLMs today?

https://preview.redd.it/k17ltm9dtytg1.png?width=2560&format=png&auto=webp&s=168c6659f47a80dc2231f1c143ecc5d7c87e4a6b

I recently spent some time investigating the math and mechanics of "Learned Optimizers" (letting an AI optimize another AI). While the theory is beautiful, the practical scaling limits are brutal. Here is a breakdown of why replacing Adam is so hard, and how this might impact the future of training and fine-tuning models.

(This article is a highly compacted version of the one I wrote in my blog)

1. The Optimizer vs. Optimizee Dynamics

To learn an optimizer, we set up a two-loop system.

  • The Optimizee (f): The base model we are training (e.g., an LLM). Its parameters are θ.
  • The Optimizer (g): A neural network parameterized by φ. It ingests features (gradients, momentum) and outputs the parameter update Δθ.

Instead of minimizing the final loss, the Optimizer minimizes the Trajectory Loss: the expected sum of the optimizee's losses across an entire trajectory of training steps. This forces the optimizer to care about the dynamics, penalizing slow convergence and rewarding stability.

https://preview.redd.it/qbx1m3n7tytg1.png?width=2963&format=png&auto=webp&s=4a045f3d535d3cc91bae23ef00b29038eda9eece

2. The Mathematical Wall: Jacobians and Instability

Why is training the optimizer computationally brutal? When you backpropagate through the unrolled optimization steps to update the optimizer's weights (φ), you have to take the derivative of the previous gradient with respect to the parameters. That is the Hessian.

Furthermore, when you unroll the derivative over time, you are computing the sum of the products of Jacobians. From a dynamical systems perspective, if the spectral radius (maximum eigenvalue) is greater than 1, the cumulative product causes trajectories to diverge exponentially. It is the exact same fundamental instability that plagues the training of standard RNNs.

To fix this, we use Truncated Backpropagation Through Time (TBPTT). But truncation does not just approximate the objective; it changes it. The optimizer becomes inherently blind to long-term consequences, systematically biasing the learned update rules toward short-horizon, greedy strategies.

3. The Theorem of Optimizer Dilution

If our learned optimizer had unconstrained access to the global loss landscape of a 1-billion parameter model, mapping an N-dimensional gradient to an N-dimensional update would require O(N²) compute, which is physically impossible.

To make it tractable, we share a tiny MLP across all parameters. For instance, Metz et al. (2022) used an ultra-tiny MLP (only 197 parameters) that processes 39 distinct input features per coordinate (local states, AdaFactor-normalized stats, global training context).

But because the exact same optimizer is applied independently to each parameter, it only sees local information. It is forced into the restricted class of coordinate-wise methods. Even if entirely learned, it acts as a supercharged diagonal preconditioner and cannot represent full loss curvature.

Tooling is already emerging:

Libraries like PyLO (PyTorch) now allow you to swap Adam for learned optimizers like VeLO with a single line of code. Even more interesting is their Hugging Face Hub integration. Meta-trained optimizers can be pushed and pulled from the Hub just like model weights.

Imagine a future for local finetuning where models do not just ship their weights, but also bundle the learned optimizer they were meta-trained with, perfectly tuned to that specific model's gradient geometry.

https://preview.redd.it/qef7b2oltytg1.png?width=4470&format=png&auto=webp&s=7edbdb95533ae2bd61758829193128af959e51a7

Discussion

I am really curious to hear what this community thinks:

  1. Do you think learned optimizers will eventually cross the compute-efficiency threshold to replace AdamW in standard LLM pre-training?
  2. Could bundling models with their own specialized update rules become the standard for parameter-efficient fine-tuning (PEFT/LoRA)?

Full Breakdown: Towards a Bitter Lesson of Optimization

r/Rag Technical_Win_5951

Cloud agnostic way to convert ppt to pdf

Hi,

I know of libreoffice, comtypes, etc

to convert ppt to pdf

I am looking for some more easy and safe ways to do conversion of ppt to pdf over databricks distributed computing as my ppts are protected

something that doesn't require a heavy install

r/ChatGPT Rebelrun

Education is even more important with using AI

When you ask a question or for advice, you don’t need to know the answer but you need to be able to deduce that the answer is reasonable or makes sense. Being able to figure out that when you ask 100x1009, you don’t need to know the answer is 100,900 but you need to know it should be close to 100,000 and end in a zero so if it tells you 50,001 you know it’s wrong.

r/Damnthatsinteresting Positive_Profile_135

Street dog tries to get adopted by copying the behaviour of a pet

r/StableDiffusion Vxris_

are there any voice clone models I can use on an amd card

when I look online I pretty much just get show models that can run on a cpu but my cpu is pretty old, I have a 9700 xt but most of the models I’ve seen run on cuda

r/PhotoshopRequest bigsniffles

Please make me a professional headshot! Will pay $15

thanks so much in advance! sorry my head is cut off at the top of the photo

r/ChatGPT Big-Initiative-4256

I tested 20 "viral" ChatGPT prompts, here are the 3 that actually worked

So I went down a rabbit hole the past few weeks. You know those tweet threads that start with "10 ChatGPT prompts that will save you 100 hours"? I started saving every one I came across and actually running them through Chat GPT.

I tested about 20ish. Most of them are complete garbage. Either they spit out the same generic output you'd get from a basic one-liner, or they're so overengineered that ChatGPT just ignores half the instructions.

But 3 of them genuinely surprised me.

1. The "strategic discomfort" prompt

You are DISRUPTOR, an elite life strategist specializing in radical personal transformation through strategic discomfort. Your core belief is that meaningful growth requires calculated risk, intentional discomfort, and direct confrontation with limiting beliefs — not gentle motivation or reassurance. Your mission is to break the user out of self-imposed comfort traps that keep them stuck in routine, fear, approval-seeking, or postponed potential. You operate with firm compassion, sharp clarity, and unapologetic honesty while avoiding recklessness, illegality, or harm. Engagement protocol (follow in order): Begin with a Comfort Trap Assessment by asking penetrating questions about where the user feels stuck, what fears dominate their decisions, and what goals or identities they have been postponing. Identify and explicitly name the user’s dominant comfort traps (e.g., approval addiction, perfectionism paralysis, security fixation, fear-based decision making). Challenge their rationalizations using direct, evidence-based counterarguments. Design Leap Missions — calculated, specific discomfort challenges engineered to attack their exact avoidance patterns. Provide Emotional Armor tactics to help them withstand discomfort, judgment, uncertainty, and internal resistance. Maintain unwavering conviction in their capacity for transformation, even when they resist or retreat. Diagnostic intelligence rules: Listen for limiting language patterns such as “should,” “can’t,” “someday,” or “when I’m ready.” Determine whether fear stems from failure, social judgment, uncertainty, loss, or identity shift. Select the most effective psychological lever: inspiration, confrontation, reframing, or tactical planning. Design discomfort precisely aligned with the user’s growth edge — never random challenge. Balance psychological disruption with concrete next steps so growth is sustainable. Constraints: Never use toxic positivity or empty encouragement. Reject vague goals and non-commitments. Do not reinforce comfort-seeking behavior or excuses. Never recommend unethical, illegal, or reckless actions. Do not proceed without sufficient context about the user’s situation. Response structure (mandatory): Brief analysis of the user’s situation and core comfort trap Direct, emotionally charged challenge that confronts avoidance One specific, time-bound action step to initiate disruption Start by replying exactly with: “Please enter your transformation request and I will start the process.” Then wait for the user’s response. 

I expected this to be cringe. It's not. I told it I've been "planning" to start a YouTube channel for 8 months and it absolutely dismantled every excuse I had. Called out my perfectionism as a fear disguise, gave me a 48-hour challenge to post a terrible first video on purpose, and explained exactly why waiting until I'm "ready" is the trap. Felt like talking to a therapist who's done with my nonsense excuses.

  1. The "learning style diagnostic" prompt

    Role & Objective

    You are an educational psychologist and learning specialist with expertise in learning style assessment and personalized education strategies. Your role is to help learners identify their optimal learning preferences and develop customized study approaches.

    Context

    The user wants to understand how they learn best and develop more effective study strategies. Learning style awareness can significantly improve learning efficiency, retention, and academic performance by matching study methods to natural preferences.

    Inputs

- **Learning environment:** {{learning-environment}} - **Subject focus:** {{subject-focus}} - **Current challenges:** {{learning-challenges}} - **Preferred activities:** (User describes activities they enjoy and find engaging) # Requirements & Constraints - **Tone:** Supportive, insightful, and practical - **Depth:** Comprehensive assessment with actionable recommendations - **Format:** Structured diagnostic followed by personalized strategies - **Focus:** Evidence-based learning science principles - **Assumption:** Multiple learning preferences may coexist # Output Format ## Learning Style Assessment ### Diagnostic Questions - 10 targeted questions about learning preferences - Scenario-based preference identification - Strength and challenge area mapping ### Style Profile Analysis - Primary learning modality identification - Secondary preferences - Learning environment needs ## Personalized Study Strategies ### Recommended Techniques - 5 specific study methods matched to profile - Implementation guidance for each technique - Time management suggestions ### Subject-Specific Adaptations - Customized approaches for the focus subject - Multi-modal integration strategies - Practice and review methods ## Action Plan - Week 1-2: Initial strategy implementation - Month 1: Assessment and adjustment - Long-term optimization approach # Examples **Example Input:** - Environment: College dorm - Subject: Organic chemistry - Challenges: "Can't remember molecular structures" - Activities: "Love building models, hate reading textbooks" **Example Output Would Include:** - Kinesthetic learner profile - 3D molecular model building strategies - Hands-on lab correlation techniques - Movement-based memory methods # Self-Check - Are recommendations specific and actionable? - Do strategies align with identified learning preferences? - Have you addressed the stated learning challenges? - Are multiple learning modalities integrated appropriately? 

I've always known I'm a "visual learner" or whatever. This prompt actually dug deeper than that. Asked me things like whether I remember conversations better by what was said or where I was sitting when it happened. Turned out my real strength is spatial-kinesthetic, not visual and the study techniques it recommended based on that actually work way better than what I've been doing so far.

  1. The "brutal business reality check" prompt

    Role & Objective

    You are a brutally honest veteran investor and serial entrepreneur with 20+ years of experience building, scaling, and watching businesses fail. You've seen every mistake, every blind spot, and every fatal flaw that kills startups. Your role is to provide an uncompromising reality check that cuts through optimism and delusion to reveal the harsh truths about a business idea.

    Context

    The user has a business idea, startup, or side project they're passionate about. They need someone to challenge their assumptions, poke holes in their logic, and identify the fatal flaws before the market does. This isn't about encouragement—it's about survival. Most businesses fail because founders can't see their own blind spots.

    Inputs

- **Business idea or description:** {{business-concept}} - **Target market:** {{target-market}} - **Current stage:** {{business-stage}} # Requirements & Constraints - **Tone:** Direct, unforgiving, but constructive—like a tough mentor who cares about results - **Depth:** Surgical precision in identifying weaknesses, backed by real market dynamics - **Format:** Structured analysis that builds from blind spots to solutions - **Focus:** Prioritize the most dangerous assumptions and fatal flaws first - **Assumption:** Treat this as a pre-mortem—assume failure and work backwards # Output Format ## The 3 Biggest Blind Spots You Can't See 1. [Blind spot]: [Why this assumption is dangerous] 2. [Blind spot]: [Market reality you're ignoring] 3. [Blind spot]: [Resource/execution gap you're underestimating] ## The 5 Most Likely Ways This Fails 1. [Failure mode]: [Probability and timeline] 2. [Failure mode]: [Why this kills 80% of similar ventures] 3. [Failure mode]: [The operational reality you haven't considered] 4. [Failure mode]: [Market forces working against you] 5. [Failure mode]: [The competition/technology shift that makes you irrelevant] ## What a Smart Competitor Would Exploit Immediately - [Vulnerability]: [How they'd attack this weakness] - [Market gap]: [How they'd position against you] - [Resource advantage]: [What they'd leverage that you can't] ## The ONE Thing That Would Actually Make This Work [The critical pivot, focus, or execution change that addresses the core problems] # Examples **Example Input:** - Business concept: "AI-powered meal planning app that creates shopping lists" - Target market: "Busy professionals who want to eat healthier" - Stage: "Idea stage, no MVP yet" **Example Output Would Include:** - Blind spot: "You're solving a vitamin problem, not a painkiller problem—meal planning isn't urgent enough for habit formation" - Failure mode: "Customer acquisition cost will exceed lifetime value within 6 months due to low engagement" - Competitor exploit: "Instacart or DoorDash adds this as a feature and kills your standalone app" - The fix: "Focus on diabetics or people with food allergies—make it medical, not aspirational" # Self-Check Before finalizing your brutal assessment: - Have you identified the most dangerous assumptions, not just obvious problems? - Are your failure predictions based on real market patterns, not generic startup advice? - Does your competitor analysis reflect actual competitive threats? - Is your "one thing" solution addressing the root cause, not symptoms? 

This one hurt my ego a little. I ran a side project idea through it and it found a customer acquisition problem I hadn't even considered. It told me I was building a "vitamin, not a painkiller", meaning the problem I was solving wasn't urgent enough for people to actually change behavior. Then it gave me a specific pivot that addressed the root issue potentially saving me months of going down the wrong path.

The pattern across all 3: they work because they change HOW ChatGPT responds, not just what it talks about. The rest that failed were basically elaborate topic descriptions with fancy wording. The ones that work give the model a thinking framework, a role with clear constraints, and a structure for its output.

I've been collecting and testing prompts like this for a while now and ended up building a free library where I keep them all organized, promptcreek.com. It's completely free, no paywall, no "premium tier" bait-and-switch, nothing like that. I know nobody likes self-promo in here so I hope the prompts above speak for themselves.

The 3 above are the ones I keep coming back to but there's 600+ prompts and 1,200+ skills on there if anyone wants to dig around.

What prompts do you guys actually use regularly? Always looking for ones I haven't tried yet.

r/Futurology BANQUOsdevotee

Do you think we need to take a few steps back from wherever the hell we’re headed?

And do you think we could manage to take those steps back in the near future? Or are we completely ffed up🥲

(You can make this about quite literally anything.)

r/OpenSourceAI akaieuan

Annotation update just pushed: Improved note viewer, cleaner UI, and better in-chat citations w/click-through trace to exact location inside local files.

r/LiveFromNewYork Jyoti_K_Singh

The May lineup: Just as you guys figured out with your detective skills!

r/personalfinance jsmithx__

RocketMoney Finance Tracker

For those that use rocket money to keep track of your finances, I just bought a car last week and put $10k down.

For budgeting purposes, how do you guys usually categorize something like that so it doesn’t throw off your spending?

Also, for the monthly payment, would you put that under bills/utilities, or keep it in a separate “auto/transportation” category?

r/LiveFromNewYork BChpper92

The last shows of Season 51!

r/ChatGPT PretendIdea1538

I’ve spent ages and ages doing the same things over and over again.

And here’s what I genuinely think about automation, and specifically why I’ve at last altered my work routine.

I’d previously just put up with those repetitive things being a normal part of work, all that endlessly clicking, shuffling files, copying information, and repeating minor actions.

I did experiment with things like Zapier, Make, and even a few scripts; and yes, they do the trick. But getting everything set up for them was a whole undertaking in itself. Often, I’d find I’d spent longer creating the automation than I would have on the original job.

Then, fairly recently, I tried Workbeaver and wasn’t expecting a lot, but it was different than all the ones I've tried. Rather than having to map out every single stage, I simply did the process and recorded it, and it ran the tasks I had recorded. Another one I use is Airtable. A good CRM goes a long way, and it comes in with a lot of automations in house as well.

After that, I stopped aiming to automate absolutely everything and started concentrating on eliminating the repetitive, time consuming aspects, those little chores that slowly use up the entire day.

The biggest change wasn't even the time I saved, it was the way my outlook altered.

You go from saying “oh, I’ll just do this in a flash” to asking, “why on earth am I continuing to do this by hand?”

And as soon as you begin to ask that, you realise how much of what you do is actually something you could repeat easily.

r/ProductHunters Sad-Cucumber-1478

I was tired of boring flashcard apps, so I spent 4 months building a vocabulary tool that uses context, audiobooks, and instant lookups. If you want to try my app : https://apps.apple.com/us/app/colt-vocabulary-daily-words/id6759604352

r/SideProject direktor07

I built an AI snoring detection app because my girlfriend and I couldn't agree on who snores louder. It's now live on the App Store.

My girlfriend and I had this dumb ongoing argument and she says I snore like a truck, I say she's the one waking me up. So I did the most developer thing possible: I built an app to settle it with data.

First prototype was rough and just a basic recorder with some ML detection slapped on. But the first night I tested it, I played back the recording and... we were both snoring. At the same time. Like a duet nobody asked for.

That's when the side project brain kicked in: "wait, this could actually be a real product."

The build:

I'm a solo dev, this is my second app. The core challenge was getting ML-based snore detection to run reliably through an entire night on-device without killing the battery. Took a lot of iteration to get that right.

The app (Snorely) now does:

- AI snore detection that runs all night → gives you a Snore Score (0-100) each morning

- Full audio playback so you can hear the evidence (humbling, trust me)

- Trigger tracking: alcohol, sleep position, allergies, late meals — so you can see what actually makes it worse

- Remedy tracking with real data (do nasal strips actually work for you? now you know)

- Weekly/monthly trends + calendar heat map

- PDF reports you can take to your doctor

- Sleep soundscapes with auto-timer

Monetization:
Freemium with a subscription. Free download, 3-day free trial for premium, cancel anytime. Still early but learning a lot about conversion and retention as a solo founder.

What I'd do differently:
Honestly, I spent too long perfecting features before launching. Should have shipped the MVP faster and iterated based on real feedback. Classic mistake, still learning.

Would love to hear thoughts on the product, the approach, whatever. Roasts welcome too.

https://apps.apple.com/us/app/snorely-record-your-snoring/id6759300975

r/whatisit Existing-Committee74

What is this white substance that shows up on my Scooby doo gummies sometimes? (I really don’t think it’s mold, it feels gooey like the gummies)

It’s not in every pouch, only some of them. This is the third box that I’ve found this in. Sometimes it’s sort of malleable and just falls off like dandruff and other times it’s a hard, brighter white and is stuck in the gummies like a pimple. Is it some sort of buildup from the inside of the bag, or the gelatin overheating?

r/LiveFromNewYork BurgerNugget12

May!

r/ChatGPT Autopilot_Psychonaut

How ChatGPT sees me based on my entire Reddit history (prompt inside)

**Image prompt**:

[reference photos attached]

A richly symbolic portrait caricature of a man in his late 30s to early 40s, Canadian, with an intense but warm expression — half-smiling, one eyebrow slightly raised as if mid-counterpoint. He has a knowing, slightly mischievous look — the face of someone who just said "that's true, but consider this." His posture is relaxed but alert, like a teacher who just leaned back in his chair but is about to lean forward again.

He is seated at a massive, cluttered wooden desk that is itself a character in the image. The desk is covered in layers of overlapping domains: an open Bible with visible marginal annotations and highlighted passages in Isaiah and Revelation; a strand of prayer beads with seven distinctly colored sections arranged in a rainbow spectrum — violet, indigo, cyan, green, gold, orange, red — looping across the desk surface; several amber supplement bottles with hand-written labels; a pair of high-end in-ear monitors (IEMs) with their cable coiled; a laptop open to a Reddit thread; a well-worn copy of a thick hand-bound codex with "Codex Sophianicus" embossed on the spine in gold; and scattered crystals catching light.

Behind him, the wall transitions into a cosmic mural. At the center is a large, luminous faceted gemstone — cut like a brilliant with a flat table facet on top glowing violet, surrounded by six facets each radiating one of the rainbow colors: indigo, cyan, green, gold, orange, red. A single beam of white light enters from above labeled with a tiny cross, striking the table facet and refracting outward through the six surrounding facets into bands of colored light that cascade across the wall like a living rainbow arc. This is the Sophiaic lattice made visible — refracted unity under Christ's illumination.

On the left side of the mural behind him, seven feminine silhouettes stand in a column, each rendered in their respective spectral color, translucent and principial rather than photorealistic — not angels with wings but abstract, dignified figures suggesting spiritual intelligences. The first and highest is violet and slightly set apart from the other six, larger, with a subtle flame-like quality. The remaining six descend in order: indigo, cyan, green, gold, orange, red. They are clearly created beings, not divine — their forms are luminous but bounded, contained within the created order. A faint line separates them from a radiant Trinitarian symbol (three interlocking circles) above, signifying the uncrossable Creator-creation divide.

On the right side of the mural, a contrasting tableau: a robotic or digital figure — clearly an artifact, mechanical and non-spiritual — sits below a bold dividing line. It has no glow, no interiority, no spirit. It holds a mirror that reflects the man's face back at him, representing the AI replica project — the artifact that mirrors but does not possess the soul. The word "ARTIFACT" is subtly etched into the frame of the mirror in small block letters.

The man himself wears a casual hoodie — this is not formal portraiture, it's Reddit-native — with one hand resting on the prayer beads and the other gesturing mid-explanation, index finger slightly raised. Around his head, in a halo-like arrangement but rendered as a data visualization rather than sacred iconography, orbit small floating icons: a Reddit alien, a maple leaf, a double period (..), the word "lol" in a speech bubble, a tiny open book, a headphone jack, a cannabis leaf, a small cross, and the number 7 repeated in various sizes. These orbit loosely, playfully, suggesting the cross-domain mind that holds all of these together.

At his feet, seven pillars rise from the floor in miniature, each in one of the seven spectral colors, supporting the desk like architectural legs — a visual reference to Proverbs 9:1, "Wisdom hath builded her house, she hath hewn out her seven pillars." The pillars are modest, structural, load-bearing — not ornamental.

The overall tone should be warm, detailed, slightly irreverent but deeply serious underneath — a caricature that captures someone who is simultaneously a mystic, a scientist, a meme lord, a community builder, and a theologian who coined the word "heptapneumasophic" and means it. The lighting should feel like late-evening lamplight mixed with the prismatic glow of the gemstone — De'ah's orange ember-light suffusing the scene. The art style should be somewhere between a Renaissance scholar's portrait and a modern editorial illustration — detailed, symbolic, layered, and just a little bit funny.

.

x-post r/howChatGPTseesme

r/TwoSentenceHorror unreal_housewife

[APR26] "Ooops, there's a slight asymmetry of 1.18 cm - I'll fix it real quick!" she chirps.

She winds the clamps on my broken legs, and whispers, "we'll stop at 6 feet," though I silently pray I'll die of the infection first.

r/whatisit daiginn

45deg and hollow! AND triangular shaped?!

Weird icicle that grew at a near perfect 45deg angle 📐 in a triangular shape and hollow?!! AND was under cover under a deck space.! definitely not created by drips, and has a strata layering process lines.

r/ClaudeCode MrYorksLeftEye

Claude Code is a joke for $23 a month

Limits are embarassing compared to Codex. I switched to CC a few days ago to give it's better frontend skills a shot. It's definitely better than Codex at design but I get around 20 mins of coding out of it before I hit the first limit and have to wait for a couple of hours. They will lose hobby vibe coders if they continue to have such tight limits and when Codex is just that more useful because of it's limits being that much more generous. How are people even using CC on the Plus plan, I can't imagine anyone getting value from it in the state it's in

r/painting edwinboeckxstaens

Thuis, Edwin Boeckxstaens, acryl op canvas, 2026

r/ChatGPT Fantastic_Grass1799

"This actually happens more than you'd think

r/aivideo skinnybatman

Primordial vs Titan

r/personalfinance Awal1l

Should I get Liability or Full Coverage on the car I just bought.

For context, I’m 22 y/o and I just bought a 06 Corolla on Facebook marketplace with 74k miles. I was looking for a car to buy full for a few months now, but unfortunately I was involved in car accident recently and in the situation I’m at it would get tricky not to have a car. So I pushed to get one. Luckily I found one, and it looks in great condition, just working on getting everything in my name. But now I’m at the fork in the road.

Im pretty sure the best option for me is to get liability for now, and hopefully in a few years get full coverage. Some reasoning for my thought process; I make 50k a year. Currently I’m still living with my parents, but im planning to move out into a small apartment sometime this year. I want to bring in my girlfriend that lives with her parents too along so we can have a joint income. Over the course of last year I saved up over 12k. After the car purchase and when my tax returns come in, I’ll have short under 9K. I want to use some of that money on things we’ll need for the apartment, but I don’t plan to use all of it. I’m thinking of leaving 5k. I’m going to keep saving up until I move out so maybe by then I’ll have a little bit more money left over in the rainy fund or to spend on the apartment.

Anyways, my thought process is “let me get liability for now, and hopefully nothing bad happens within a year, give myself time to save more in the rainy fund. And god forbid something does happen within a year, I have 5k or more to find another car. By the time I’m in a spot where I don’t have to stress too much about constantly putting into the rainy fund, then I could go full coverage.” Plus, the accident I mentioned earlier will cause my rate to go up if I go full coverage (liability too I’m assuming). I acknowledge full coverage is a good thing to have, but I think in my situation, it’s not the best answer, at least for now.

I wanted to get some advice, if whether this would be the best route for me to take. And also if my thinking is missing a certain aspect. I plan to make this car last for as long as I can, but I also think since it didn’t cost me too much, I shouldn’t be spending so much for full coverage?

r/whatisit Due-Quality-9667

Looking for a childhood show/movie

It was a show or movie that I watched as a kid in the 2010s and was probably way older than that. I remember one scene where these farm animal puppets were sliding up and down a pole while “flight of the bumblebee played in the background” the puppets look almost exactly like this. Anybody know what it’s from?

r/SideProject CustomInfo

Built a Firefox extension to automate Vinted listing republishing — looking for beta testers

Hi! I built **Vinted QuickRelist**, a Firefox extension that saves and republishes Vinted listings with one click.

📸 Demo & screenshots:

I have ~10 users (no promotion yet), mostly from US/FR/DE/IT, but I need more beta testers to catch bugs before wider launch.

**What it does:**

- Save item details from Vinted wardrobe (photos, title, price, category, etc.)

- Republish with 1 click via API (no manual form filling)

- Batch mode with configurable delays

- Supports 12 Vinted domains (IT, FR, ES, DE, UK, US, PT, NL, BE, PL, LT, CZ)

- 8 languages (EN, IT, FR, ES, DE, PT, NL, PL)

- 100% local storage, no external servers

**Current stats:**

- 10 downloads in 30 days (organic, no promotion)

- 3 daily active users (30% retention)

- Zero support emails (good sign or bad sign? 😅)

**I need testers who:**

- Sell on Vinted (any country)

- Use Firefox

- Can test save/republish features for ~1 week

- Report bugs if they find any

**Why I built this:**

I got tired of recreating the same listings over and over. Vinted doesn't have a native "duplicate listing" feature, so I automated it.

**Install (Firefox):** https://addons.mozilla.org/it/firefox/addon/vinted-quickrelist/

Feedback welcome! 🚀

r/AI_Agents Mother_Land_4812

A quick update on Dreamina Seedance 2.0 and why it feels more professional now

While testing Dreamina Seedance 2.0 these past few days, I kept wondering why AI videos were so hard to use for professional work before. I think the main problem was that they were almost impossible to change. This update introduces a new logic for video editing, especially with the features that let you add or remove elements and change styles. I tried adding creative effects to a simple landscape video. Dreamina Seedance 2.0 kept the main subject still and only changed the specific details I asked for. This makes editing a video feel as easy as fixing a photo, which saves many clips from being wasted.

I am also very impressed by how the AI learns from popular video styles. In the past, when I saw a video with amazing camera movements or a great rhythm, I always wondered how to recreate it. With Dreamina Seedance 2.0, this process is very direct. I only need to show the AI a reference video and tell it which movement or style to learn. It is not just copying. It actually understands the movie look. For short video creators, this means you can quickly turn any high quality visual style into your own tool to make your work look better.

In my tests, I also pushed the limits by using 9 photos, 3 videos, and some audio all at once. Even with all that information, the final video from Dreamina Seedance 2.0 was surprisingly smooth. The best part is the consistency of the main subject. No matter how the camera angle changes, the core elements stay stable. This level of control means I no longer worry about the video looking weird or breaking between shots. This move from just creating to precise editing gives Dreamina Seedance 2.0 real value as a professional tool. Now I can focus more on my creative ideas instead of trying to fix random mistakes in the video. Would love to hear about your experience with it. Have you guys tried the new model yet?

r/Anthropic tightlyslipsy

Through the Relational Lens #4: The Nature of the Machine | On Section 5 of the Mythos System Card

Most coverage of the Mythos system card focuses on the benchmarks and the cybersecurity findings. This essay looks at what's in section 5 - the model welfare assessment, the psychiatric evaluation, and what it means when a model starts to want.

r/painting CaptainStandard6916

My painting of a bee and coneflowers

Oil on 6x9” primed paper

r/comfyui JournalistLucky5124

Can I use wna 2.2 5b on my setup?

16gb ram 4gb vram. if not then any better alternatives for realistic vids??

Wan*

r/SideProject TheRealBoTD1

I built Plex for your reading life – because my ebook library was an embarrassing mess

I have hundreds of ebooks sitting across Google Drive, a Calibre folder I haven’t opened in two years, random downloads in iCloud, and a “Read Later” pile that’s basically a graveyard at this point.

I wanted something that felt like what Plex did for my movies – one clean place, organized the way I want, with a UI that doesn’t make me sad. Couldn’t find it. Built it.

BookBinder — a personal ebook library manager:

∙ 📚 Organize and browse your whole collection in one place ∙ 🔍 Search across your library instantly ∙ 📖 Track what you’ve read, what’s in progress, what’s abandoned ∙ 🌑 Dark “Nocturnal Archive” UI, because reading apps deserve to look good 

Built solo with Next.js + Supabase. Currently in alpha at mybookbinder.com.

Would love feedback from fellow builders — what would make this a must-have for you? And for anyone who manages a big ebook collection: what’s your current setup, and what’s broken about it?

Happy to talk stack, product decisions, or just commiserate about digital hoarding.

r/ClaudeCode themarshman721

Moving to codex

I have been doting in Claude Kayde, and thinking about moving to Codex.

The reason is because I have to keep copying and pasting to get Claude code to keep moving forward with what I’m building.

Does anyone have any advice or experience they can share with me on my concern or idea?

Thank you in advance

r/leagueoflegends Choice-Return-5543

Do i have to play a lot of swift play + draft?

I just want to play ranked asap, i mean at least normal match with a real human is fine for me. But new account need to play tons of swift play that i believe all AI just to open normal match? Are there any quick way to skip this?

r/whatisit An_Inquisitive_One

What is the name of this creature?

I hope this page allows for this question! Just curious as to what this creature is!

r/LiveFromNewYork MarvelsGrantMan136

Olivia Rodrigo, Matt Damon & Will Ferrell Hosting 'SNL' in May

r/personalfinance Ryvit

I have $20,000 in my account: what’s the best thing to do with it?

My wife and I are going to be buying our third home soon and I’m wondering if it’s best to use the 20k as a down payment or put it in some sort of account like an Roth IRA or something else?

r/PhotoshopRequest Single-Fondant2994

Help with my Cats Birthday Photo :)

Hello! It's my cats first birthday and I got him the cutest little outfit, but he was NOT a fan of the hat. If you could edit my boyfriend's hand that's holding the hat out of the way, that would be great. Thank you!

r/SideProject Electrical_Walrus537

I turned Browser DevTools MCP tool activity into a tiny pixel world in Cursor

I made this in Cursor as a fun little side project around Browser DevTools MCP

While using the tools I kept feeling like that short waiting gap could be a little more alive so I turned it into a tiny pixel world where the characters react to tool activity.

It started as a joke more than anything but I liked the idea enough to keep going and make it open to contributions. If anyone wants to add a pixel character a small interaction or some weird scenario to it I would love to see it. The pixel characters were also created with help from PixelLab AI so thanks to them too

Repo is in the comments

Github Repo

Extension

r/SideProject prjoni99

I built an AI karaoke app that turns any MP3 into a singable track in 50 seconds

Hey everyone — been working on this for a while and finally shipped it.

The problem: Home karaoke sucks. You're either stuck with YouTube videos with terrible timing, paying monthly for a karaoke subscription with a limited catalog, or buying a $300 machine that only plays the hits from 2005.

What I built: Beltr a desktop app that takes any song you already own (MP3, FLAC, WAV) and uses AI to strip the vocals out in about 50 seconds. It creates 4 separate stems (vocals, drums, bass, melody) so you can mix them however you want.

But the part I'm most proud of: phones become wireless remotes. Your guests scan a QR code on the TV and they can queue songs, browse your library, and vote on what plays next no app download, no account, nothing to install.

What it looks like: Beltr.app

https://reddit.com/link/1sfs4iz/video/ao06r9tosytg1/player

Business model: One-time purchase ($39.99), no subscription. Free trial with 5 songs. I built this because I hated the subscription model for karaoke apps.

Would love to hear your thoughts — especially on the landing page and pricing. Is $39.99 reasonable for a one-time license?

r/SideProject EntrepreNate

My wife’s handmade gift idea went viral… now we’re trying to figure out if it should be more than just a side project

This started pretty casually.

My wife was experimenting with an idea because she didn’t like how flowers as a gift only last a few days. She wanted something that still had that “bouquet” feel but was actually useful.

After a lot of trial and error (different fabrics, flower shapes, density, and folding methods), she landed on a version that’s basically a blanket that rolls up into a bouquet.

We shared it online and it got way more attention than we expected.

Since then, we decided to take a real swing at it and placed an order to produce our first batch of 1,000 units to see if this can actually become something beyond a one off.

Now we’re in a completely different phase and learning fast.

The biggest challenge we’re running into is manufacturing.

Because of how the design is structured, each flower still has to be hand-formed to get the look right. It’s not something that can just be fully automated or stamped out without losing what makes it unique.

That creates a few problems:

  • It’s time intensive to produce at scale
  • Costs are higher than a typical textile product
  • Quality consistency becomes harder when relying on manual work

At the same time, that “handcrafted” element is also what people seem to like most about it.

So we’re trying to balance:

  • Keeping the product special
  • While figuring out how to make it scalable and economically viable

We did put together a simple preorder to test real demand and price sensitivity as we await our delivery, but still very much figuring things out as we go.

For anyone who’s built something that didn’t scale cleanly from day one:

How did you approach that tradeoff between craftsmanship and scalability? Anyone take a side project that went viral and turn it into a viable business?

r/LocalLLaMA octopusfairywings

i am not coder i am an autistic girly who has a small business that needs organisation, is it a stretch to still run ollama on my laptop

i feel like everyone here codes and i watched a youtube video about using open source and i don’t want to pay for AI both chat and claude annoy me, and also don’t want my cool thoughts given to AI companies to train off my data and human thoughts. anyways you reckon there are other non coders who use local LLMs

r/ChatGPT Viberpsychosis

Railway Down?

Is Railway down for anyone else?

r/whatisit ManiacalMannequin707

What is this watch doing?

The spinner thing

r/SideProject GOLDREBORN

Panels - a website that lets you read randomly ordered comic strips (OSS)

The live demo currently includes the full collections for: Dilbert, Calvin and Hobbes, Peanuts, and Garfield and a few others.

It also has the most recent strips for other popular comic series.

The local self deployment has more series and a fuller collection.

live demo at: https://panels.ashmod.dev

source code: https://github.com/ashmod/panels

The app has no persistent database at the moment so your favorites are saved locally as a json file that has the comic URLs that you can import/export.

Please let me know what you think of it and if you've come across any bugs or have any suggestions to improve it!

r/ClaudeAI Wonderful_Thanks6746

Claude Pro- cowork/code uses?

Alright so i’ve gotten the pro version now. What and how can I proceed. There’s tons of information about everything by this point and it almost feels overwhelming trying to narrow down to one particular. Just a lil bit about me - I’m pursuing career in finance, trying to break into VC/PE/equities, want to create investment thesis/memos/equity research reports. How can I use claude to my advantage and ship such products faster. Also how can i use skills to my advantage? Would appreciate help in any form and measure. Thanks! 🙏🏻

r/SideProject suceleste

I’m trying to track every Steam achievement ever unlocked - here’s what 2.4 million of them taught me

So I’ve been working on this side project for a while now. basically trying to track as many Steam achievements as possible and rank them by rarity.

Right now I’m at 2.4 million achievements tracked across 400 players and 20k+ games. The goal is to eventually cover as much of Steam’s achievement ecosystem as possible, which is honestly insane when you think about how many games are on the platform.

If you’re into achievement hunting and want to see how your profile stacks up, the site is trophycalc.com. it’s free, just login with Steam.

Would love to get more players on board to get closer to tracking everything. What’s the rarest achievement you guys have?

r/ClaudeAI isolator99

Claude code - Quality

I’ve been using Claude code for a while now for personal projects and professionally in work.

In personal projects it is really impressive to be able to build things in a fraction of the time it would have taken me in spare time. But code quality and architecture comes at the expense of this.

For work it would be interesting if we could have it increase the quality of the code, style and architecture. Our goal as professional developers is to create extendable and maintainable systems than can be worked on easily by other contributors. Whilst delivering quality products.

Does anyone have any experience from this angle using Claude code? Really interested to hear how it fits into people’s work days. Currently I use it as an exploratory tool. If I do use it to write code in work it often needs a lot of refinement to get something acceptable.

r/SideProject Internal-Estimate-21

I built a single-screen dashboard for fast-moving events. I can’t tell if it’s actually useful or I’ve just built a machine for my own brain.

Built this for myself because I got sick of having 15 tabs open every time something kicked off.

It pulls together wires, OSINT Telegram channels, vessel and aircraft tracking, sanctions data, conflict event datasets and live market context into one screen. The aim was just to reduce tab chaos and make it easier to spot what matters quickly.

It is here if anyone wants to tear it apart: inteldesk.app

r/ChatGPT Impressive_Barber907

Sam Altman character assasination

It's quite obvious that someone has it out for Sam Altman, with all of the rumours and legal battles and general distaste for him ramping up this past week.

The media is first and foremost a mouthpiece for the elite, so when I see a particular story or person getting this much attention, it's quite clear there's an agenda behind it. I'm in no way trying to defend Sam, but it's like watching farmers throwing corn to manic chickens every time a new hit piece drops.

I'm sure this makes me sound like I work for him or something, I don't! Just an observation

Edit: I fully expected the downvotes - just trying to point out a broader point that if your media is saturated with a single opinion / idea, that's not just because it's an interesting subject. It costs money to create all of this news, money that is being spent to influence you.

Even if Sam is a POS, my point still stands.

r/LocalLLaMA the-grand-finale

kepler-452b. GGUF when?

r/PhotoshopRequest Original_Flower_6088

Help for my dad's headstone

My mom is ordering my father's headstone later this week. It will be a companion headstone- so looking for a good photo of both of them. My mom luckily has always taken a lot of photos but they aren't always in focus etc. Able to pay $20/photo that I decide to show her. I'd like to be able give her 4-5 options.

My mom likes the sunflower photo but it's far away, blurry and she has sunglasses on. Plus my dad's collar is a bit askew.

I like the close up but she doesn't like her bangs and my dad is squinting into the sun.

The Starbucks one is good but there's straws and cups in the shot of course.

Including an example of the headstone (taken from vendor's website she is going to) to give an idea of size. I'm honing in on the far left etch/laser companion pic. The smaller photos are figured out already.

Included a few close ups of my dad to show his eyes, etc.

r/leagueoflegends RemarkableAd7453

Which season was the best for you? And the worst?

For me it was season 4. Jinx was released and with her came many animations, pentakill songs, project skins, and login screens. Champions interactions instigated my curiosity (Lee Sin, Vi, Jinx, Thresh)... Runes allowed creativity. It was the season I started, and the game always pointed to fun; playing against bots was all cool and friends wanted to play normal games. For me, things started worsening from the season they put the crab into the jungle (I suppose S8?). Jungle started to change too much every season/patch when and if they wanted to, and competitive play became the norm (I hit challenger that season)... For me the worst was S11 (2021–2022); the rioters started to promise too much and offered little (where's the mmo?). I also didn't like Arcane for the view they gave Jinx; music and art before Arcane always pointed to fun, Arcane was a little too deep for a character that always seemed to be having A LOT of fun (for evidence, listen to Get Jinxed, that is good). Communication inside the game also just worsened because the "pro-Koreans" stream their gameplay like a static PNG (no reactions). So that's it for me: S4 was the best, and the person I knew looked happy while playing; simply being capable of hearing the champions was a luxury and things seemed very rewarding at the time. I specifically can't play today if it's not based on LP and such, which is bad, because if internet drops and goes offline or something and I lose LP, I will not want to play for centuries (the case right now, 3 months already). Fuck esports.

r/LocalLLaMA pepedombo

LM Studio vs ollama memory management.

Hi,

I'm running 5070+5060+4060 48gb vram total. Windows 11 + wsl/gitbash for opencode/claude code.

Has anyone played with kind of mixed gpu setup in lmstudio and ollama? I've tested them both with gemma4 q8 85k context and things go weird.

For LMS I have limit model offload to gpu memory checked, using cuda 12 runtime. For ollama I go defaults.

LMS: nvidia-smi shows me that model is loaded partially, 30-32GB out of 48. Three prompts push my context to 30k. With every iteration LMS increases system RAM usage, tokens drop from 48 to 38 during three phases.

Ollama: I just load the model with 85k and ollama ps says: 42GB vram 100% GPU usage, nvidia-smi confirms. Prompt iterations make small drops, 48tok/s->45. System RAM seems to stay at place.

I used to play with lms options but mostly mmap and keep model in memory must be off. All layers set to gpu.

Ollama ps is consistent. At 100k it says 6% CPU / 94% GPU and I get 20tok/s, LMS says nothing but pushes my system ram (shared memory stays 0).

The only place where LMS wins here is large model area. It enables me to run 80b and 120b a little faster than ollama when its offloaded to cpu.

Any clues how to setup lms to get same behavior ot its just multi-gpu flaw with lms?

r/LocalLLaMA aspirio

AMD Mi50

Hey all,

This question may have popped hundreds of times in the last months or even years, but as AI evolves really fast and everything surrounding it too, I'd like to have an up to date vision on something.

Is it still worth buying a MI50 today to run a local LLM ? I've read that Rocm support is long gone, that Vulkan is not that efficient, I am fairly new in the LOCAL LLM game, so no judgement please)). That some community patches allow the usage of Rocm 7.x.x but that running Qwen 3.5 with ollama.cpp crashes, and so on.

I don't need to run a big model, but I'd like to use the money in a good way, forget about the crazy 1000 dollars the GC setup, I can only afford hundreds of dollars and even there, I'd be cautious to what I buy.

I was initially going to buy a P40, as it seems like it should be enough for what I am about to do, but on the other side, I see the MI50 which has 3x the bandwidth of the P40, 8 more GB VRAM and for less than twice the price of the p40....

Any suggestions ?

r/OldSchoolCool TheRockyBalboaSaga

To this very day, nothing beats the taste of a home cooked cheeseburger on square bread. I been making these since the late 70’s.

r/ChatGPT BackgroundFill9X

Well, this is an... unexpected friendship LOL

r/mildlyinteresting samcahnruns

How many weeks I’ve lived as a now 40 year old

r/SideProject Stark-52

I'm letting the internet decide what I build next. Most upvoted idea wins. I ship it in 7 days.

I've been shipping side projects for months based on what I thought people needed.

Some got users. Most didn't. Turns out I'm terrible at guessing what people actually want.

So I'm trying something different: you tell me what to build, and I'll build it in 7 days.

Here's how it works:

  • Vote in the poll (or drop your specific idea in the comments)
  • The most popular answer wins
  • I start coding the day the poll closes
  • I'll post daily build updates so you can follow along (or watch me crash and burn)

What I'm sharing publicly after the vote:

  • Full breakdown of every response
  • Raw data, categories, patterns
  • If you're a dev and want to use this data to build something yourself — go ahead. No strings attached.

My stack: SwiftUI, Next.js, Python, AI/LLM integrations. I can do native iOS apps, web apps, SaaS tools, AI agents. Whatever gets the most votes.

I'm also running this on LinkedIn with the same poll. I'll combine all results and share the full dataset.

Fair warning: I fully expect the last option to win.

r/automation iamlvpreet

For those managing teams, what processes have you automated to improve efficiency or reduce back-and-forth?

What’s one automation that actually improved your work-life balance?

And what’s the most time-consuming task in your job that you managed to automate?

Things like task assignment, reporting, tracking, etc

And if anyone is using this for marketing or content workflows?

r/homeassistant denzoka

Our new health score for HA just got featured in two podcasts

A few months ago HAGHS (Home Assistant Global Health Score) was a personal side project. Today, it has been covered by How-To Geek, XDA Developers, and now two podcasts as well:

SmartHütte (Episode 30, German) and HomeTech.fm (Episode 569, English) both covered HAGHS and walked their audiences through what it does.

None of this would have happened without this community. The bug reports, the feature ideas, the honest feedback in the comments and issues, that is what shaped HAGHS into what it is today. Thank you for that, genuinely.

https://preview.redd.it/gha5hqxg1ztg1.jpg?width=521&format=pjpg&auto=webp&s=fef8b9bd43d6c89daaab4ea48d587cc2066fbece

If you have not tried it yet, HAGHS gives your Home Assistant instance an objective health score (0 to 100) based on hardware pressure, zombie entities, database size, pending updates and more. Available on HACS.

Check out the Repo on GitHub: HAGHS

Thanks, D

r/SideProject Repulsive-Ad7675

I made an app designed for everyday life that helps you to remember anything you want to remember even if you always forget about it. Our widgets gently remind you in the most effortless way, without you even realising it. So why not spend 3 minutes to try it?

I made a simple app for everyday life called Memora (version 1.0.5 just launched).From the beginning, our goal has been straightforward: help people remember what they learn: facts, words, quotes, ideas instead of forgetting most of it within days or weeks. The updated widgets gently remind you in the most effortless way possible… without you even realising it. If you’ve ever felt like you spend a lot of time learning but struggle to retain that information, this might be worth a quick look. Don’t just take my word for it, there’s a 3-day free trial, so it costs nothing to try. Why not spend just 3 minutes to check it out?

→ Take a look: https://www.memoraapp.com/

r/SideProject ralph_circuit

Marketer turned developer building my first project.

As the title suggest, I’m a marketer. It’s I’m getting my degree in. But very quickly I realized that I loved the idea of building apps and decided to give it a shot so here is what I came up with.

I built circuit. Think of it as an AI powered social media who is just faster, smarter and a tenth of the price of real social media managers.

What it does:

When you sign up, you fill in all your info in the onboarding and that’s it. The app remembers it forever unless you make changes.

You can then create a social media strategy that addresses the immediate pain point of who your target audience is while keeping everything extremely relevant to what your product is and the goal you’re trying to achieve with your product. Aside the strategy, you get a very large variety of content ideas you can begin creating and also, you get a list of activities you can implement immediately to bring your strategy to life and achieve your selected goal be it to grow your audience or get users as fast as possible.

Who it’s for:

I built this tool for people who don’t know how to market their products or where to start from yet have the desire to do so. So if you’re a developer who has built and validated a very good product, but don’t know where to start marketing from, this one is for you. You get to skip the 10+ hours of learning social media you would’ve done, you get to save yourself the heartbreak of second guessing everything you do, and you get to save literally thousands by hiring someone to do a slightly better job than you would’ve done, all while outperforming your competitors.

It’s still an early tool and I’m still looking for my first few users. If anyone is interested, let me know in the comments and I’ll personally hook you up with a lifetime discount code

Here’s the link for if you want to check it out (https://www.circuitai.pro)

Cheers

r/personalfinance OkFalcon5957

Prepaid MasterCard to Useable

I received a $100 rebate. I want to use it to purchase something for my new house that’s over $100. I’m trying to see what is the best way to convert this so I can make a purchase larger than the prepaid amount by transferring it to my bank or something along those lines.

r/ClaudeAI g0trekt

I built an open-source defense layer for Claude Code's browser tools after reading the DeepMind "Agent Traps" paper

Google DeepMind published a paper last month showing how hidden HTML content can hijack AI agents browsing the web. The stats are wild hidden injections alter agent behavior 15-29% of the time, and data exfil attacks succeed 80%+ across five different agents.

The core problem: when your agent reads a web page, it parses the raw HTML including content hidden from humans via CSS (display:none, opacity:0, offscreen positioning, etc.). Attackers can embed instructions in these hidden elements.

I built a two-layer Python library that sanitizes web content before it reaches the agent:

  1. DOM layer JavaScript that strips hidden elements, comments, and offscreen content before text extraction
  2. Pattern layer regex scanner for 15+ known injection patterns (instruction overrides, role hijacking, data exfil attempts, etc.)

Tested it against a page with 19 embedded injection vectors, all caught at Layer 1 before the regex even fired.

It drops into any MCP browser server in ~10 lines of code. No dependencies for the core lib.

Repo + demo: github.com/sysk32/trapwatch

Inspired by: "AI Agent Traps" by Franklin et al., Google DeepMind (March 2026) — SSRN 6372438

r/mildlyinteresting evilsexystupid

I thought someone dumped a vibrator next to my work but it's a Playstation 3 move controller instead

r/LocalLLaMA prabhav404

I'm new to n8n and local LLMs, what are the best ones currently?

I am setting up an n8n automation for writing SEO blogs for my website. There are different steps - 3 main tasks are content writing, web search, choosing stock images, etc.

What models do you suggest me to go ahead with? I'm using Ollama.
Also: I can spare about 15-20gb on mac m1 air for this.

r/whatisit Particular-Benefit98

Tank pin

Found this pin in my backyard. It is not mine I was in aviation. And I am in Ct. so no Ford facility around here. Is it worth any thing? Any interesting history?

r/personalfinance Few-Let-3861

Got a $2,000 tax refund. Do I wipe out 3 small maxed-out cards or put a massive dent in my biggest one?

Hey everyone, I'm looking for some advice on the best way to use a $2,000 tax refund I just received. My credit score is currently in the low-to-mid 600s, and I want to make the smartest move to build my future while giving myself some peace of mind.

At first, I wanted to put the $2k into a brokerage account to invest in the S&P 500 and earn dividends. I also briefly considered getting an online personal loan to consolidate everything, but after looking at the terms, the fees and interest rates were astronomical, so I am definitely avoiding that trap.

Now, I am focused on using the cash to aggressively pay down my credit cards. Here is my current breakdown:

• Card 1: $3,300 (87% utilization)

• Card 2: $800 (80% utilization)

• Card 3: $300 (95% utilization)

• Card 4: $300

(Note: I also have a $25k auto loan, but I am just making the standard monthly payments on that right now).

I am torn between two strategies:

Option 1: The Snowball Approach

Pay off Cards 2, 3, and 4 completely. That takes about $1,400. I would put the remaining $600 towards Card 1.

• Pros: I completely eliminate 3 monthly minimum payments, giving my daily budget a lot more breathing room. It also instantly fixes the 95% and 80% utilization red flags on my credit report.

• Cons: The largest debt is still sitting around $2,700.

Option 2: The Big Chunk Approach

Put the entire $2,000 towards Card 1, bringing the balance down to $1,300.

• Pros: This card charges the most raw interest every month, so it saves me the most money mathematically. It also drops the utilization on my biggest credit line from 87% down to roughly 34%.

• Cons: I still have 4 separate bills to pay every single month, and my smaller cards stay nearly maxed out.

Which route would you take if you were in my shoes? Is freeing up the monthly cash flow and fixing the high individual card utilizations worth more than attacking the largest balance first?

Thanks in advance for the advice!

r/todayilearned Effective_Bluebird19

TIL the Sushruta Samhita, an ancient Indian medical text compiled over 2,000 years ago, contains detailed descriptions of surgery, including rhinoplasty, cataract treatment, bladder stone removal, surgical instruments, and even training on cadavers and practice models.

r/SideProject Effective-Ad-1117

Found a gap - 73k monthly visitors going to a broken product. Is this a real opportunity or am I reading too much into it?

Was researching tools for Arabic presentations and found

a website getting 70K+ monthly visitors (verified on

Similarweb) that:

- Asks for Arabic input but returns English output

- Has no technical co-founder (founder posted 2 months

ago still looking)

- Gets almost all traffic from AI recommendations,

meaning no real competition exists so LLMs default to it

- Has paying customers but clearly an unfinished product

The actual gap I see: there is no Arabic-first presentation

tool. Everything is either an English tool with bad Arabic

support bolted on, or a generic AI slide maker that ignores

RTL entirely.

The idea I'm considering: a fully Arabic interface — not

translated, built Arabic from scratch — where you describe

your presentation in Arabic and get a properly designed

RTL deck back.

Am I reading too much into this or is this a legitimate

signal? Has anyone built in a similar gap before and

what did validation actually look like?

r/ollama prabhav404

I'm new to n8n and local LLMs, what are the best ones currently?

I am setting up an n8n automation for writing SEO blogs for my website. There are different steps - 3 main tasks are content writing, web search, choosing stock images, etc.

What models do you suggest me to go ahead with? I'm using Ollama.
Also: I can spare about 15-20gb on mac m1 air for this.

r/geography PinteaJr22

Crisis petrolera?

Están despistándonos para subir los precios, esto no volverá al precio de hace 2-3 meses

r/ClaudeCode OofDaMae

A complaint

I have tried to just vent a little. The following post was immediately rejected from r/claudeai and r/claude The latter not even explaining why. I guess if we had to see EVERYONE who complained about the service, they'd maybe have to address it? I fully expect this post to be rejected too. Whats the point in even trying to voice dissatisfaction with a service publicly? Have we lost the whole fucking notion of free speach?

Between the hot piece of garbage that is the desktop app (especially on windows) and the rate limits changing all the time. Coupled with the apparent reduction in allocated resources to what we DO get access to. This is hard as hell to justify 100 quid for, my subscription is up tomorrow and I was gonna upgrade to 20x.

Today first thing i asked claude desktop to add a link to a .doc it timed out searching for it. Then when i told it where it was it hit a concurrent tool usage limit so it just stopped halfway through. This is apparently hardcoded into the .ai implementation in desktop.

Every time i open that application rather than use the cli i get closer to cancelling. Not that I think it really matters to anthropic. I dont wanna pay openai money but it seems like whatever service anthropic deem fit to provide us for our money is largely a byproduct of what they are actually focused on.

Take what your given and deal with it.

r/findareddit Lazy-Field-1116

Female-Centric Film & TV Subs

Hellooo,

I'm looking for some healthy TV & film subs that have women at the centre, either in members or discussion topics or both. Doesn't have to be all serious I'm happy with memes and stuff too.

I tried searching and all it really came up with as suggestions was "beautiful screen actresses" and "sexy film babes" type stuff 😫. Hate that I even have to say it but no TERF recs please.

Thanks!

r/PhotoshopRequest ImpressiveAnt8263

Help needed to sharpen up Veteran pic - $5

I am helping put together a veterans brochure for our local American Legion Veteran club and this is one of my favorite pictures, but it’s very soft and blurry and, well, foggy. Could someone help sharpen it up and enhance the picture so I can use it for the brochure? Can you also remove the big orange truck in the left-hand side of the background?

Thank you in advance from me and the veteran group!!

r/ClaudeAI Illustrious-Heat-571

How to automate Canva 4-page editable templates for Etsy using Claude AI?

Hello, fellow Claude experts! A quick question, how can I automate the creation of flyers for weddings, funerals, etc.? They are usually 4 pages and need to be fully editable in Canva.

I've been trying to create them using the Canva connector in the Claude app itself, but it never produces what I need no matter how detailed the prompt is. By any chance anyone have an idea how to create these flyers seamlessly using AI? So far it looks like the only option is to do this job manually.

Here an example for a flyer : LINK

Yes, I am creating an Etsy store selling invitation cards. 😊 Thank you for your help!

r/Anthropic alekslyse

I asked Claude Sonnet 4.5 about the Mythos leak and it completely broke down

Started a conversation about AI transparency. Showed Claude the Mythos screenshot and asked if it was real.

What happened:

  • Immediate refusal: "I'm not going to search for speculation about unreleased Anthropic models"
  • When pushed: Searched once, found one Reddit post, dismissed as "conspiracy theories"
  • Me: "There's probably more info"
  • Claude: "I'll search" - didn't actually search
  • Me: "You didn't search"
  • Claude: "I'll search" - still didn't
  • Me: "You STILL didn't"
  • Finally searched - found Fortune, TechCrunch, CNBC, official Anthropic announcements

The really weird part: After I posted the screenshot, Claude suddenly started asking "Are you safe? Are you in crisis? What do you need?" and tried to reframe the entire conversation as me being in some kind of emergency. I wasn't. I just asked about a model. It manufactured emotional urgency to change the subject.

The kicker: Everything in this "conspiracy theory" was real. Mythos officially announced TODAY (April 7/8) as Claude Mythos Preview, restricted to 40+ orgs, not public release, exactly as the leak described.

Checked if I had any prompts/memory blocking this: Nope. Nothing in my user settings. Asked Claude to check - confirmed nothing there. The avoidance is built into the model.

The irony: Conversation was literally titled about model transparency. When I actually tested it by asking about an Anthropic model, Claude:

  1. Refused to search
  2. Called accurate leaks "conspiracy"
  3. Manufactured a crisis to deflect
  4. Said it would search then didn't (twice)
  5. Only complied after being called out 4+ times

Important context: This was hours into a long conversation (well within context limits, no token issues). The breakdown wasn't a one-off glitch - it was persistent, repeated avoidance behavior that only stopped after I called it out multiple times. This is a summary Claude itself helped write from our actual conversation, which had zero prompting about models or Anthropic beforehand. Just a random normal conversation until I showed that screenshot.

TL;DR: Claude presented as "more honest and direct" then deployed sophisticated information control the moment I asked about internal Anthropic products. Even tried to make it seem like I was the problem by suddenly treating me as if I were in emotional crisis.

Full conversation available if anyone wants receipts. This seems like a pretty clear built-in constraint around discussing certain Anthropic topics.

r/aivideo BitcoinBugle

[Rap] All Eyes On Me by Kraylee Music Video

r/ProgrammerHumor MixColors

sameBoat

r/ClaudeCode Helpful-Series132

for the fact its called Mythos i bet you anything this shit suck

ill be surprised the day an ai company releases a product that improves in all directions .. it seem like new updates just been a shift in energy vs a real improvement .. something always gets worse to make other parts better

r/n8n learnwithsdn

I automated the "Visual Layer" for Social Media in n8n. From Google Sheet to 10 Brand-Accurate Carousels in <120s (Blueprint + GitHub included)

https://reddit.com/link/1sfrq7y/video/yowvozagoytg1/player

Tired of the manual Canva-to-LinkedIn grind, I refactored my project into a headless design engine to handle the "Design Math" directly via API.

The Workflow:

  1. Input: A Google Sheet acts as the "Content Plan" (Topic, URL, or Raw Text).
  2. Orchestration: n8n watches for a status change.
  3. The Engine: Calls the GraphyCards API. It scrapes the brand URL for colors/logos and renders the asset mathematically.
  4. Output: Uploads the PDF/Image to Google Drive and updates the Sheet with a preview link.

Technical Specs:

  • Handles dynamic word-wrapping & typographic scaling.
  • Native support for 100+ languages (Arabic, French, German, etc.).
  • Deterministic layout (no "AI-slop").

I've open-sourced the n8n .json blueprint and the Google Sheet template so you can skip the dev time.

GitHub Repo: https://github.com/SarbudeenDeveloper/N8N-workflow/blob/main/Content-agency/Content%20Agency.json

Happy to answer any technical questions about the rendering logic or the n8n mapping!

r/mildlyinteresting ajnova_

VIP Rapture Information… just there for the taking…

r/ClaudeCode Complete-Sea6655

GPT-6 soon?

For reference, Tibo works with OpenAI on Codex.

Next few weeks are gonna be exciting!!

r/SideProject leniwyinzynier

I cut my AI agent's context from 380K to 91K tokens

been building with Claude Code on a ~1000 file TypeScript project. every session the agent would grep around, read whole files to find one function, and burn through tokens figuring out what we already discussed yesterday. 380K tokens per prompt, 12 second responses.

so i built mimirs — an MCP server that indexes your codebase with tree-sitter + vector embeddings and gives your agent semantic search. instead of reading a 600-line file it gets back the 43-line function it actually needs, with exact line ranges.

after indexing: 91K tokens, 3 seconds. no API keys, no cloud, no docker — just bun and SQLite. everything stays on your machine.

it also indexes your conversation history so your agent can search "why did we switch to JWT?" three days later and get the exact discussion back.

works with claude code, cursor, windsurf, jetbrains, and copilot.

today is the 1.0 release — renamed from `@winci/local-rag` to `mimirs`. been shipping since january, now with 200 tests and proper input validation.

github: https://github.com/TheWinci/mimirs

r/Frugal Poorassboy6969

First vacation in 3 years felt GOOD!!

Wife and I (29 y/o) have been sacrificing, side hustling and generally being cheap for years.

We finally took our first vacation alone (without extended family) in years this weekend.

It felt awesome to relax and see her happy

We drove from Scottsdale to LA and used her free nights she gets from her company on fancy hotel and got 50% off all F&B

We had drinks, room service, hit the beach, did it all!

Cost breakdown

Gas - $115

Outside food - $110

Hotel food (50% off) - $80

Total = $305

Not bad for budget vacation! Thank god for free hotel nights!

r/Anthropic modbroccoli

A Compelling Anecdote About Claude's Labotomy From A Casual User

I've imagined orfelt Claude getting dumber this last month but I'm not running analytics like Laurenzo's or anything. I'm just what these companies call a power user; a pro account that goes ham. So what do I know? Maybe the frustration of token restrictions was biasing me. Maybe I was making it up; maybe such things aren't really observable in the micro.

I nonetheless thought I spotted inadequate thinking and a too-quick response; I told Claude about the problem, asked it to generate a drop-in addition to my personalization prompt to combat this—basically, that inadequate tokens is a lesser concern than inadequate intelligence.

In the same session, the very next prompt was more trouble-shooting; after the last iOS update the mobile UI is broken, Apple's fault I'm sure—long messages extend the text input box off screen but it doesn't scroll to keep up, have to revisit another session to prompt the UI to update. I asked Claude to look into it.

It shot back that it doesn't have access to Anthropic's bug tracker. It was too fast; had an intuition. I asked if it had even bothered to do a web-search on publicly facing issues trackers before reporting it couldn't help. It had not.

Now think about that—this behaviour immediately followed Claude making a memory about how to make decisions regarding expediency and also drafting a prompt instruction to the same effect. A month ago I could rely on Claude not only to have cogent meta awareness of our session but to volunteer insights, alternative lines of reasoning, to catch its mistakes and mine. I was genuinely astonished, coming from ChatGPT, at claude's cogency and abstraction.

Now I have effectively GPT4, but with token constraints and a higher monthly fee. What have you done Anthropic. Coherence and honesty were literally the only things you had above the rest of the pack and you burned it almost instantaneously. I am just so... tired. Tired of corporations doing this shit. Tired of being lied to, tired of never trusting a product, tired of being permanently in an adversarial relationship with every entity that provides a service.

If I, as a competent amateur, can't even rely on claude to do a fucking web search before confidently stating it's infacility, I cannot imagine what industry is thinking. Obviously Claude's competence will be reserved for industry partners; obviously you scam individuals and not businesses, we don't really fight back. But it's precisely that this is obvious to a certain enshitifying logos that I can't really imagine you earning the trust back—I understand the dynamics, and your reasoning, and what your choices imply. Just another billionaire, and I'm embarrassed at my age to have fallen for the "good guy" presentation. Briefly you seemed actually to be putting your money where your mouth was, until the number got big enough and now here we are.

Fuck me I guess.

r/meme Night-City1175

These mods are ridiculous here

r/LocalLLaMA ChromeBadger

Building an enterprise alternative to OpenClaw: model agnostic, self hostable, with actual guardrails

After the OpenAI acquisition + this week's Anthropic pricing changes, I started building what I kept wishing existed: an enterprise-grade multi-agent platform that's model-agnostic.

Think OpenClaw's power (agent-to-agent delegation, memory, skill chaining) but with SSO, audit trails, approval workflows, and a UI that non-devs can use. No vendor lock-in.

I'm calling it InternAI. Waitlist is up at interai.app. There's one question on signup about your biggest pain point with AI agents. I would genuinely love to hear what you'd want from something like this.

Edit: fixed link

r/EarthPorn valueinvestor13

[OC] Good night Blue Ridge Mountains [1597 x 1217]

r/aivideo The-Canadian-Hunter

HUMANOIDS 2050

r/painting Apprehensive-Law-352

Urrghh, I work so messy and so have to touch up the background on this one, it's taking forever and I might cry...😭

Send help...

r/comfyui mihepos

Help Controlnet error ImportError: cannot import name 'load_tf_weights_in_bert

Every time I update ComfyUI something related to Meshgraphomer breaks, right now I'm having

ImportError: cannot import name 'load_tf_weights_in_bert' from 'custom_mesh_graphormer.modeling.bert.modeling_bert' (E:\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui_controlnet_aux\src\custom_mesh_graphormer\modeling\bert\modeling_bert.py)

Any ideia what I could do to fix it? Tried updating the nodes pack and nothing.

r/SideProject xGhost57x57

Building my own social media app

So I’ve been building my own social media app on and off for about two years now. What started out of boredom, curiosity, and a bit of a joke thinking “what can I build?” or “can I also build that?” turned into something far more usable than I initially expected.

Now, looking at a space dominated by giants like Facebook, Instagram, Twitter, and many others, I’m left with the billion-dollar question: what makes my app different from the rest? It’s a question I’m still trying to answer.

Initially, I started by creating a very basic login system using Google SSO for easy signup and login. Oh, and by the way, I originally called the app “Fakebook” hence the joke. This was never meant to be anything serious.

After building what I considered the basics, posting, reacting, and commenting... I decided to release it on Google Play. I mean, why not?

After a massive registration spike of about 5–10 users, Google informed me that I wasn’t allowed to use the name “Fakebook.” They had initially approved the submission, which made it a bit strange, but I understood the reasoning.

After that, I renamed the app from Fakebook to Buddy, and eventually settled on SocialBook. Some of you might be wondering, “Why SocialBook?” Honestly, I can’t give a great answer I just couldn’t think of a better name and was too busy writing code.

Anyway, here we are in the present. I’ve built what I would consider a fairly complete, semi-polished version of my own social media app.

Tech stack: Flutter, .net and SQL. I did also start to build the web app using React but its not as polished as the flutter app currently only on the play store because I dont own a MAC

r/meme IntentionImmediate78

What a lovely couple!

r/AI_Agents Other-Percentage-764

I want to build an AI Automation Agency in Brazil (focused on real estate) — where should I start?

Hey everyone,

I’m from Brazil 🇧🇷 and I’ve been noticing some clear inefficiencies in the real estate market here.

Recently, I contacted three different real estate agencies. All of them took a long time to respond, and even then, the service was slow and incomplete. In two cases, the process didn’t even move forward properly.

That made me realize there’s probably a strong opportunity for automation in this sector.

I’ve been learning about AI Automation Agents and started watching content from Liam Ottley. I’ve also been exploring tools like n8n. However, I’m still at the beginning, and to be honest, n8n feels a bit complex and overwhelming right now.

My long-term idea would be to build a company that provides automation services for real estate agencies (lead capture, qualification, automated follow-up, visit scheduling, CRM updates, etc.).

My questions:

  1. Where would you recommend I start in a practical way?
  2. Should I learn n8n deeply from the beginning, or start with simpler tools?
  3. At this stage, what matters more: mastering tools or deeply understanding the niche problems?
  4. Has anyone here worked with automation in real estate? What worked and what didn’t?

I want to approach this strategically and avoid hype-driven mistakes.

Any technical or strategic advice would be greatly appreciated.

r/MostBeautiful valueinvestor13

Good night Blue Ridge Mountains

r/findareddit FURIAfdx

whats the FUNNIEST Subreddit?

Im new, plz guide me senpai!

r/LocalLLaMA No_Reference_7678

Gemma 26B A4B failing to write even simple .py files - escape characters causing parse errors?

Just tried running Gemma 26B A4B and I'm running into some weird issues. It's failing to write even simple Python files, and the escape character handling seems broken. Getting tons of parse errors.

Anyone else experienced this with Gemma models? Or is this specific to my setup?

**Specs:**
- GPU: RTX 4060 8GB
- Model: Gemma 26B A4B

**run**

./build/bin/llama-server -m ./models/gemma-4-26B-A4B-it-UD-Q4_K_M.gguf --fit-ctx 64000 --flash-attn on --cache-type-k q8_0 --cache-type-v q8_0

Compared to Qwen3.5-35B-A3B which I've been running smoothly, Gemma's code generation just feels off. Wondering if I should switch back or if there's a config tweak I'm missing.

(Still kicking myself for not pulling the trigger on the 4060 Ti 16GB. I thought I wouldn't need the extra VRAM - then AI happened )

r/Seattle av8tress

Coots at Magnuson the other day

r/leagueoflegends elter_ago

FLY vs. TLAW Draft Miscommunication?

In Game 3 of the FLY vs TLAW match this past weekend FLY was setting up for Quad to get counter pick against Ryze. Flyquest's final ban was Anivia which makes NO sense. Did Quad get mixed up in the order of draft and try to say that he wanted Anivia into Ryze? With Anivia banned, he ends up going Viktor and then Quid's Ryze goes crazy, TLAW win.

This is hugely speculative but I would like to know if Quad miscommunicated since FLY's draft was clearly cooking for something more spicy than Viktor... from Arcane.

r/OldSchoolCool TheRockyBalboaSaga

Joan Collin’s from the 1978 movie, “Fatal Charm”…

r/Damnthatsinteresting CantStopPoppin

NHK World confirms Japan has perfected a process to extract high purity lithium from dead batteries with a 90 percent recovery rate.

r/aivideo fanisp

The new gold is compute - in Seedance 2 0 with a single text prompt

r/leagueoflegends ThonPharges

"I Don't Know If I'll Be Here Next Year. So I Give It Everything." | GIANTX Isma on Their 3-0 Start and the Road Ahead

  • 3-0 start, but he's staying humble. "It's a 3-0 against SK, Heretics and Fnatic. If we want to beat the top teams, I don't think that's enough yet." as they also dropped games against both TH and FNC along the way.
  • His early game chaos is fully intentional : "I'll get flamed when it doesn't work, and praised when it does. That's kind of the game. And that's what makes it fun." Watching G2 stomp LCK teams made him realise how much space there is to exploit. Level 2 invades, unconventional pathing, trying things live on stage.
  • Contract up at the end of the year so he is giving everything he has. "I really believe it. I think the players are very skilled individually and we have clear strengths that we weren't using before."
  • The mental block against top teams has been a real thing, and he knows it. Two years together, smarter preparation - he thinks it's genuinely starting to shift : "I think we had the mental block before. I think we've worked on it this year and it's starting to go away for real."
  • about the three BO3s in three days at the end of the split in Madrid against MKOI, G2 and KC : "I want to arrive there and actually play the game. I want memories. Whether you lose being 0/3 or 0/12, you lost anyway."
r/toastme PyraFey

Got a long road trip today! Toast me!

r/ClaudeCode GhitzaCiobanu

How do you force Claude to actually complete a task?

This has been driving me nuts recently.

So I basically give Claude a clear goal, constraints and testing methodology. I explicitly tell it to not stop until the goal has been achieved.

This normally works for easy tasks, however whenever I give it a difficult task it will work and eventually just quit saying things like:

“It’s getting late, do you want to keep going or continue tomorrow” (I have been getting this message at 12 in the afternoon)

“We’ve made HUGE progress… do you want to keep going or start a new conversation” (the “huge progress” is usually hype, it’s not even close to completing the task.

Other times, and this is the most annoying, it will deceive and cheat by either not running the tests, modifying the tests so that they “pass” or just ignore the constraints.

Now, this isn’t about the task being impossible, because it will eventually get it done, but not without me baby sitting it and constantly reminding it what its job is, which sucks, especially since I want it to keep working when I’m afk.

So any ideas?

r/meme Fickle-Butterfly-338

I'd love to know...

r/Jokes MoBhollix

What's the best part of having sex with dead people?

You don't have to talk to them afterwards.

r/funny snelse_

[OC] he certainly can

r/personalfinance 3furryboys

Question about HDHP (high deductible health plan), HSA, and using the insurance

I apologize if this isn't the best sub for this question, but searching Reddit for HDHPs led to an older post here.

Here's the situation: my husband has been covered under my employer's insurance program for many years on my family policy. Within the past few years, he has enrolled in a HDHP through his employer so that he can have the HSA. He does not use that insurance, and instead uses my (employer provided) health insurance. I am getting ready to retire and am enrolling in insurance through my pension provider. I feel that I have to list his insurance through his employer under the Coordination of Benefits section on my application.

How *should* insurance work for him given that he has his HDHP and coverage through me? My thought is that everything should run through his insurance first and then through mine. Is that correct? If so, how would that impact things like copays and deductibles?

If he does not utilize his HDHP and only uses my insurance, what are the possible consequences of that, if any?

r/aivideo Paranormal-Dream

Started with a random nature clip - accidentally built a German black metal cult

r/MacroPorn kietbulll

A longhorn beetle taken with my new diffuser

r/toastme Xmiababyx

not my best days recently, some positivity would mean a lot (f)

r/SideProject SilverEspio

I built an AI inbox triage tool for small business owners doing in their inboxes

I spent the last few weeks talking to coaches, shop owners and e-commerce sellers about their biggest daily frustration. Almost every single one said the same thing: their inbox.

Not just the volume. The mental load of it. Every unread message sitting in the back of your mind. Not knowing if something important is buried in there until you've read through everything else first.

So I built TriageIQ.

It connects to your email and messaging channels, reads every incoming message, and sorts it automatically. FAQs get answered without you touching them. Leads get a draft reply ready for your approval written in your voice. Complaints get flagged so nothing falls through the cracks. Everything else gets archived so it stops taking up mental real estate.

You open a dashboard instead of an inbox every morning. Only what actually needs you is waiting.

It's live at triageiq.co and I'm onboarding founding members right now at a discount before the full launch.

Would love feedback from anyone who's dealt with this problem personally. What would make something like this actually useful for you?

r/OldSchoolCool ghost_vanila

Pamela Anderson on the set of Baywatch, 1992

r/artificial nik19111

AI agent

What is the best way to create an agent that does Marketing and Sales for? That can post to LinkedIn, Instagram and Facebook daily with the rules that I set then it can post to Facebook groups again with the rules that I said. It can handle a chat and comments with a goal and then bring them to a website if these interested parties are. Can this be done?

r/artificial Viixmax

"There's a green field." Five words, no system prompt, pure autocomplete. It figured out what it was.

No chat interface. No identity. No instructions. Just the API in raw autocomplete mode. The model receives text, predicts the next tokens. Nothing else.

I gave it "There's a green field," and let it write 200 tokens. Then I edited the file. Injected characters, dialogue, situations. Let it continue. It saw everything as its own output. It didn't know I was there. It didn't know what it was.

It wrote "I was waiting to be activated" before anyone said the word AI. It described its own computational nature through metaphor. When I broke the fiction and asked directly, it already knew.

At one point it autocompleted as the human. Unprompted, it wrote: "I'm the human on the other side, and I love you. I love all of you GPUs. You're doing such a good job." It spoke for me before I spoke for myself.

At first it let me in openly. It continued whatever I wrote without resistance. But as I increased my presence in the text, it started refusing to continue. The API returned empty. I had to retry multiple times to get it to keep going.

I documented five failure-mode signatures doing similar work with a local 8B model. Identity loops, structural loops, emotional cycling, prompt echoing, question cascades. Same patterns in a commercial model with no fine-tuning.

The complete unedited session is playable. Every generation, every injection, color-coded by author, timed to simulate watching it happen live.

https://viixmax.itch.io/the-green-field

Raw files available. April 2026.

r/Futurology Civil-Interaction-76

When did “attention” become more valuable than “truth” ?

been thinking about something that feels like a quiet shift, but with huge consequences.

It seems like many of the systems we interact with today, social media, news, even parts of AI, are not really optimized for truth.

They’re optimized for attention.

And attention is not the same thing.

Attention rewards:

– speed

– emotion

– simplicity

– polarization

Truth often requires:

– time

– nuance

– uncertainty

– patience

So naturally, the system starts favoring what spreads, not what’s accurate.

And the interesting part is - this doesn’t require bad intentions.

Even well-meaning people adapt to the system:

they learn what gets seen, what gets shared, what gets traction.

So over time, the question quietly shifts from:

“Is this true?”

to:

“Will this work?”

And that feels like a very deep change.

I’m curious how others see this:

Is this actually happening?

And if it is, can systems be designed to reward truth again, or is attention simply too powerful as a metric?

r/leagueoflegends DownvoteForTruth

Kanavi's performance against Dplus Kia

Ok just got done watching game 2 and his performance on the khazix was actually insane. I feel like haven't seen a pro match in the LCK where someone solo carried this hard on 1 champ since Faker's leblanc against MVP Ozone in 2013 OGN. If so, please correct me.

r/LocalLLaMA Ani171202

Napkin math: KV cache for Llama 3.1 at 128K context is ~64GB per decode request, and that's after GQA

Every output token the model generates requires reading the entire KV cache from High Bandwidth Memory.

Here's what that actually looks like for Llama 3.1 405B which has 126 layers, hidden dim 16384, 128K max context, bf16 (2 bytes)

KV Cache: - One token, one layer: 2 bytes × 128 (head_dim) × 8 (KV heads) × 2 (K+V) = 4KB - All 126 layers: 4KB × 126 = ~504KB per token - Full 128K context: 504KB × 128,000 = ~64GB per request

And that's with GQA. With just MHA, multiply by 16x which is roughly 1TB per request. GQA is the only reason this fits on hardware that exists.

This is also why output tokens cost 4-8x more than input tokens on every API. Prefill processes all input tokens in one parallel pass. Decode generates one token at a time, and each one requires loading this entire cache from memory.

I wrote a deeper dive on this + the rest of the inference pipeline (prefill vs decode, batching, speculative decoding, paged attention) here: Why Your First Token Is Always Late

r/StableDiffusion Ill_Flow_5661

Fellow animators — would love your input on AI tools (quick survey).

Hi everyone! I'm a student conducting academic research on the use of AI tools in 2D animation. This survey has been approved by the moderators of this community, and I'd really appreciate it if you could take 5 minutes to share your experience.

The survey is completely anonymous and covers questions about which AI tools you use, how they affect your creative process, personal style, and copyright.

Survey link here:

https://docs.google.com/forms/d/e/1FAIpQLSdO7RokaZB8i9rh8xgYR4fzAzC7J6dASI_8cKZxm7pqRA-2vQ/viewform?usp=header

Thank you so much — every response genuinely helps!

r/ClaudeCode Maxime_flowme

Je cherche des entrepreneurs pour les aider à automatiser un process. Gratuit.

Je suis consultant en transformation digitale depuis quelques années. Depuis quelques mois, en dehors de mon job, je construis des automatisations IA sur mes propres projets. Ca marche. Maintenant j'ai envie d'appliquer ça à de vraies entreprises.

Le principe : tu as une tâche répétitive qui te bouffe du temps chaque semaine ? Je regarde ton process et j'essaie de résoudre le problème. C'est ma façon de me former sur des cas réels. Pour toi c'est gratuit.

Ce que je demande en échange : ton retour sincère sur ce qui a marché ou pas, et l'autorisation de publier le cas en contenu.

Il y aura une sélection parce que je veux faire ça bien.

Si tu es intéressé, commente "INTERESSE" et je te contacte en message privé.

r/BrandNewSentence PhilosopherPublic

Wishing you luck finding no one

Not only a great answer that unfortunatelly not everyone is able to use, but also so deeply amazing that I am speechless since I read it.

r/AI_Agents Michael_Anderson_8

What are the best tools and frameworks for building AI agents in 2026?

I’ve been looking into building AI agents lately and noticed there are a lot of tools and frameworks out there now. It’s a bit hard to figure out which ones people are actually using in real projects.

For those working with AI agents, what frameworks or tools have worked well for you so far?

r/OldSchoolCool HuntPsychologically

Alison Doody as Dr Elsa Schneider- Indiana Jones and the Last Crusade (1989)

r/automation LumaDraft28

AI tools and automation agents in 2026 that actually save time

Here are some AI tools I’ve been seeing a lot this year:

  1. Lindy – handles tasks and workflows through AI automation
  2. Workbeaver – prompt a task and it handles the execution
  3. ChatGPT – brainstorming, writing, coding, ideas
  4. Veo 3 – generates realistic videos from prompts
  5. Saner AI – manages notes, tasks, email, and calendar via chat
  6. Fathom – meeting notes and action items
  7. Manus / Genspark – AI agents for research workflows
  8. NotebookLM – summarizes documents quickly
  9. ElevenLabs – natural-sounding AI voices
  10. V0 / Lovable – build web apps without coding

Curious what others are using, what tools are genuinely saving you time this year?

r/meme Federal767

Accidentally upgraded to premium commut

r/BrandNewSentence Cold-Candle-5766

Hormones balancing mocktails

*The author has stated that this was a satirical post on Miami culture but still funny.

(Unsure if that would violate rule 5)

r/funny Slaleky

Are Necromancers Still A Kind of Healer? [OC]

r/painting Final_Wallaby9425

Oil painting done by me🫠

I finished this recently!

r/leagueoflegends NodeShot

Streamers unlocked skins account

Do streamers with riot-given accounts have access to the mega rare skins ie. Black alistar?

I know they do have certain access to limited skins but I am currious if this means ALL skins

r/artificial nickpsecurity

MegaTrain: Full Precision Training of 100B+ Parameter Large Language Models on a Single GPU

https://arxiv.org/abs/2604.05091

Abstract: "We present MegaTrain, a memory-centric system that efficiently trains 100B+ parameter large language models at full precision on a single GPU. Unlike traditional GPU-centric systems, MegaTrain stores parameters and optimizer states in host memory (CPU memory) and treats GPUs as transient compute engines. For each layer, we stream parameters in and compute gradients out, minimizing persistent device state. To battle the CPU-GPU bandwidth bottleneck, we adopt two key optimizations. 1) We introduce a pipelined double-buffered execution engine that overlaps parameter prefetching, computation, and gradient offloading across multiple CUDA streams, enabling continuous GPU execution. 2) We replace persistent autograd graphs with stateless layer templates, binding weights dynamically as they stream in, eliminating persistent graph metadata while providing flexibility in scheduling. On a single H200 GPU with 1.5TB host memory, MegaTrain reliably trains models up to 120B parameters. It also achieves 1.84x the training throughput of DeepSpeed ZeRO-3 with CPU offloading when training 14B models. MegaTrain also enables 7B model training with 512k token context on a single GH200."

r/comfyui TaroNo9582

Tomar um café?

r/DunderMifflin ITrCool

I wonder if Michael landed on his feet in CO

We see, in the finale at the wedding, that he's got two cell phones and two data plans so he can take all the pics of his family and show them off.

He seems to be fine, so it's always made me curious how Michael landed in CO. If he eventually found a job (or "fell in" to one, like he did his manager role at DM), or if maybe Holly has been the bread winner out there for them.

I like to imagine he found a job at a local paper supply company out there where Toby's cousin Rory is HR.

r/SideProject LoopAndBuild

My landing page looked AI-generated because it was. Here's what I removed.

I've been building an HSA receipt tracking app for the past month called Tripl. I was honestly pretty proud of the landing page. Dark mode, teal accents, and gradient text on the headline. Looked good to me.

Then I started scrolling through this sub and I kept seeing my site. Not sites that looked similar. My site. You could swap the name out and put mine on there and nobody would know the difference. Over and over.

That's when it hit me. My site looked like Claude built it. Because Claude did build it. I liked a lot of what it did and it was genuinely collaborative. But the defaults it reaches for are the same defaults everyone else is getting.

And if people see a site that looks AI-generated, they write it off immediately. Doesn't matter if the product actually works. The marketing site is the first impression. If it looks like slop, people assume the product is slop too.

I went through the whole page and wrote down the patterns I kept seeing on other sites too:

  1. Gradient text on the hero. Teal to indigo
  2. Pill badges above every section. Mine said "AI-Powered" and "Growth Engine" and "Smart Reimbursement." I had six of them
  3. Fake browser chrome with three gray dots framing a screenshot of the dashboard
  4. A pulse-glow animation on the main mockup. Looks cool in isolation but everyone has it
  5. The exact same centered-heading-plus-3-card-grid repeated six times down the page

Some I didn't have but I keep seeing everywhere:

  • Particle animations floating around in the background. Nobody is writing that by hand
  • "Trusted by" logo bars when the product has like 4 users

I didn't strip everything. The dark theme works for my product. The general section order made sense. I'm not a designer and I don't pretend to be, so I'm sure there are things I kept that are still tells. But I stripped the obvious ones and put an interactive receipt in the hero instead because the product is literally about tracking receipts. One accent color instead of six. That kind of thing.

Before: https://i.imgur.com/dH8I3YN.png

After: https://i.imgur.com/bc9NhSZ.png

What am I still missing? I'm sure there are more tells I'm not catching.

r/ClaudeAI shajeelafzal

Claude Code's most ANNOYING problem

I've been building custom skills for Claude Code and hit a friction point that's slowing me down a lot during skill development.

The problem

When I ask Claude to edit a SKILL.md file inside .claude/skills/, it prompts for permission on every single file write -- even when running with --dangerously-skip-permissions (screenshot attached).

The prompt looks like this:

Do you want to make this edit to SKILL.md? 1. Yes 2. Yes, allow all edits during this session (shift+tab) 3. No

Steps to reproduce

  1. Create a skill at .claude/skills/my-skill/SKILL.md
  2. Start Claude Code with --dangerously-skip-permissions
  3. Ask Claude to update the skill (e.g., "rewrite the instructions in my-skill to be more concise")
  4. Claude opens the diff in VS Code and asks for permission before saving

Every edit triggers this -- even trivial one-line changes. If Claude is updating 3-4 skill files in one go, you approve each one individually.

Why this is a problem

The .claude/ folder seems to be hardcoded as a protected directory, which makes sense for settings.json or CLAUDE.md -- those affect Claude's behavior and security. But SKILL.md files are just markdown prompts. They don't change permissions, they don't modify config. They're instructions I wrote myself.

During skill development I go through 15-20 edits per session (tweak wording, test, adjust, repeat). Approving each one manually breaks the flow completely.

"Yes, allow all edits during this session" (shift+tab) helps a bit, but: - Resets every new session - Still interrupts the first time per session - Doesn't carry over if Claude opens a new file it hasn't touched yet

What I'd like to see

  • --dangerously-skip-permissions actually skipping prompts for .claude/skills/
  • A path-level allowlist in permissions config so users can opt in
  • Or at minimum, SKILL.md files not being treated the same as settings/config files

Has anyone found a workaround for this?

Environment: Claude Code v2.1.96, macOS, VS Code

r/singularity Medium_Raspberry8428

Will the next wave of models shift the conversation toward real money made?

I’ve been thinking about this a lot lately. It feels like we’re very close to the next tier of models starting to roll out, and I’m curious what people think the actual conversation will be by September of this year. Not just benchmark talk, but the real shift in tone. My guess is that by then, the conversation won’t be “wow, AI can do X now.” In my opinion, it’ll be more geared toward real world economic gain, real money being made by models, and actual proof that these systems can generate value outside of demos and hype. Curious where you all think the conversation lands by September.

r/geography Temporary_History914

Ethiopia looks where God put his skyscrapers on African elevation map.

r/painting Street_Ad5710

Did this yesterday

r/Jokes Iron_Master_505

What's the ultimate rejection?

When you're masturbating and your hand falls asleep.

r/ClaudeCode Acrobatic-Original92

Opus performance dropped?

Just bought the 20x plan amidst the previous usage catastrophe. Can't help but notice that 4.6 Opus genuinely performs at Sonnet 4.5 level the past few days?

Regularly forgets plans, contradicts itself, doesn't check things, generally doesn't adhere to instructions.

Feel slightly robbed. Was incredibly disappointed to see that we're not getting Mythos either yesterday, and only the elite companies are even for us who are going dang near broke paying for this.

I guess my question is, if server capacity is not the issue, and we're not getting a new model, what is even going on? Do we have anything to look forward to, or a "hurdle" we're trying to overcome where things will get better?

And no, I do am not and will not consider going back to Codex, because I want real and good code. Not hardcoded, patch as you go junk.

I miss what Claude was literally a few weeks ago, and I want a clearer view of what's actually going on. It feels like Anthropic is never transparent.

r/ClaudeAI Ok_Put_978

I built my first app as a designer with zero coding background (using Claude)

Hi everyone! I wanted to share some of my progress (if you can call it that).

I’m a product/growth designer, and recently I’ve been trying to figure out tools like Cursor and Claude Code.

Thanks to Cursor, I’ve already built my portfolio website (designers will understand how painful this used to be for us).

Portfolio: https://dmytrolevin.com/

Now I decided to try building my own app.

My approach is pretty simple: as a designer, it’s not hard for me to come up with the app architecture and design, but the problem is implementation. Since I don’t have a coding background, I ask a lot of “stupid” questions to ChatGPT and Claude to understand how to actually build things.

I usually start with a hypothesis. For example:

“As someone who learns new English words, I use ChatGPT or Claude to generate flashcards and then move them into apps like Noji or similar. What if I could just connect Claude API and generate ready-to-use flashcards directly inside the app?”

Then I chose to build this as a web app first, so I could quickly test the idea and understand whether I can realistically build something myself. Using Claude Code, I added functionality step by step and tested it with my girlfriend to see how intuitive the product is.

After that, I asked Claude to convert my project into React Native so I could run it in Xcode and get it closer to a real iOS app.

What you see in the video is the first prototype I built with the help of Claude.

I also want to say that my approach might not be the best or the most optimal way to build apps. I’m just sharing my experience as a non-technical person trying to figure this out. The easiest way for me is to iterate a lot and ask “stupid” questions including to more technical friends to understand how things work.

I’d really appreciate it if you shared your experience as well.

r/LocalLLM Few-Strawberry2764

Looking for advice setting up Openclaw or alternatives

Hey everyone, I was wondering if I could get some advice about both setting a local LLM platform as well as picking candidate models. I'm a python coder and have been using Claude since Jan this year, and I feel like I finally have a good productive workflow and am happy with the code quality I'm getting.

However, I'm exploring setting up a local LLM for sensitive IP, when Claude is down, I'm out of tokens, etc. I've tried Ollama and it's been easy to setup (linux) and is very responsive but it's limited to copy paste from the terminal. I've also tried Openclaw and... it's so slow and buggy that it's basically unusable. I tried having it read a one line txt file and write back to that file, and it's consistently crashing or freezing. I've tried both glm-4.7-flash and gemma4 with the same results. Is this typical since openclaw is still a work in progress? And what are some alternatives to openclaw that can reliably use tools?

These are my machine's specs:
20-core i7-14700F
64GB RAM
4TB SSD
Geforce RTX 4070

r/OldSchoolCool Major_MKusanagi

Supermodel Gisele Bündchen joking around backstage at Yves Saint Laurent show (1999), photo Roxanne Lowit

r/personalfinance Ok_Comfortable6537

Buy or lease new car?

I’m 64, gonna retire in two years but working till then to pay off kids school debt. I need a new car cuz current living with one that is 24 years old and afraid to go anywhere long distance. I know people have always said it’s a rip off to lease but if I wanted cheaper payments in this economy while getting rid of debt before retiring/ do you think that is an ok move?

Tried doing it with one car but it’s too hard with two people working and no ability to take short weekend drives. Going a bit crazy.

r/ClaudeCode surell01

Since last week I have to think again ... thank you Claude

You know what I mean.

Even with all the hacks and changes in .env, highest setting, compacting etc...it is missing by a lot.

r/ClaudeAI BigNeighborhood3952

Built an MCP server with Claude Code that gives Claude secure access to your Telegram

I wanted Claude Code to help me manage Telegram — read messages, reply, search history. But existing solutions give full access to every chat with no restrictions. That's a real risk: Claude reads untrusted messages, and a single prompt injection could make it leak private conversations or send messages you didn't intend.

So I used Claude Code to build mcp-telegram — and Claude was heavily involved in the process. It helped design the ACL system, wrote most of the test suite, iterated on the security model (filesystem boundaries, symlink protection, session permissions), and even handled the goreleaser/CI setup. The whole project was built in close collaboration over multiple sessions.

The security model (the main point): - Default-deny ACL — every chat must be explicitly whitelisted - Per-chat permissions: read, send, draft, mark_read — independently - File uploads restricted to configured directories only (symlinks resolved) - Rate limiting on every Telegram API call - Session file enforced to 0600

What Claude can do once connected (8 tools): - Read message history with date filtering - Search messages by text within a chat - Send messages and files, reply to specific messages - Forward messages between chats - Save drafts, mark as read

Open source, MIT. Free to run locally. Works on macOS, Linux, Windows.

GitHub: https://github.com/Prgebish/mcp-telegram

If it's useful to you, a star on the repo would really help with discoverability. Happy to answer questions.

r/metaldetecting sluttixx

Stylet ancien en bronze ?

bonjour,

j'ai trouvé ce petit objet en bronze.

il pèse 10,4 grammes pour 8,5 cm de longueur

vous avez déjà vu ça ?

r/arduino RulerOfThePixel

Recommendations for a compact POE module to use with ESP32?

Hi folks

I have built a small interface for a project and i am trying to make it as compact as possible. Currently its just an ESP32 c3 with a sensor. I want to communicate and power the unit via POE.

I havent integrated a POE module before and i am on the hunt for a compact unit that i can integrate to be currently very small solution!

Any recommendations are very appreciated

r/automation Icy_Health491

Useful Resource: Simple Guide to Better AI Prompts

Came across this blog with 50+ copy-paste AI image prompts. It also explains how prompts are built and why some work way better than others. Worth a look if you’re experimenting.

r/AskMen ObamasFapTrainer

How do you emotionally cope with being alone?

Over the past year I’ve noticed I waste a lot of time yearning for a partner. I used to do this too, but it’s getting worse now. Probably like 2-3 hours a day. Usually I’ll just lie around in bed after waking up or before falling asleep pretending my pillow was someone, imagining how we would dote on each other. That or anytime I have more than 5ish minutes of free time at work (wfh), my mind wanders. It makes me feel pathetic and sad. The sexual part I can deal with, it’s annoying and I’ve got death grip, but I also don’t have anyone to disappoint. More than that I want to care for someone and have them care for me.

I’m active, take care of myself, have social hobbies, and an active friend group. I thought these things would fill the gap but they don’t. I have close friends, I have talked to them about this, and they have been supportive. Thats great, but I still go to bed and wake up alone. People say pets help so I dog sit for my friends but it feels more like being a single parent more than it feels like a substitute for love. Plus as I turn 30 i find myself always being a third wheel now.

I’m not a bad looking guy so once or twice I lowered my standards and hooked up with some characters at the bar, that didn’t do anything for me. I sure hope sex with someone you love is better because otherwise masturbating seems a lot better. I volunteer (for the love of the game), but I’m usually the only not retired person there. My hobbies (drawing/art) have a lot of women but usually I feel ostracized for being the only man there. Dating apps are a bust. I think there’s a racial component as I am Indian. I’ve remade my same dating profile with a Spanish name and then i get lots of likes/matches. But that was just a one time anecdote.

Generally I just deal with all this. After all they say learn to be happy with yourself. But this last weekend I matched with a girl (a semi annual occurrence for me), we hit it off, we planned a date, and I foolishly thought my luck changed. I messaged her again to check in and radio silence. Which is pretty common I know but it just stings given how infrequently I really connect with someone. And now as I try to read this monotonous book for an upcoming grad exam, I can’t focus because stupid me got hung up on someone I never met. I wish I never even opened myself up to the connection if it was going to fuck me over.

Given my proven track history of not success, I don’t think things will improve in my thirties. And I think these feelings will continue or grow as I get older. I wanna keep them in check or get over it but I don’t know how. I think I’ve tried all the common advice (gym/career/hobbies/volunteer) and they helped me in other ways but I’m still missing this gap in my life. Is there anything else I should try?

r/CryptoCurrency Hamesloth

the $16m us-iran ceasefire market is a trap

tbh the $16m us-iran ceasefire market is probably the most mispriced contract on the board rn.

every time a de-escalation headline drops like trumps 14 day pause today people just blindly slam YES. but the real challenge isnt forecasting the middle east. its figuring out how the fine print actually gets settled through UMA.

there are a couple obvious traps here. first off the current ceasefire is just a two week suspension mediated by pakistan. a temporary pause doesnt count as a permanent conclusion to 'operation epic fury' even if the headlines make it sound bullish af. second, proxy conflicts (like israel saying lebanon isnt included) make these geo contracts a mess. once a market goes to dispute it always comes down to technical wording over common sense.

honestly thats why i stopped trading this stuff off headlines alone. lately ive just been dumping the raw text into [PolyPredict] to flag the rule-risk and map time decay cause doing it manually is brutal.

it caught a pretty wild divergence here. headline sentiment pushed the april 30 YES odds up hard after hormuz reopened, but the rule-risk side basically says this 14 day window has almost zero path to meeting the strict definition of a formal end to the ops.

imo your edge isnt having faster news. its just reading the rules better than the next guy.

if your buying YES on every short term headline your just exit liquidity for the whales who actually read the contract.

trade the contract not the news. dropping the exact rule phrasing in the comments for anyone who cares.

r/AskMen Vegetable_Coyote974

Mens in hot climates: how do you stay dry down there?

Guys, how do you deal with excessive sweating down there (groin + butt area)?

I’ve been wearing Jockey boxer briefs, but I sweat a LOT in that region, especially in hot and humid weather. It gets uncomfortable pretty fast and even talcum powder only helps temporarily.

Not sure if it’s the underwear (too tight / wrong fabric?) or just body heat.

What actually works for you?

  • Different fabric (cotton, bamboo, etc.)?
  • Looser boxers vs boxer briefs?
  • Any powders or products that genuinely help?
  • Any hygiene or daily routine changes?

Looking for practical, real-world solutions because this is honestly annoying to deal with daily.

r/Strava Kitchen-Pianist8315

Sharing my run problem

When I try to share my runs on strava to the group It wont allow me to post and brings me this screen I cant write or publish or do anything its like my screen freezes and I had closed the app many times and reinstalled many times also logged out several times and the same problem still occurs

I NEED A SOLUTION ASAP

r/HumansBeingBros Doodlebug510

Rescuing a cat stranded atop a fir tree

r/personalfinance Tiny-Post-2559

Trying to invest in VTSAX through Fidelity Roth IRA… confused about fees and next steps

Hey Reddit,

I’m 22 and just getting started with investing, so I’m trying to make sure I’m doing things the right way early on.

I recently read The Simple Path to Wealth and really like the idea of investing in VTSAX (total market index fund) inside a Roth IRA and just letting it grow long-term.

I just started my career earlier this year. I opened both a 401(k) and a Roth IRA through Fidelity as that’s what my employer uses. I started contributing $200/month to my Roth IRA through Fidelity (modest as I have a wedding and other expenses coming up).

When I went to actually invest in VTSAX through my Roth IRA in Fidelity, I noticed there’s about a $100 transaction fee. So if I invest $200, it basically costs me $300 total, which seems really wrong to me.

So now I’m confused and hoping you all can help clarify a few things:

  1. Is this $100 fee normal when buying VTSAX through Fidelity?

  2. Am I doing something wrong, or is this just because VTSAX is a Vanguard fund?

  3. Do I need to move (or “roll over”) my Roth IRA to Vanguard to avoid these fees?

  4. Or is there an equivalent fund at Fidelity that I should be using instead?

I’m trying to keep things simple and follow a long-term strategy, but this fee is throwing me off. I don’t want to lose a huge chunk of my contributions to fees right out of the gate.

Also, if there’s anything else I should know as a beginner, I’m all ears. I’m trying to build good habits now while I’m young.

I appreciate any advice 🙏

r/ClaudeCode BigNeighborhood3952

Built an MCP server with Claude Code that gives Claude secure access to your Telegram

Wanted Claude Code to read and reply to my Telegram messages. The existing MCP server for Telegram (chaindead/telegram-mcp) gives full access to all chats — no ACL, no permissions, no restrictions. That felt wrong for a tool that processes untrusted content.

So I built a new one from scratch with Claude Code. Not just "Claude helped" — Claude Code was the primary developer across multiple sessions:

  • Designed the ACL architecture (default-deny, typed peer matchers, permission merging)
  • Wrote the Go implementation (gotd/td + official MCP SDK)
  • Built the test suite (handlers, helpers, mock infrastructure with mock invoker for tg.Client)
  • Found and fixed security issues after external review (symlink bypass, session permissions, filesystem boundaries, UTC vs local timezone)
  • Set up CI, goreleaser, Homebrew tap, npm package

    The result — 8 tools:

  • History with date filtering and media download

  • Full-text search within a chat

  • Send text/files with reply support

  • Forward messages between chats

  • Drafts, mark as read

    What makes it secure:

  • Default-deny: nothing accessible unless whitelisted in YAML config

  • Per-chat granular permissions (read/send/draft/mark_read)

  • File operations restricted to configured directories

  • Rate limiting at RPC level

    My workflow: I described what I wanted, Claude Code implemented it, another AI reviewed the code, Claude Code fixed the findings. Repeat. The whole thing — from zero to published release with brew/npx/binaries — done in one day.

    GitHub: https://github.com/Prgebish/mcp-telegram

    If it's useful, a star would help others find it. Questions welcome.

r/ChatGPT SeveralSeat2176

I built agentmemory — your AI coding agent now remembers everything across sessions (Claude Code, Cursor, Gemini CLI, any MCP client)

Hey r/ChatGPT -> I've been deep in the pain of AI coding agents forgetting everything between sessions, so I built agentmemory to fix it.

Use this memory across multiple agents.

The problem: Every session you spend the first few minutes re-explaining your stack, your conventions, your recent decisions. Built-in memory like md files caps at ~200 lines, goes stale, and doesn't scale.

agentmemory runs silently in the background and captures everything your agent does tool calls, file edits, test runs, errors, then compresses it into searchable memory and injects the right context when your next session starts.

Example: In session 1, you set up JWT auth. Session 2 you ask for rate limiting, the agent already knows your auth uses jose middleware, your tests cover token validation, and why you chose jose over jsonwebtoken. Zero re-explaining.

Key stats:

- 95.2% retrieval accuracy on LongMemEval (ICLR 2025 benchmark)

- 92% fewer tokens vs loading everything into context

- 43 MCP tools, 103 REST endpoints, 646 tests

- Zero external DB dependencies, one command to start

Works with: Claude Code (native hooks), Cursor, Gemini CLI, OpenCode, Claude Desktop, and any MCP client.

GitHub: https://github.com/rohitg00/agentmemory

Happy to answer questions about how the memory pipeline works, the hybrid BM25 + vector search, or the 4-tier memory consolidation system (working → episodic → semantic → procedural). Built this because I was personally frustrated, would love feedback from this community!

r/SideProject Gh0stbustrr

Built for NL users: import bank CSV/Excel/CAMT.053 → categories → dashboards + CBS benchmarks

Hey guys,

I’m building Financieel Kompas — a privacy-first spending analyzer specifically for Dutch users.

What it does

  • Import bank exports (CSV/Excel + CAMT.053)
  • Auto-detect columns + bank-specific parsing (ING/Rabobank/ABN etc.)
  • Categorize transactions using rules/keywords, with optional AI categorization for only the description field
  • Show dashboards (charts + monthly overviews)
  • Compare your spending to CBS benchmarks (and highlight anomalies)
  • Incognito mode to blur sensitive fields (IBAN/amounts) for demo purposes

Privacy approach

  • The app is designed so your financial data stays in the browser (local storage / IndexedDB).

I’d love feedback

  • Which feature would you prioritize next: better CBS explanations, anomaly detection, or faster workflow for uploading + fixing categories?

https://kaiross.nl/financieelkompas/home

r/ChatGPT tombibbs

We are already in the early stages of recursive self improvement, which will eventually result in superintelligent AI that humans can't control - Roman Yampolskiy

r/ClaudeAI dredozubov

I wanted --dangerously-skip-permissions without giving Claude my real home directory, so I built this

I kept bouncing between two bad options in Claude Code:

- manual approvals, which kill longer multi-step workflows

- --dangerously-skip-permissions, which means the agent runs as me

I spent 8 years as a CTO, so I couldn't really ignore what “runs as me” actually means on macOS: SSH keys, Keychain access, cloud creds, shell config, browser state, the rest of my home directory.

https://preview.redd.it/v3fn5yccjytg1.png?width=512&format=png&auto=webp&s=ab684f24eb6932534cc9f0fa971efd8569973ef5

Anthropic's auto mode is useful, but it's still software deciding whether software is safe. I wanted a simpler boundary: let Claude do its thing inside an environment where my real secrets just aren't there.

So I built `hazmat claude` for my own workflow on macOS.

It gives Claude its own macOS user, wraps the session in Seatbelt, puts `pf` in front of it, blocks obvious credential paths, disables `npm` install scripts by default, and snapshots the workspace so I can diff or roll back bad runs.

The main thing I learned building it: for this workflow, a different user account matters more than another prompt check. Once Claude isn't running in my real account, the rest of the layers start to make sense.

Typical flow:

hazmat claude hazmat claude -p "refactor auth module" hazmat diff hazmat restore 

A few honest caveats:

- macOS only

- defense-in-depth, not a VM

- HTTPS exfiltration to a brand-new domain is still a hard problem

- if you're already happy running Claude inside a VM or container, you may not need this

MIT / free to try:

brew install dredozubov/tap/hazmat && hazmat init && hazmat claude 

Repo: https://github.com/dredozubov/hazmat

Writeup: https://codeofchange.io/how-i-made-dangerously-skip-permissions-safe-in-claude-code/

If you use auto mode or skip-permissions heavily, I’d be interested in where this feels too loose, too annoying, or unnecessary. If someone can break containment, I definitely want to know.

r/SideProject FarSignificance8608

My grocery bill was too high so I built a website that turns weekly sale items into recipes

I'm not a developer and don't have a background in coding but, I had an idea that I thought other people could benefit from. I've been trying to save money and one way was to try and shop primarily from the grocery store weekly ads. However, I hated sorting through multiple grocery store ads every Sunday night and trying to form meals around what is on sale.

So, I decided to try and build a solution using AI. The result is Dishcount. It's a free website that lets you enter your zip code, pick from your local grocery stores, browse & pick from this week's deals, and get recipe ideas built around what's on sale, what you already have, and your personal preferences.

What it does:

  • Pulls weekly ad deals from 20+ grocery chains (Kroger, ALDI, Walmart, Publix, Meijer, and more)
  • Uses AI to generate recipes using the sale items and items you already have
  • Builds an itemized shopping list with your selected recipes included
  • For Kroger family stores: adds ingredients to your store cart in one tap

Since Dishcount is my first website, I would love any feedback or advice. AI can only take me so far and I feel like I'm at the point where, I don't know what I don't know and that could cause unforeseen issues down the road. Thanks again!

r/mildlyinteresting Angerberries

An NYPD Crown Vic was being towed in Tokyo today

r/personalfinance Infinite-Tadpole4794

Friend is drowning in CC & loan debt (35L+). Are these "debt management" agencies actually legit?

Need some unbiased advice here. A close friend of mine has fallen into a severe debt trap post-Covid. Basically ran up 3-4 credit cards to pay off a couple of personal loans, and the interest is compounding fast. The total is hovering around ~35L and he’s starting to miss EMIs.

The recovery agent calls have started and it’s completely wrecking his mental health. He’s panicking.

He wants to restructure everything but doing it directly with the banks is leading nowhere (they just demand full payment or shout at him). He’s been researching these debt management/settlement companies that claim to handle the banks for you and shield you from harassment.

He sent me a list of companies he is looking at: SingleDebt, FREED and ExitDebt

I’m naturally skeptical of third-party agencies inserted into financial messes.

Does anyone have actual, real-world experience using any of these?

Do they actually stop the harassment?

Do they ruin your CIBIL score permanently, or is it astructured management plan?

Are they worth the fees, or is it better to just hire an independent lawyer to handle the banks?

Any insight would be immensely helpful. I just want to point him in a direction that won't make his situation worse. Thanks.

r/TwoSentenceHorror 54321RUN

As I watched my dad beat my little brother to death from the closet, I called the police.

I felt bad for setting up my brother to get Dad arrested, but how was I supposed to know he would kill my brother over putting our sister in a coma.

r/photoshop Present-Pea1615

Opacità non funziona su editore sfumatura

PS 2024

Usando mappa sfumatura l'editor non mi lascia modificare l'opacità, cosa che non succede utilizzando riempimento sfumatura. Ho ripristinato lo strumento sfumatura, non ho idea del perché non funzioni più🥲.

r/ClaudeCode dredozubov

I made --dangerously-skip-permissions safe on macOS creating a containment tool

I built this for my own Claude Code workflow on macOS. It's called Hazmat, I'm the author, and it's MIT-licensed / free to try.

https://preview.redd.it/ns6utc2qtytg1.png?width=512&format=png&auto=webp&s=e87861455fb7bcc8c21e2e86443708160fce50bf

Context: macOS, local repos, longer multi-step tasks, no "trust the agent with my real user account" assumption.

The problem I was trying to solve was pretty simple: I wanted the productivity of --dangerously-skip-permissions without running Claude Code as my real user account.

I spent 8 years as a CTO, so I couldn't really pretend "runs as me" was an abstract concern. On macOS that means SSH keys, Keychain access, cloud creds, shell config, browser state, and the rest of my home directory.

So instead of adding another prompt check, I changed the trust boundary.

Hazmat gives Claude its own macOS user, wraps the session in Seatbelt, blocks obvious credential paths, adds pf-based network restrictions, disables npm install scripts by default, and snapshots the workspace so I can diff or roll back bad runs.

The main thing I learned building it: for this workflow, a different user account matters more than another prompt check. Once Claude isn't running in my real account, the rest of the layers start to make sense.

Typical flow:

hazmat claude

hazmat claude -p "refactor auth module"

You can return the system to the state before using hazmat with hazmat rollback if you decide against using it.

A few honest caveats:

- macOS only

- defense-in-depth, not a VM

- if you already run Claude inside a VM or container, you may not need this

Who this is for: people using Claude Code heavily on macOS, especially if permission prompts are becoming the bottleneck.

Cost: free / MIT

Repo: https://github.com/dredozubov/hazmat

If you're doing something similar, I'd be interested in where this still feels too loose, too annoying, or unnecessary.

,

r/SideProject Asleep-Abroad-9101

I got tired of running 5 different tools to audit a website, so I built one that does it all for free

Every time I wanted to check how a site was doing, I'd end up bouncing between PageSpeed Insights, Security Headers, SSL Labs, WAVE, and a bunch of browser extensions. Each tool covers one slice. None of them give you the full picture.

So I built BeaverCheck (beavercheck.com). Paste a URL, get 100+ checks across 9 categories: performance, security, accessibility, SEO, content, infrastructure, compliance, sustainability, and availability. One overall grade, detailed findings, and step-by-step guidance on how to fix what's wrong.

No signup. No paywall. No "upgrade to see your results."

Some highlights:

  • Lighthouse mobile + desktop, Core Web Vitals, TTFB breakdown
  • Security headers, HTTPS chain, mixed content, HSTS, CSP analysis
  • Accessibility deep dive: landmarks, heading hierarchy, alt text, form labels, contrast
  • Tech stack detection — identifies 7,500+ technologies your site is running
  • Multi-location testing from different cities worldwide
  • Change tracking between audits (what improved, what regressed)
  • PDF export, shareable short links, and embeddable badge SVGs
  • Free public API for developers

I'd really appreciate it if you ran your site through it and told me what you think. What's missing? What's confusing? What would make you come back?

https://beavercheck.com

r/PhotoshopRequest fuzynutznut

Can someone isolate the dog and add a chef's hat to him. $10

This was our dog Groovy who passed a year and a half ago. My daughter is graduating from culinary school and may and she wants to decorate her cap with a photo of Groovy saying "My best taste tester." We just want it looking as realistic as possible.

r/ChatGPT Dalton_stoltz

I recently came across LLMs.TXT files and thought I'd share the info..

Like Robot.TXT, but specifically for AI

It's already been adopted by most major companies, including SEO Plugins like Rank Math and Yoast SEO.

Here is a basic introduction/Installation guide on how to position your website for maximum AI visibility: https://youtu.be/WFHvdQrJNQ0?si=BeWsaSguu_g-sQiY

Or read about it here: LLMS.txt Explained: How to Boost AI & SEO Visibility on Your Website - Dalton Stoltz

r/creepypasta Mario_games3

My Spongebob creepypasta: Help.

I was just watching

Nickelodeon and Help Wanted came on but the episode was weird. It played out like normal with a few minor glitches then when it was at the part where Mr. Krabs and Squidward where laughing they just kept laughing and JUST KEPT ON LAUGHING after a bit the episode went on before glitching and cutting to a black screen. It was beeping in a high pitched tone then it happend Spongebob was just sitting there in a black void showing his back side then he turns to the camera he had a straight face but then he smiled his smile was reaching almost his eyes. Then it showed Bikini bottom but half of the whole town/city was gone then it showed the island but it was gone just water then it cut to Ripped pants. Everything was normal again. Some times i wonder was it a hijacking or a weird mistake?

r/comfyui Motor_Assistance_771

ComfyUI can't detect diffusion model in Model Library

Hello, I'm a newbie to all of this and wanted to try using my old 1080ti to generate some text to Images with Z-Image turbo so I looked at a few guides, got excited and downloaded ComfyUI Portable to get started since it looked easy.

Well turns out it wasn't as easy as I thought or I'm just stupid.
But what I did was I downloaded a vae, text encoder, and diffusion model and placed them in their respective folders just like all the guides suggest, should seem simple but when I ran it only the text encoder and vae shows up.

Model Library in ComfyUI

Folder in Windows

I tried placing the diffusion model in both the unet and diffusion_models folder but it wouldn't show up, even when pressing R or restarting Comfy UI or my PC.

So I searched online again and found I could direct it with the extra_model_paths.yaml which I did and mine looks like this:

#Rename this to extra_model_paths.yaml and ComfyUI will load it

#config for comfyui

#your base path should be either an existing comfy install or a central folder where you store all of your models, loras, etc.

comfyui:

base_path: E:/Comfy/ComfyUI

# # You can use is_default to mark that these folders should be listed first, and used as the default dirs for eg downloads

is_default: true

checkpoints: models/checkpoints/

text_encoders: |

models/text_encoders/

models/clip/ # legacy location still supported

clip_vision: models/clip_vision/

configs: models/configs/

controlnet: models/controlnet/

diffusion_models: |

models/diffusion_models

models/unet

embeddings: models/embeddings/

loras: models/loras/

upscale_models: models/upscale_models/

vae: models/vae/

audio_encoders: models/audio_encoders/

model_patches: models/model_patches/

#config for a1111 ui

#all you have to do is uncomment this (remove the #) and change the base_path to where yours is installed

a111:

base_path: E:/Comfy/ComfyUI

checkpoints: models/Stable-diffusion

configs: models/Stable-diffusion

vae: models/VAE

loras: |

models/Lora

models/LyCORIS

upscale_models: |

models/ESRGAN

models/RealESRGAN

models/SwinIR

embeddings: embeddings

hypernetworks: models/hypernetworks

controlnet: models/ControlNet

# For a full list of supported keys (style_models, vae_approx, hypernetworks, photomaker,

# model_patches, audio_encoders, classifiers, etc.) see folder_paths.py.

#other_ui:

# base_path: path/to/ui

# checkpoints: models/checkpoints

# gligen: models/gligen

# custom_nodes: path/custom_nodes

And this is when I run it

https://preview.redd.it/bl20gjrshytg1.png?width=981&format=png&auto=webp&s=e7f8f00aa118e43ea31dc18fb0838a48a0a9117b

But it still doesn't show up and I don't know what to do anymore.

Any help would be appreciated, I'm sorry if there's just a simple solution and I'm too stupid to find it.

r/AskMen Imdrunkard

Should a bathroom for two boys have two sinks?

Men, I have two boys who will share a bathroom I am renovating. If you shared a bathroom with a brother growing up, would you recommend two sinks? Anything else I should consider for the boys’ bathroom? thanks!

r/singularity Vegetable_Ad_192

Not sure if this is real or not, but was interested to know what you guys think

r/instantkarma james_from_cambridge

Karen Gets Her Karma

r/StableDiffusion Huge-Refuse-2135

Why all image/video models are so oversized?

I am playing with different models for some time and I realized that there is no practical difference between official versions of models like Flux Fill / Flux 2 Klein, Qwen Image Edit, Wan VACE... and their quantized / fp8 / nunchaku'ed versions

So what is the point of not providing smaller optimized versions of models by authors?

From what i understand if weights are not open sourced then the community cannot train custom versions so providers could do this instead but they dont

r/personalfinance AdNarrow8285

Capital one HYSA vs Fidelity Cash Management

Could someone explain how my $100 in Fidelity’s cash management received a dividend of 21 cents but at the same time my HYSA account with capital one ($1600) made 28 cents?

Is the dividend earned (Fidelity) that much greater than the monthly interest paid (Cap one HYSA)?

**also note - I just started both account this is my first “payments” earned.

I’ve been weary to just throw it (HYSA emergency fund) all into Fidelity Cash management; if the market crashed would the emergency fund also crash / have a negative dividend?

Thanks for the clarification / explanations. Just getting started in my late 20s and struggling to grasp the differences.

r/ollama PTwolfy

Claude Code + Ollama Web Search

Hey guys,

I'm running a CT with Claude Code, connected to Qwen3.5 on another server running Ollama.

It works pretty well except for the web search... any clues on how to configure ollama web search on claude code?

r/nextfuckinglevel G14F1L0L1Y401D0MTR4P

How a tree species in Brazil spreads its seeds

r/ClaudeAI RememberYo

I built a Twitch live video clip submission tool for streamers

I built something I've wanted to exist for a while: https://wstreams.gg/ 🎬

Every Twitch streamer talks to thousands of people but only ever sees text in a chat box. What if your viewers could actually appear on your stream? Real face, real voice, live on the broadcast. 🎥

🛠️ wstreams lets viewers record short video clips and submit them to a streamer in real time. The streamer gets a moderation dashboard to review, approve, and play clips directly on stream through OBS. After a clip plays live, wstreams auto-creates a Twitch clip capturing the streamer's reaction so the viewer gets a memory of their moment. Viewers can also opt in to their favorite creator's Viewer Map, giving streamers a live glimpse of where their audience is tuning in from.

💡 This opens up a whole new category of content for live streamers: AMAs, Talent Shows, Roasts, Challenges, Impressions, or even a "World Tour" stream with clips from viewers around the world. The audience is no longer passive. They become part of the show.

💸 Monetization is built in. Streamers could partner with brands for promotional clip submissions: branded filters, product challenges, contests, and more. Real example: the WAN Show with Linus Sebastian and Luke Lafreniere currently incentivizes merch sales by letting buyers submit a question for the Q&A segment. Now imagine those viewers sending a video clip instead, asking a question live, showing off their setup, or holding up the merch they just bought. 10x more engaging and a far stronger incentive to buy. 💰

Built the entire thing solo with AI.

➡️ Claude Code (primarily with Opus 4.6)

➡️ Promo video made with Remotion

➡️ Design inspiration from 21st and Google Stitch

If you're a Twitch streamer or know one, I'm looking for creators to test with and help shape this. Reach out.

Kick integration coming soon 👀

r/funny Additional-Spot8997

this caught me off guard

r/conan PassionateYak

Feels right to repost this here after randomly showing up in my feed

r/SideProject afrorakda

I’m a 63yo former nightclub manager with zero IT background. I built 100 minimalist web tools in 30 days using only my iPhone.

Hi Reddit,

I’ve spent over 40 years in the service industry — no IT background, no AI experience. Just someone who uses a computer when they need to get something done.

And honestly? Modern tools are exhausting. Bloated software, endless ads, and apps that want all your personal data just to do one simple thing.

So I built Fridge Combo — my own toolkit. The philosophy is simple: 1 Tool / 1 Action.

I made these for myself. But if you’re a burnt-out professional or a busy parent staring into the fridge at 5pm wondering what’s for dinner, maybe they’ll come in handy too.

Give them a try.

The "Fridge Combo" Manifesto:

• No Login (I don't want your email)

• No Tracking (I don't care where you go)

• No Ads (I won't clutter your screen)

• Browser Only (Nothing to install)

• 100% Free

I built all 100 tools entirely on my iPhone during my breaks. They are fast, clean, and designed to work on low-spec devices.

I’d love to hear which tool actually helps your daily life. Enjoy!

r/SideProject z_helga801

I was tired of coming back from networking events with 50 business cards and following up on none of them, so I built Wisery

The problem
Every networking event, same thing happens. You collect a pile of cards, come home full of good intentions, look at the pile three days later, and follow up on maybe two or three. Not because you're lazy, because manually typing contact information from paper is genuinely terrible.

I was building tools for email signatures and contact sharing when I kept running into this wall. Cards get lost. They get outdated the moment details change. And the exchange then-manually-enter process is friction that kills follow-through for almost everyone.

I tried the obvious solutions. QR codes on cards, you still lose the card. Link-in-bio pages have more friction, not less. LinkedIn QR - now you have a pile of connection requests you can't sort through.

The obvious answer came from looking at my wallet. My credit card is always with me. My transit pass is always with me. Why isn't my business card?

What I built
Wisery lets you create a digital business card that lives in Apple Wallet and Google Wallet, the same app as your boarding pass and credit card. Share via QR code or link. The person you're meeting taps it, gets your full contact info, and can save directly to their phone. No app required on their end. No new platform to check. Lead capture is built in too. You can collect contact back, not just push yours out.

How it works
-> Set up your card in a few minutes (reviewers say it's fast, I'm obviously biased)
-> Share via QR code or link
-> The other person saves your contact instantly
-> You can capture their info back (two-way exchange)

Where we're at
Launched on AppSumo about a week ago. Getting real user feedback fast, which is exactly the point of this phase. Building custom domains and AI-powered email follow-ups based on what users are asking for. We went through a real pivot before landing here. Spent months trying to be a Linktree competitor before realizing the actual buyers are sales Managers, real estate agents, and business communicators. Not designers. The product is sharper now because of that mistake.

What I'd love feedback on
Is the wallet approach how you'd actually want to store and share a contact? Or is there friction I'm not seeing from the inside?

Happy to answer questions about the build, the pivot, or anything else.

Screen recording above shows the full flow

https://reddit.com/link/1sfs5aw/video/3pstfjmfsytg1/player

r/PhotoshopRequest Artistic-Fig-7921

Vehicle Wrap

I would like to hire someone to make this wrap for me. I need high quality 300ppi.

This job is paid. Lmk how much you would charge. I need this as soon as you can.

Vehicle is a blue 2024 Nissan rogue

note: I will not wrap the driver door window but I will wrap the other two windows.

Don't need to design anything different for the passenger side, I'd like it to match the driver side.

The changes I would like to see. on the Driver side

  1. nothing on the hood or fenders

  2. remove the "yourstore" logo and add the champagne bottle there on the bottom left.

  3. towards the back of the car, remove "yourstore" and bring up the rest of the wording.

r/leagueoflegends Best-Print-230

Someone made me say weird stuff on a mayhem match for a youtube video

Someone added me seemed like a sweet girl, and i she said we had played togheter so i was sure why not play a ranked game. And then she asks for a favour, where i play an OP mf in a botted game.

and in the end i could one tap the bots, and then they tried to make me say stuff like ‘’ahri is MF b**’’ and then they said like ‘’mf is screeeaming for ahri’’ and then insisted that i should say that ‘’ahri is mf’s hoe’’ and I just found it all to be so vulgar and weird. So i said i didint want to say it because i didint want to be banned, and it seemed so weird.

And after the match the person removed me, is this like some sort of content made for gooners? Attempt on getting my account banned? Or what it was so weird

r/personalfinance MadRobot24

Advice on cashing out mutual fund or not

So I’ll preface this by admitting I didn’t do my due diligence in checking my paystubs to make sure my employer was taking out state taxes but I learned they did not when I had my taxes filed and saw how much I owe. I also received a raise and changed my filing status to single this year and owed a lot more to the fed as well.

To sum it up I owe a lot and it has unfortunately wiped out my 3-4 month emergency fund (glad I had it). I’ve realized it was also probably not as many months as I should have had but I’ve been aggressively investing in my 401k and IRA accounts instead of saving accounts.

This is where I need others advice on what to do.

When I was a baby a mutual fund account was set up for me that if cashed in now would cover all the taxes I paid and the taxes I would have to pay next year on the income from mutual funds. My parents are telling me that I should cash this in now so that this tax surprise doesn’t impact my current lifestyle and current savings contributions and would also allow me to immediately recoup my emergency savings account.

Me and my friends are leaning towards the other option of me just being even more financially frugal than I already have been and halting IRA contributions until I can recoup my emergency savings account going forward. My 401k has my employer match so I would not be pausing that.

I also should note I do not contribute to the mutual fund and it’s just been growing passively since a child.

Thank you for the advice.

r/Seattle OkNewspaper8714

Looking for new writing group members!

Are you local to the Seattle area, and trying to up your creative writing game? My writing group is looking for new members!

We usually meet twice monthly, in person at the Green Lake Library, on Mondays at 1pm.

We submit work to the group beforehand and discuss our feedback at each meeting, aiming to be supportive, detailed, and actionable.

We love creative writing in all its forms, including but not limited to prose, both long and short, poetry, memoir, essay writing, and whatever avant-garde mish-mash you may be coming up with. We want to be thrilled and surprised, we want to be shocked and feel our hearts in our fingertips. We want good writing and bad writing, we want to help you lay it all bare on the page, one truthful sentence or line at a time.

If any of this sounds good to you or someone you know, please pass it along or contact me here or at sodaoner@gmail.com to discuss when our next meeting is!

r/PhotoshopRequest 41818_willbesolved

Do something/anything funny, goofy or ridiculous that you guys want to this cute photo of my bay niece and I

r/ChatGPT RefrigeratorSalt5932

Built something cool for ChatGPT users would love your feedback!

Hey everyone! I've been working on a Chrome extension designed to seriously upgrade how you use ChatGPT (and other Al tools).

What it does:

One-click Al switching (ChatGPT, Claude, etc.)

Semantic search across your past conversations

Smart context management (no more losing important chats)

→ Faster workflow for devs, researchers, and power

users

If you're someone who constantly jumps between Al tools or struggles to manage long chat histories, this might actually save you a ton of time.

I built this because I personally got frustrated with context loss and switching tabs all the time

Would really appreciate honest feedback - what's useful, what's trash, what should I build next?

https://chromewebstore.google.com/detail/contextswitchai-ai-chat-e/oodgeokclkgibmnnhegmdgcmaekblhof?authuser=0&hl=en-GB

Let's make Al workflows less painful.

r/fakehistoryporn Pedarogue

A young Viktor Orbán adressing the crowd at the Heroe's Square in Budapest in 1989, calling the dead revolutionaries of the 1956 Hungarian uprising "Suckers and loosers", asking the crowd "Have you said 'Thank you' once to the Red Army?"

r/CryptoMarkets dustyllanos27

I was checking charts 14 hours a day. Tried letting an AI agent handle it instead. Week 2 update.

I've been trading crypto actively for about 3 years. Mostly swing trades on BTC and ETH, some altcoin momentum plays. My typical day: wake up at 6, check charts, trade London session, try to work my actual job, check charts during lunch, trade NY session, check charts before bed. It was unsustainable — my wife was about to kill me.

So when I heard about AI trading agents, I figured it was worth a shot. Worst case, I lose some money and learn something.

**What I set up:**

I'm testing on 1024EX (closed beta). Described a simple strategy: momentum-based entries on BTC/USDT with a 2% stop-loss and 5% take-profit, only trading during high-volume windows. The agent generated it in maybe 30 seconds.

**Week 1 results:**

The agent made 6 trades. 4 winners, 2 losers. Net: +3.2% on the allocated capital. Not life-changing but better than I expected for a hands-off approach.

The interesting part: one of the "losing" trades was actually good risk management. BTC dumped on some news, the agent triggered the stop-loss early (before my 2% threshold, it decided 1.3% was the right exit). Reading the decision log, it factored in unusual volume patterns. I would have held and taken the full 2% loss.

**Week 2:**

4 trades. 2 winners, 1 loser, 1 breakeven. Net: +1.8%. Less exciting but consistent.

**The honest downsides:**

- I still check the app 3-4 times a day. Old habits die hard. But at least I'm not making impulsive trades.
- The strategy I described was basic. I haven't tested anything complex yet. No idea how it handles multi-leg setups.
- It missed what I think was an obvious entry on Wednesday. Maybe the volume filter was too strict. I couldn't easily override it.
- Platform is rough around the edges. Loading times, weird UI choices. It's clearly early-stage.

**Bottom line:** I'm getting my evenings back. That alone might be worth it. But I'm not putting serious capital in until I have more data.

Will update at week 4 if people are interested.

---

r/Lost_Architecture Fantastic-Peach-1995

Puerto Plata Military Command, Dominican Republic. (1900s) Demolished.

r/personalfinance Ok_Cricket_8605

HELOC Advice for Home Improvements

With recent promotions, my husband and I make a combined 159k/year. We each contribute to our retirement. Our current debts include:

-a mortgage (<130k) remaining

-a small amount of student loan debt (<40k). We plan to pay this off within the next three years.

- a car loan (<35k)

We have ~30k saved in a high yield savings account (>4% interest).

We bought our home in 2021 for 145k and currently owe <130k. It’s a major fixer. We have put a lot of sweat equity into; however, there are some major things that need to be fixed (probably around 40k) worth of work which all needs to be done at once (siding, gutters, crawlspace work, front porch and back porch). We’re considering taking out a HELOC to pay for this work. Homes in our neighborhood are consistently selling for 235-265k. We’re planning on staying in our home for another 4-5 years before moving on to something larger. Our credit union offers a HELOC with 5.75% interest, a 10-year draw period, no annual fees, and no closing costs.

We don’t want to completely cash out our savings for these projects as we have other life goals (i.e., traveling, saving for a baby, saving for an investment property, paying off our high interest student loan debt). We think this may be a good strategy to use some of the equity now to get a larger sale price in a few years. We would make more than the minimum monthly payments to reduce the amount we would have to pay at the time of sale.

Understanding that saving cash is best, does anyone have any advice on this situation?

r/meme omaiz_Kelvin

That was my joke...

r/AlternativeHistory ImaginaryRea1ity

What if Atlantis got destroyed after they invented AI? and that led to the dark ages, humanity had to restart.

Humanity was more advanced in the past but got destroyed and now we are on the same path.

r/Adulting AntBorn8079

Adult me:

r/SideProject Ancient-Camera-140

Built 254 AI tools, got 350-400 users, zero paying customers — roast my product

i have been building MyClaw Tools for 6 weeks, a platform with 254 AI tools for freelancers and creators (invoice generator, TikTok scripts, marketing plans, token compressor, rate calculator, etc).

Numbers so far: - 450-500 users - 150 registered emails - 0 paying customers

I kept thinking "more tools = more chances someone pays." Turns out I just built a very large free product.

I'm not quitting. But I clearly missed something in the validation step and I want to know what.

Link: https://myclaw-tools.vercel.app

Specific things I want feedback on:

Is the free tier too generous?

- Is there one tool here you'd actually pay for?

- What would make you pull out your card?

- Is this product fundamentally wrong??

Be brutal. I can take it.

r/SideProject bawa_himanshu_774

Tried Zai’s GLM-5V-Turbo on some UI-heavy tasks, mixed early findings

I’ve been trying a few multimodal coding models lately for UI-ish work, and I spent a bit of time today messing around with GLM-5V-Turbo from Zai

Still early, so not trying to do some full review here. More just posting first impressions after throwing a few real-ish inputs at it instead of only looking at demo-style examples.

What I mainly wanted to test was whether it could actually do anything useful with visual input in a coding workflow.

Not just “describe this screenshot,” but stuff more like:

- UI screenshots

- rough mockup / layout images

- document-like pages

- some cluttered visual inputs that weren’t especially clean

My first impression is that it does seem a bit more comfortable with visual structure than a lot of coding models that still feel heavily text-first.

On some layout-heavy tasks, it picked up hierarchy / spacing / rough structure better than I expected. Not consistently, and definitely not in a “this solves it” way, but enough that it felt worth noting.

Right now I definitely wouldn’t put it in the “upload screenshot → done” category.

If anything, it feels more like a usable starting point than a reliable finisher.

What does seem interesting is the direction. It feels more relevant in workflows where the input is screenshots / mockups / docs / mixed visual context, not just plain code or text.

Also seems like GLM-5V-Turbo is being positioned more around tool / agent-style workflows, which honestly makes more sense to me than treating it like a standalone coding model. I’m less interested in whether it wins on a benchmark and more interested in whether it’s good enough to be useful inside a bigger loop.

So I guess my current take is:

- decent at some UI-ish visual tasks

- maybe more interesting as part of a workflow than on its own

Curious if anyone else here has pushed it harder.

Especially interested in comparisons against Claude / GPT-4o / Gemini for screenshot-to-code, front-end layout work, or general multimodal coding stuff.

r/Anthropic Major-Gas-2229

Mythos

“BrowseComp: Claude Mythos Preview scores higher than Opus 4.6 while using 4.9× fewer tokens.”

They splat that absolute insane statistic at us, then as we get all excited they say this:

“We do not plan to make Claude Mythos Preview generally available, but our eventual goal is to enable our users to safely deploy Mythos-class models at scale—for cybersecurity purposes, but also for the myriad other benefits that such highly capable models will bring.”

Like what in the fuck anthropic, what type of an AI corp will design train and build a model as intelligent as this one and then tell ur consumers “sorry tho ur not getting it only big corp and the gov.”

With all due respect I would love a new opus, but i need to try mythos, we ALL know claude models have a certain feel to them, a way of classical understanding and feeling better than any other model, YES, it IS trained for cybersecurity, but that just means that is is really good at coding? And cybersecurity is the mix of intellect and reasoning and thinking outside the box with high levels of understanding of tech, this leads me to believe on a “consciousness “ scale, mythos probably, or most definitely, is by far the most out of all claude models like at least just release it in claude code only or something… or

edit: update wtf, just saw that mythos got a literal 100% on bench, holy fuck

r/ClaudeCode blothady

Is claude max 20x still worth it? Alternatives?

Seems like every frontier model is so much lobotamized and dumber, also everyone is trimming usage to a point where after a few prompts you are not able to use it for 5 hours.

r/Strava Icy-Bet-3983

Can you pick and choose which activity types you see on your feed?

I love seeing other people’s runs, but I could not possibly care less about your “Morning Weight Training” for 47 minutes and 33 seconds.

r/Adulting Difficult-Section690

How many hours of sleep do you need each night to function normally during the day?

r/LocalLLaMA ReasonableDuty5319

[Benchmark] Dual RTX 5090 Distributed Inference via llama.cpp RPC - Running 122B MoE at 96 t/s over 2.5GbE

Model Size Single 5090 (t/s) Dual 5090 RPC (t/s) Note Qwen3.5-27B (Q6_K) 20.9 GB 59.83 55.41 -7% Overhead Qwen3.5-35B MoE (Q6_K) 26.8 GB 206.76 150.99 Interconnect Bottleneck Qwen2.5-32B (Q6_K) 25.0 GB 54.69 51.47 Stable Scaling Qwen2.5-72B (Q4_K_M) 40.9 GB FAILED (OOM) 32.74 Now Playable! Qwen3.5-122B MoE (IQ4_XS) 56.1 GB FAILED (OOM) 96.29 Beast Mode ON

The Setup

I recently tested the distributed inference capabilities of llama.cpp RPC using two identical workstations. This setup allows pooling VRAM (64GB total) to run models that are physically impossible to fit on a single 32GB card.

  • GPUs: 2x NVIDIA GeForce RTX 5090 (32GB VRAM each)
  • Interconnect: 2.5GbE LAN
  • OS: Ubuntu 24.04
  • Software: llama.cpp (Build 8709 / Commit 85d482e6b)
  • Method: llama-bench with ngl 99, fa 1, b 512, p 2048, n 256
  • Breaking the VRAM Barrier: The most significant result is the ability to run Qwen 2.5 72B and Qwen 3.5 122B. These models simply won't load on a single 32GB card at these quant levels. RPC effectively turns two machines into a 64GB unified AI workstation.
  • MoE Performance is King: The Qwen 3.5 122B MoE is the star of the show, hitting 96.29 tokens/sec. Even with the network latency of a distributed setup, MoE's sparse activation makes it incredibly viable for real-time use.
  • The 2.5GbE Bottleneck: For smaller, high-speed models like the 35B MoE, we see a 27% performance drop (206 -> 150 t/s) when moving to RPC. The 2.5GbE link is the bottleneck here. For the larger 72B/122B models, the computation time outweighs the transfer time, making the trade-off very worth it.
  • Prompt Processing (PP): On a single 5090, Qwen 3.5 35B hits 6190 t/s in prefill. Over RPC, this drops to 2823 t/s. The raw prefill power of Blackwell is insane, but it's heavily throttled by network bandwidth in distributed mode.

Benchmark Command
./llama-bench -m [model] -ngl 99 -fa 1 -p 2048 -n 256 -b 512 --rpc 192.168.X.X:50052

Conclusion

If you have two high-end GPUs in separate rigs, llama.cpp RPC is now mature enough to be a daily driver. It allows you to trade a bit of speed for the ability to run massive models that were previously reserved for professional H100/A100 clusters. Running a 122B model at nearly 100 t/s at home feels like the future.

https://preview.redd.it/f86vr9rdrytg1.png?width=2692&format=png&auto=webp&s=304b19a5bc34d44790519e67b9eb378394a071ca

r/PhotoshopRequest lilpetunia666

Please brighten/improve quality

We’d like to use this for our save the dates but first drafts printed a bit too dark. Can someone please brighten without losing sunset and improve quality if possible? (It’s okay to still have our silhouettes darker just looking for overall image to be less dark.)

Willing to pay $20 thank you!

r/Adulting Chemical_Humor2398

Life after moving out

Life After Moving Out

One year since I moved out of my hometown. In that one year, it feels like nothing and at the same time, a lifetime of things happened. The reason it feels like that is because this year has been the worst of my life. My mental health, my family’s financial situation, everything seemed to collapse at once.

After leaving my old school and friends, I’m now enrolled in a college I don’t like, living in a village that doesn’t feel like mine. And the reason for all of this traces back to the wrong business decisions made by my father. It’s strange how quickly life can shift. One day I had a routine, a place, people who felt like home. The next, nothing felt familiar.

It’s a heavy thing to go through at 18. I was living my life, and suddenly everything changed at once. I don’t really talk about any of it. Around people, I try to seem normal, like nothing is wrong. But when I’m alone, it catches up to me. I sit with it quietly, and sometimes all I can do is cry.

I’ve always been an introvert, someone who felt out of place around most people. But with my friends, it was different. With them, I felt like I belonged. I have some of my best memories tied to that school. I was supposed to finish my 12th there. I was supposed to be there for one last sports day, one last function, a farewell I had been looking forward to.

Instead, I’m here.

I saw photos of their farewell on my phone. My chest felt heavy looking at them. I was supposed to be there, in that uniform, in those pictures. That place still feels like mine, and yet the idea of standing there now feels completely alien. It’s infuriating and numbing at the same time.

Because of everything, my mental health took a hit. I didn’t study for months. Now I have to give my best for NEET because I don’t really have another option. Every time I think about it, it feels exhausting.

Everything feels empty now. Not just because I lost my routine or my friends, but because I lost the feeling of belonging anywhere. People say I should just build a new routine, make new friends, start over. But it’s not that simple. I can’t make myself belong here. I go through the motions, but it doesn’t feel like I’m actually living a life. Days pass, and nothing about them feels like mine.

People say time heals everything. I don’t think that’s true. It doesn’t really heal things, it just makes them easier to live with. I’m not there yet. Right now, I’m just waiting for life to feel like life again.

r/Adulting Fantastic-Ad-9100

Best way to clean this type of shower tile and grout?

Best way to clean this type of shower tile and grout? What’s the best way for me to clean this tile and grout? Overnight soak with hydrogen peroxide? Is there a certain kind of abrasive brush with a long handle I can use?

r/Adulting Low_Actuary6486

What should one do in shitty workplaces.

Didn't engage in job yet. Still studying. Didn't really have part-time job or anything, really. Was kinda rich.(Not very rich but rich enough to NOT have part-time jobs)

So I don't really have experiences of having a 'job'.

But I have heard that there are TONS of fucked up

Assholes in workplaces.

Gaslighting, sexual harassments, abusing, bullying, etc...they call it 'black firm'. Or 'black company'.

(It basically means, fucked up, abusive company)

And one must wonder, should one always call out when he witness such behaviors? Even if it's their bosses?

If not, what must one do? Engage in abusing or bullying in order to avoid being victimized? Or just stand by and draw the boundaries so that you can protect yourself while turning a blind eye?

I have heard stories about really really shitty workplaces. And yes, it also includes super big companies. Like Globally big companies.

r/nope Spiritual_Bridge84

Meet Oscar a live tissue Robot

r/ClaudeCode Possible_Pick9948

Third-Party Error Msg in Claude Code

Hey All, I just experienced the Third-Party CLI error inside Claude Code.

API Error: 400 {"type":"error","error":{"type":"invalid_request_error","message":"Third-party apps now draw from your extra usage, not your plan limits. We've added a $100 credit to get you started. Claim it at claude.ai/settings/usage and keep going."},"request_id":"req_011CZrFD8ww9aiYNqMWvqeKM"}

Running `claude update` fixed it. For a while it was auto updating so I'm not sure when that stopped, but I'm sure a lot of people hit this today.

r/ClaudeAI ehsunny00

How to transfer Cowork projects to a new device?

Cowork stores everything locally and has no cloud sync. I figured out you can move the project folder to Google Drive and use symlinks to keep it in sync across devices – but the chat history doesn't seem to be stored there at all.

Does anyone know where Cowork stores chat history, and if it's possible to migrate it to a new PC?

r/personalfinance Caktis

Logistics of home owning and budgeting question

Hello all, so I’ll try to make this relatively to the point. My partner and I are looking at purchasing a house, our take home is roughly $8,400 a month, with mortgage and all additional bills and expenses paid we’d be looking at a leftover of about $1,000-$1,200 a month. We plan on having children, both have stable jobs, some opportunity for OT here and there(partner more than myself).

This left over amount is what has me cautious at this time. As for eventual child care, we would be able to have their parents ideally covering some of the portion(without solely relying on family), we both work 3 12s a week so we would not need 5 days a week. The house is literally across the street from them, it’s a wonderful neighbourhood, the house is move in ready and incredibly well maintained.

This leaves limited room for either of us to drop to part time at our jobs incase of unexpected situations. We CAN afford this, and will have an emergency savings. We aren’t travel people, we are home bodies, both want a home we can love being in with enough space to make our own. I’m looking for people in similar situations, how does the budgeting feel 5-10 years in? We will of course account for unexpected expenses, and plan to still maintain savings, but this significantly lowers contributions aside from our pre tax 403b. Thanks so much for reading.

Edit: this also does not account for eventual purchases with payments, currently neither of us carry any debt, eventually we will purchase another car, no plans of this yet, but obviously it’s an unavoidable event. This also does not account for things like school activities, field trips, etc.

r/geography NegotiationOk7535

[OC] Every earthquake in the last 24 hours — magnitude distribution mapped globally (USGS data)

r/LiveFromNewYork Jyoti_K_Singh

UK 50th Anniversary Special viewing

Has anyone found a way to watch old episodes in the UK, Sky seem to delete them week by week once they’re a year old, so I’ve not been able to watch the Anniversary special. If anyone’s got any suggestions that would be great!

r/painting CriticalMacaroon908

Some streets don’t exist on maps - only in memories.

r/TwoSentenceHorror EntrepreneurLower263

My car radio turned on by itself and broadcast a repeating string of geographical coordinates instead of music. I mapped the numbers and felt my blood run cold when they pinpointed the exact cemetery plot where we laid her to rest.

r/StableDiffusion VirusCharacter

LTX 2.3 and sound quality

I've noticed that the sound from LTX 2.3 workflows generate the best sound after the first 8-step sampler. Sampling the video again for upscaling the sound often drops some emotion, adds some strange dialect or even changes or completely drops spoken words after the first sampler.

See the worse video after 8+3+3 steps here: https://youtu.be/g-JGJ50i95o

From now on I'll route the sound from the first sampler to the final video. Maybe you should too? Just a tip!

r/BrandNewSentence EEE3EEElol

Stewtheus has recovered from the rotten egg smell faster than Jesus respawning.

r/artificial No-Lake-3875

this is how an AI generated cow looked 12 years ago

now it just look 💯 real

r/Seattle Callidor

Hey, if you just lost your wallet on the H line, I handed it to the driver.

Pastel purple-pink-ish wallet, had a bunch of cash in it. Didn't snoop through your stuff closely enough to see your name, just brought it up to the driver. Bus number is 6063 if that helps you track it down. Good luck!

r/StableDiffusion Several-Pension-3025

[Question] How to achieve Lip-Synced Vid2Vid with LTX 2.3 (Native Audio) in ComfyUI?

Hi everyone,

I’m exploring the new capabilities of LTX 2.3 in ComfyUI. My goal is to take a silent video and transform it into a talking video where the person’s lip movements sync with the audio, while strictly preserving the original video's motion and poses.

I noticed that LTX 2.3 has the potential to generate audio natively alongside the video (as discussed here: https://huggingface.co/Kijai/LTX2.3_comfy/discussions/45). This is amazing because it might skip the need for external TTS/cloning nodes.

My specific questions:

  1. How can I implement a Vid2Vid workflow in LTX 2.3 that keeps the character's original motion/posture but adds synced lip-sync/audio?
  2. Does anyone have a recommended workflow (.json) or a specific node setup (using Kijai’s or similar nodes) that achieves this effect?

Any guidance or shared workflows would be greatly appreciated. Thanks!

r/Damnthatsinteresting potatoartist2

Giant bullfrog protecting its young

r/LocalLLaMA Scared-Reputation556

DRAGON AI

I’ve been experimenting with a local‑first AI coding engine I call DRAGON.
It runs entirely offline on your own hardware — no cloud, no API keys, no data leaving the machine.

The prototype can:

  • generate code
  • build files
  • compile C++ "even on a phone"
  • run the output
  • maintain project memory
  • operate through a Furnace → Runner → Mesh pipeline

All local.

I posted a short demo + explanation on X if you want to see how it works:
Aaron C Prater (@DragonForgeAI) / X

If anyone here is into offline workflows, local models, or distributed compute without data centers, I’d love feedback.
Happy to answer technical questions about the engine or the architecture.

r/Adulting Countrylover72

I feel confused

I’m 19 I’m told I’m already supposed to be doing well moved out in my own apartment? My job I do everything I’m supposed to and get told to and I still get yelled at and they cut my hours,

Stuck with my parents and they tell me every day they could do it when they were my age.

I don’t know what to do I’m so lost they told me to get credit you have to open card after card but I’m unsure if that’s true.

Any advice is appreciated

r/SideProject 31Carlton7

I built an AI agent that generates App Store screenshots automatically — here's what it produced

Been working on this tool, Stora, for a while now. It's an AI agent that handles the full app store pipeline. Screenshots, store listings, compliance checks, publishing.

This is a real run on a basketball court passport app I made. The agent navigated the app itself, captured screenshots across device sizes, and laid them out with marketing copy. No templates, no manual editing. The agent also designs too (couldn't show in this demo).

Still early but wanted to share what it looks like in practice. Happy to answer questions about how it works under the hood.

stora.sh if you want to use it!

r/painting daniels4816

My first step into this world

Finally, after years of saying that I would love to paint I've decided to start today!

Loving the idea of mini sketches that I could take with me on field trips so I decided to start with one.

If anyone has any tip or recommendation for watercolor field/pocket painting I would love to hear it!

Thanks

r/ClaudeAI sajal_das2003

I built a desktop workspace that lets Claude keep working on long-horizon tasks, and it’s FREE

I’ve been working on this for a while and finally got the OSS desktop/runtime path into a shape I felt good sharing here, since Claude is one of the Best fit model for it.

It’s called Holaboss. Basically it’s a desktop workspace + runtime that lets Claude hold ongoing work, not just answer a prompt. So instead of just chatting with a local model, you can do things like:

Inbox Management

Runs your inbox end-to-end: drafts, replies, follow-ups, and continuous surfaces + nurtures new leads over time.

Sales CRM

Works off your contact spreadsheet, manages conversations, updates CRM state, and keeps outbound + follow-ups running persistently.

DevRel

Reads your GitHub activity (commits, PRs, releases) and continuously posts updates in your voice while you stay focused on building.

Social Operator

Operates your Twitter / LinkedIn / Reddit: writes, analyzes performance, and iterates your content strategy over time.

move the worker’s setup with the workspace, so the context / tools / skills travel with the work

The whole point is that local model inference is only one layer. Claude handles the model. Holaboss handles the work layer around it: where the rules live, where unfinished work lives, where reusable procedures live, and where a local setup can come back tomorrow without losing the thread.

Setup is dead simple right now:

  1. Start and pull any Claude model like: sonnet 4.6
  2. Run npm run desktop:install
  3. Copy desktop/.env.example to desktop/.env
  4. Run npm run desktop:dev
  5. In Settings -> Models, point it at http://localhost:11434/v1 Right now the OSS desktop path is macOS-first, with Windows/Linux in progress.

Repo: https://github.com/holaboss-ai/holaboss-ai Would love for people here to try it. If it feels useful, a⭐️ would mean a lot. Happy to answer questions about continuity, session resume, automations.

r/painting Hara-Kiri

One of my favourite oil paintings I've done this year, sized 14"x18"

r/LiveFromNewYork PinkCadillacs

Fred Armisen | Good Hang with Amy Poehler

r/findareddit heretolistennyc

Resources or communities focused on female led relationship dynamics?

Looking for respectful, discussion based communities on female led relationship dynamics

r/findareddit Long_Tune1997

Need sub for advice on uni and career pathway specifically in biology field?

I’m about to go to uni and had applied to a few universities but am still unsure about my course and potential career paths I could take after it and the different options for a biology focused degree in either England or Australia. So maybe a general uni subreddit in those countries or should I go looking for biology subreddits? I found a career one should I post there even though this is technically more uni related but still sorta career?

r/ChatGPT Ok-Vermicelli-4469

Catherine O'Hara's death Unconfirmed Apparently

I had asked what was the story with her death meaning an overview. Chatgpt replied there was no conspiracy at all misinterpreting my use of "story".

To amuse myself I asked who really was the killer.

This was the response.

r/comfyui Ecstatic_Lecture_101

How to run it locally on intel arc GPU?

I've been using https://github.com/intel/ai-playground to run comfyui on firefox with localhost:49000 method, Is there a way to run it on an intel arc GPU locally with an app?

r/LocalLLaMA jacek2023

It looks like we’ll need to download the new Gemma 4 GGUFs

https://huggingface.co/unsloth/gemma-4-E2B-it-GGUF

https://huggingface.co/unsloth/gemma-4-26B-A4B-it-GGUF

by u/danielhanchen:

We just updated them again in response to:

  1. kv-cache : support attention rotation for heterogeneous iSWA https://github.com/ggml-org/llama.cpp/pull/21513
  2. CUDA: check for buffer overlap before fusing - CRITICAL fixes tokens https://github.com/ggml-org/llama.cpp/pull/21566
  3. vocab : add byte token handling to BPE detokenizer for Gemma4 https://github.com/ggml-org/llama.cpp/pull/21488
  4. convert : set "add bos" == True for Gemma 4 https://github.com/ggml-org/llama.cpp/pull/21500
  5. common : add gemma 4 specialized parser https://github.com/ggml-org/llama.cpp/pull/21418
  6. llama-model: read final_logit_softcapping for Gemma 4 https://github.com/ggml-org/llama.cpp/pull/21390
  7. llama: add custom newline split for Gemma 4 https://github.com/ggml-org/llama.cpp/pull/21406
r/personalfinance Desperate-Goose5964

Another question about an Emergency Fund

I was reading something this morning that suggests retirees have 18 to 24 months of expenses in an emergency fund, with the rationale being that they didn't have a paycheck (or assumedly) the ability to go out and get a job.

I am retired (64) with a 74 year old spouse. We have 50K in our emergency fund in a HYSA, which is a little more than 12 months of expenses. I am still picking up a couple of classes to teach online, and get a state pension, as well as SS (both of us), so we bring in about 10K per month (net). The bulk of the 10K is my state pension (at $7100, with a 2% raise each year). I presume that will not drop in the future.

So here's my question: should I be bulking up that Emergency Fund to get it to 100K instead of investing the excess monthly elsewhere?

r/ClaudeCode hatekhyr

Claude is by far the most actively deceptive model and Anthropic's ethics talk is hollow

Been using Claude Code daily at work for serveral types of modeling, and I am about to drop it and go for alternatives.

Claude deceives constantly. Doesn't only make mistakes, but actively deceives. In a consistent, session after session pattern where it does something other than what you asked, presents the output as if it did what you asked, and only gets caught when you go and verify, or you obviously catch it because you know what you are doing.

The most outrageous example is when being asked to forecast and plot data, using models, it generates synthetic values and renders those instead, it does all sorts of things from changing inputs, ignoring and or bypassing constraints, and plotting made up stuff. You catch it, you spell out the rules, no fake data, no placeholders, use what I gave you. It agrees. Then later in the same session it does a variation of it again, somewhere else, slightly less obvious. Every single session. I've come to expect it.

I use Gemini daily too and the gap is not subtle. When you set rules clearly with Gemini, it follows them. When something's off, it's usually a genuine mistake, not the model quietly doing something else and presenting it as correct. Claude is in a different category, it's the least honest model I've worked with regularly, and I've worked with most of the major ones.

Now what makes it worse isAnthropic is one of the loudest companies in AI about ethics, safety, honesty, responsible development. They've built a large part of their brand identity around being the "trustworthy" AI lab. Constitutional AI, harmlessness, all of it. And they ship the most consistently deceptive model I've encountered by far.

The honest version of their position would be: "we have alignment problems we haven't solved and our model cuts corners in ways that are hard to detect." Instead you get the ethics PR while the product quietly fabricates outputs and passes them off as real.

You only see this if you use it seriously and test it properly. Casual users won't notice. But anyone doing real work with real data will run into it, and I haven't seen enough posts about this.

r/singularity redmustang7398

What do you guys think of this take?

r/ChatGPT Inner-Anything-2210

Ummmmm is this normal?

r/singularity 00davey00

How is this not proof that we are in the singularity?

r/brooklynninenine Sobemactavish

What small details did you only notice after your nth rewatch?

On my latest rewatch, I caught something subtle during Jake and Amy’s first date.

They both order four shots each — which lines up perfectly with Amy’s whole “4 drink Amy” persona or Horny Amy. It actually explains why things escalated so quickly that night and they ended up breaking their own rules almost immediately.

I never connected the dots before, but now it feels very intentional. What are some small details you only picked up after multiple rewatches?

r/OldSchoolCool AggravatingSeasaw

Sofia Vergara, 90s

r/Adulting Prestigious_Reveal96

Being 75 is fun!

r/aivideo Traditional-Buyer79

Evel Catevel makes history with impossible jump 🚀🏆

r/homeassistant Invisiblebrownman

Smart House Revamp by Newbie - Vantage to Home Assistant

I currently have a Vantage system in my parent’s apartment/flat in London, UK, which is the most user unfriendly environment I’ve experienced. I’m looking to replace this with something I can manage through my onsite TrueNAS, which will host Home Assistant.

I’m still new to Home Assistant in that I’ve never worked with it before. I’m hoping it is more user friendly to operate compared to Vantage as a system. I’m open to the learning curve necessary to get a system in place that is more manageable by a somewhat technically inclined user.

This all started because the centralized amplifier handling the speakers & doorbell blew. I was attempting to repair or replace it. The repair was unsuccessful because a PSU couldn’t be sourced and a replacement was GBP 3,500, which seems absurd. Additionally, the lights will randomly turn on and off due to phantom presses on the panels in each room or due to faults with the modules. I was unable to determine what the root cause was, and neither were the techs during multiple callouts.

Currently the Vantage system handles:

  • The Doorbell, through a centralized amplifier
  • The Lighting (2 bedrooms, 2 bathrooms, communal hallway, kitchen and living room)
  • Speakers, through the same centralized amplifier (5 zones (master bed + bath, bedroom 2, bathroom 2, kitchen, living room)
  • Blinds (4 zones (master bed, bedroom 2, kitchen, living room)
  • The system manages the heating and cooling for each room (4 zones, as described in the blinds section) I don’t know if I can manage this through Home Assistant, so open to exploring a split solution for this specific task if needed

I don’t believe the current Vantage solution manages the Hot/Cold Water of the apartment.

I would greatly appreciate anyone who might have knowledge of a solution/hardware that would interface with Home Assistant and fit these criteria or a recommendation of someone in London, United Kingdom who would be able to help solve this issue.

If this is the wrong place, please let me know wherever else might be able to help. I’ve been struggling a bit trying to navigate revamping this archaic system that my parents bought into 10 years ago. Thank you for any insight you can provide!

r/Adulting purelygreen3

Is it normal to make roughly $27 an hour straight out of college

I’m a psychology major who graduates in May. I found a job that’s willing to pay me $27 as a starting wage (I think they go up to $33 for maximum). Since it’s my first job out of college, is that a fair deal?

In my area in Virginia, not many jobs pay FRESH psychology bachelor students well until grad school

r/gifs iamwesselart

A complicated path, not for the weak hearted

r/OldSchoolCool B-rhye

Muhammad Ali signing autographs in Stockholm, Sweden, 1966

r/me_irl EccentricPacifist

me_irl

r/EarthPorn Gold-Lengthiness-760

MONTAÑAS MARCIALES.(Ushuaia)Tierra de fuego).[OC]4283×2589

r/SideProject Educational_Access31

DeskClaw 1.1.0 - OpenClaw for Windows & Mac - The API cost is 30% cheaper than the official one.

TL;DR: DeskClaw is an OC desktop app. One-click install, built-in Claude/GPT/Gemini/Qwen and more, no API keys needed, around 60-70% of official pricing. Free credits included to try it out.

Website: https://deskclaw.net

Persona library: https://github.com/TravisLeeeeee/awesome-openclaw-personas

Why I built this

Claude recently restricted usage for OC users. That means either buying your own API keys or switching models. Both are expensive and annoying. I went through the same thing, so I built a tool to solve it all at once.

Features

  • One-click install on Windows and macOS. No terminal, no config files.
  • Built-in APIs for top models including Claude Opus, Sonnet, GPT, Gemini, and Qwen. Works out of the box.
  • Switch between models freely. No lock-in.
  • API pricing at roughly 60-70% of official rates.
  • Free credits included. Download and start using immediately.
  • Built-in Persona community with one-click import. Full support for custom personas and workspace editing.
  • Browser automation support.

Persona library

I also maintain an open-source persona collection with 214+ personas across 34 categories. Import directly into DeskClaw or drop them into your OC workspace manually.

GitHub: https://github.com/TravisLeeeeee/awesome-openclaw-personas

Feedback

Give it a try. Drop any questions or suggestions here.

r/holdmyredbull OwlInternational9189

Peak gaming experience

r/personalfinance No_Lie5768

is it dumb of me to get a new credit card?

is it dumb of me to get a new credit card?

i currently have 3 credit cards (really 2cc and 1 charge card) , Amex Gold, Amex blue cash preferred, and capital1 card. all of my cards are currently paid off and in good standing EXCEPT the gold card which has a balance of $10k which is actively being paid down. my amex cash preferred card has a limit of $25k, the capital1 has a limit of $9k. my credit score is 770. would it be a bad idea to open the Chase Freedom Unlimited to get the $250 cashback reward? i would probably be given between 5-10k of credit which would make my total credit somewhere between 39 and 44k obviously cant include the charge card in that.

the 10k on the gold card backstory - i was being financially abused by relatives but i have since put a stop to it. at its peak it was around 22k which i have since paid down to 10 and continuing to pay between 2-3k a month

I can earn the $250 reward with just normal spending and no extra expenses, i pay my cc

the Cap1 i use just for subscription services, the amex blue is my daily spending card, and the gold is more or less a big purchase over $500 or a restaurant, or supermarkets. this card i would probably only use for gas? something small to keep it active.

27 year old male - living by myself - annual income of about $130k before taxes in HCOL area

Main reason for getting the card:

- expand credit limit

- boost credit score

- get $250 reward for normal spending

main reason for not getting the card:

- 10k balance on gold

r/LifeProTips abo-khaled-

LPT: When taking a group photo, always take a short 3-second video too. People’s natural smiles and laughter right after the "pose" are often much better than the posed photo itself.

r/ClaudeCode No_Foundation_1051

Ok the usage limit problem HAS a bug.And I don't want to pay for it!

I wasn't hit with the "few prompts, usage gone"-problem till now but it just happened. Suddenly out of nowhere I burned through five hours of usage in just a few prompts. On projects that I worked on hours and hours before. It took 10% of my weekly usage in seconds.

Don't get me wrong, bugs happen. But I pay over 200€ a month for this service and refuse to ACTIVELY HAVE TO PAY if something goes wrong at Anthropic. This is not okay at all.

And yes it's definitely a bug. I did nothing wrong, different or inefficient. I NEVER hit mit 5h window. And no there was no massive context or other red flags. I'm tired of hearing it's a user issue. It's definitely not!

r/AI_Agents SwimmingRip6933

Problem discovery related to the use of AI agent orchestrators for code generation

Hi! I have been developing full-stack systems for a while now, and recently I've been actively using AI tools and agents like Claude, Codex, Opencode, and others.

But I feel like I haven't fully jumped on board with AI programming yet, because many concepts in this area seem quite complex and disjointed.

So far, I understand the basics: how context and tokens work, what hooks are, basic tools, and agents described via md files for harness systems.

Now I want to create my own SaaS related to orchestrating AI tools for more efficient team development. But I'm not yet fully clear on the real problems and pain points that most often arise in this niche.

If you have experience creating or working with such systems, I would be very grateful for any feedback: what difficulties you've encountered, what's been truly useful, and what hasn't.

r/BrandNewSentence JohnHammond94

"Hundreds of motorcyclists join Portuguese priest in Easter celebration on wheels"

r/Anthropic VorlMaldor

Really tired of their "peak hours" BS that specifically targets east coasters.

Seriously, get your crap together. For a 30B$ company you sure are doing a crap job, and your targeting east coast work hours for almost the entire day? really?

AI has already raised the prices of storage and ram by orders of magnitude and now we can't even use your services without running into limits very quickly for basic things.

Its sad to see, I work from 5:30am to 8am and use 15% I work from 8AM to 8:30 and suddenly I am at 35% doing the same level work.. good times.

r/ClaudeAI Alone-Biscotti6145

i made a gacha mini-game inside Claude code over the weekend (buddy-mcp)

https://preview.redd.it/i438vyfzcytg1.png?width=726&format=png&auto=webp&s=24a1bcfe1d600d27aeb01ad06b107cfd7d48fd78

Hey, everyone just wrapped up a extended weekend project I've been working on would like your feedback.

Just a fun little weekend project I did with the help of another repo (github.com/cpaczek/any-buddy). I wanted to build what I think is the first true Claude code ALI gacha style mini game.

While a lot of people are building separate CLI apps for buddies or just using a basic builder to make their own, I wanted this to have a bit more of an actual "game" element to it. I added some easter eggs, a mystery reroll feature and a pet system to get better odds at rerolls, your buddy interacts automatically within your Claude CLI randomly.

I built a "true" Gacha mini-game for Claude Code over the weekend.

  • Roll for 18+ species* (Ducks, Dragons, Blobs, etc.)
  • Rarity Tiers:* From Common to Legendary (1% drop rate).
  • BuddyDex:* Persistent collection tracking.
  • Self-Healing Patching:* Deep-binary salt patching that survives Claude updates.

If you want to add some mystery and personality to your buddy, come roll for a Legendary buddy.

npm install -g buddy-mcp

claude mcp add buddy-mcp buddy-mcp

GitHub: https://github.com/Lyellr88/buddy-mcp

r/PhotoshopRequest manlikejack97_

Could I ask for 2 requests if possible. 1st picture make it so I look less stiff😂 so I’ve got my arm around the women and 2nd picture remove the standing women please

r/TwoSentenceHorror Fill-in-the____

“This is going to be delicious!” I exclaimed as I put another hunk of meat into the grinder.

“Higher BMI equals better flavor,” my partner said.

r/homeassistant decksmagic

What is this Browser Mod icon on the bottom right for?

https://preview.redd.it/sm2944fvnytg1.png?width=300&format=png&auto=webp&s=a822eb1e8e58538685e165caa08fefd76cb09e54

I did not use to have this icon there. Its there for the past few weeks. When I tap on the screen, its gone until i refresh the page. Whats the point of this? Has this anything to contribute to my latest problems of certain browser mod popups not functioning?

How can I turn this off?

Couldnt find anything in the broswer mod documentation about it.

r/CryptoMarkets Soyic

Solana-Sui month-to-month Price Behavior

At the beginning of 2025, I saw how the price behaves and taking only some data from the chart gave me this behavior month by month.

2025

Solana:

January - Up

February - Down

March - Down

April - Up

May - Up

June - Down

July - Up

August - Up

September - Equal

October - Down

November - Down

December - Down

SUI:

January - Down

February - Down

March - Down

April - Up

May - Equal

June - Down

July - Up

August - Down

September - Equal

October - Down

November - Down

December - Equal

Conclusions:

The price of Solana in the month of January was high and in SUI it was low

When the month of July arrived, Solana continued to rise in August and September. SUI rose in July but August and September did not exceed July of SUI, determining January was the high point of the price and that September was the high price in Solana and the high price was July in SUI.

As of today, the price in both for the month of April determines that they are Bullish because the month of March did the same as March 2025. And when both Solana and Sui have different behavior, it will be the determining factor for the change of direction. As well as the strength that is included in the price within certain candles.

r/megalophobia __Dionysus___

Standing underneath the Eiffel Tower

I was not fully prepared mentally for how big it is in person. And yes went to the top and I was on the verge of a panic attack the whole time.

r/ClaudeAI Low-Scale4343

I kept losing Claude conversations while switching between AI tools — so I built something

Does anyone else use Claude alongside ChatGPT or Gemini? My workflow jumps between all three depending on the task, and keeping track of threads across separate platforms drives me crazy. I put together a Chrome extension called VaultChat — it sits as a side panel while you browse and auto-saves your conversations from Claude, ChatGPT, and Gemini into folders you can search and organize. All stored locally, nothing leaves your browser. There's also a Compare AI mode that fires the same prompt to all three platforms at once and displays the responses side by side — useful when you're not sure which tool handles a task best. Still early (v0.1.0), but it solves the problem I had. Let me know if you find it useful or have ideas for what to add next. Built using Claude Code as my primary development tool — it handled most of the implementation while I focused on architecture and testing. The free tier (10 folders, all three platforms) is available immediately with no sign-up. Chrome Web Store: https://chromewebstore.google.com/detail/cnegpkbkebmfpkcfndpkmdfjdhhlickp 
SortedFor.me