Activity
Mon
Wed
Fri
Sun
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Jan
Feb
Mar
Apr
What is this?
Less
More

Owned by Daniel

Digitally Demented

21 members • Free

AI isn't a tech problem. It's a psychology problem. Daniel Walters teaches you how to think with AI — not just use it.

Memberships

Skoolers

196.1k members • Free

21 contributions to Digitally Demented
I've built something. Want your honest take.
Some of you have heard me talk about the system I use to run my consulting business — the AI operating system I've been building for the past 10 weeks. I finally put a name on it and a website behind it. It's called Refracted Cortex. The short version: it sits on top of whatever AI you already use (Claude, GPT, Gemini) and makes it actually remember who you are. Your values, your commitments, your blind spots. It doesn't reset every session. It pushes back when your decisions don't match what you said matters to you. Everything I've been teaching in Connected Intelligence about context, cognitive architecture, and thinking with AI — this is what it looks like when you turn that into a product. The site is live and I'm opening a founding waitlist. Wanted to share it here first, before I share on LinkedIn tomorrow. 20 spots at $97/mo (locked for life — standard will be $197). But honestly? I'm posting this here first because I trust this group's judgment more than the algorithm. If you don't mind talking the time, I'd like to know: 1. Does the concept land when you read the site? 2. What's confusing or feels like a stretch? 3. Would you use something like this? Why or why not? Link: https://refractedcortex.ai Brutal honesty welcome. That's how this place works.
1 like • 8d
@Paul Harbin Currently no - but HIPPA protection is on the roadmap. It current does scan for PII, so the ability is already in the system. Just have to re-craft it for HIPPA.
Update - AL AI Innovation Summit Next Week
There are a couple of great academic events going on locally (here in Alabama) next week. The big one for me is the AL AI Innovation Summit - to which my poster presentation has been ACCEPTED!!! I'm excited to bring the idea of cognitive architecture and a working prototype to the summit next week. Also - over the weekend, I'm working on finishing getting my own cognitive architecture online, with the ability for others to use! There will be free trials available for people to be able to try it to see if it's right for them. I hope that all of you will be able to try it out. And as a thank you for being a part of this community - I'd like to extend the free trial for each of you for an additional two weeks. More details and an announcement post to come...
Your AI is only as honest as your data
I'm prepping a talk for a sales group tomorrow and I keep coming back to the same thing. Most people think AI's big promise is speed. Get answers faster, automate more, scale everything. And yeah, it does that. But here's what nobody's talking about: **AI doesn't question your inputs. It amplifies them.** If your CRM notes are written a day after the conversation — you're not giving AI the truth. You're giving it a reconstruction. If your project tracker says something is "in progress" because nobody updated the status — your AI sees active work. The project's been stalled for two weeks. This isn't an AI problem. It's a context problem. I ran into this while building my own system. The AI wasn't wrong — it was confidently right about garbage data. The output looked great. The thinking behind it was hollow. Here's the test I keep running on myself: What do I know right now that isn't in any system? That gap — between what's in your head and what's in your tools — that's where AI goes blind. What's something you know about your work right now that isn't written down anywhere? And what would change if your AI actually had access to it?
Tiago Forte just validated everything you're building.
If you follow the PKM (Personal Knowledge Management) world at all, you probably saw this: Tiago Forte — Building a Second Brain, 1M+ followers — just announced something he’s calling “Personal Context Management.” He’s launching an “AI Second Brain” cohort around the idea that your personal knowledge system needs to become the context layer for AI. Sound familiar? I’m not saying this to gloat. I’m saying it because this matters for you. When someone with Tiago’s reach tells a million people that the future is organizing your thinking so AI can actually use it — that’s not competition. That’s air cover. He just did millions of dollars worth of market education for the exact problem we’re solving. The difference is in what happens next. Tiago is selling a cohort. You’re building architecture. A cohort ends. You get frameworks, maybe some templates, and then you’re on your own. What you’re building here — CLAUDE.md files, agent systems, handoff protocols, the whole cognitive architecture — that compounds. Every session makes it smarter. Every agent learns your context better. Every workflow you design becomes infrastructure you own. Tiago’s cohort will teach people to organize context for AI. You’re already deploying it. Here’s the strategic play for this week. I’m publishing LinkedIn content that rides this wave — connecting what Tiago announced to what cognitive architecture actually looks like in practice. The timing is perfect. I need your help amplifying it. The post is up now - https://www.linkedin.com/posts/danielwalters_cognitivearchitecture-aiworkflow-activity-7441923448932765696-e7VH 1. Like them (algorithm fuel) 2. Comment with your own experience (social proof that isn’t me talking about me) 3. Share if it resonates (extends reach beyond my network) This isn’t vanity metrics. When a million people just got told “personal context management is the future,” and our community is already doing it — we want to be visible in that conversation.
90% of people using AI are using it wrong — and it's not their fault.
Harvard Business Review just published one of the most important AI studies I've seen. They tracked 2,500 employees at KPMG over 8 months. Analyzed 1.4 million AI prompts. The finding: 90% adopted AI. Only 5% use it with any sophistication. That's not a training problem. KPMG already trained these people. They had access, they had tools, they had support. And still — 85% of them are basically using a Ferrari to drive to the mailbox. Here's what surprised me most: how often you use AI has almost nothing to do with how well you use it. The "just use it more" advice is dead. The study killed it with data. The 5% who actually get results? Four things set them apart: 1. They treat AI as a reasoning partner, not a search engine 2. They delegate complex, multi-step tasks — not one-off questions 3. They define roles, constraints, and success criteria before they prompt 4. They use AI as a general-purpose thinking tool across their whole job — not just for writing emails And here's the part that matters for everyone in this community: the sophisticated users were almost all experienced professionals. Not the youngest people in the room. Not the most "tech-savvy." The people with the deepest understanding of their work. Your experience IS the advantage. Contextual range — knowing what good looks like because you've seen bad — is what makes AI actually useful. AI doesn't replace your judgment. It amplifies it. But only if you know how to think with it, not just use it. The 85% gap isn't going to close with better prompts or more YouTube tutorials. It's going to close when people stop treating AI as a tool and start treating it as an extension of how they think. That's what we're building here. **What's your experience?** Are you in the 5%, the 85%, or somewhere in between? And what do you think is actually holding most people back?
1-10 of 21
Daniel Walters
3
43points to level up
@daniel-walters-4523
Creating learning moments in people's lives, including his own... one Skool course at a time...

Active 3d ago
Joined Aug 21, 2025
INTJ
Birmingham, AL
Powered by