Friday, March 6, 2026

Problem Framing vs. Solving

 


A quick google search about "Best Books - Problem Solving" will come with 10s if not 100s of recommendations and best sellers list such as: https://www.amazon.com/Best-Sellers-Decision-Making-Problem-Solving/zgbs/books/2679

Now if you search for "Best Books - Problem Framing", you with either see the problem under the category of problem solving, maybe in few cases the books on "Systems Thinking" will appear in the search.


So while 99.99% books will teach you "problem solving", there are just handful and not well known books out there on "problem framing". You will likely come across recommendations in that category from your friends in consulting circle.


If you are ever go to one of those popular consulting schools, they taught me that brilliant people fail when they answer the wrong question.

Don’t just answer questions. Frame them. Because a brilliant answer to the wrong question is still wrong. Ask, “How do we make customer support more efficient?” and everyone races to cut headcount or automate.

You might save dollars and bleed trust. Try this instead: “What service approach builds loyalty while balancing cost?” Now you are designing for humans, not just a spreadsheet. How you frame a question shapes what you notice, what you measure and what you ship. Daniel Kahneman and Amos Tversky called this the framing effect. It’s one of the most underrated leadership skills.

I learnt the value of spending time on framing the question during my time at consulting schools.


At first it felt forced. But projects (case studies) where we invested serious time up front to define the question led to sharper insights, faster decisions and happier teams & clients.

When we didn’t take the time, chaos reigned. Put it into practice this week: 1. Question the question. ↳ What assumptions are baked in? What if you flipped it on its head? 2. Start at the finish line. ↳ Define outcome or experience you want, then trace back the decisions and actions that create it. 3. Make space for the devil’s advocate. ↳ Assign someone to challenge whether you’re even solving the right problem. If you work with data or roll out new tech, your analysis is already shaping outcomes. Make sure you’re shaping the right ones. Have you ever felt like you’ve missed the mark on the question you’re answering?
What's one question your team has been wrestling with that might need a reframe?

Further Readings:






HTH...

A Tech Artist ðŸŽ¨

Friday, October 3, 2025

A Short No Nonsense Guide To Succeed With The Agentic AI - To Drive Business ROI & Maintain Sanity

 







If you're diving into the world of Agentic AI, you're probably hearing a lot of hype. These are AI systems that act autonomously—like little digital agents handling tasks, making decisions, and even chaining actions together to solve complex problems. Think of them as supercharged assistants that don't just respond to prompts but actually pursue goals on their own. Sounds revolutionary, right? But let's cut through the noise: Agentic AI isn't some magical wand that's going to transform your business overnight. It's a tool that demands smart planning, execution, and a healthy dose of realism to deliver real value and ROI.

The truths? When done right, Agentic AI can streamline operations, boost efficiency, and open up new possibilities in areas like customer service, supply chain management, or even creative workflows. The lies? It's not "set it and forget it" tech. A lot of these systems lean on large language models (LLMs) like GPT or similar public ones to handle reasoning and language tasks. But here's the kicker: despite all the buzz, LLMs aren't truly reasoning in a human-like way. They excel at pattern matching and generating plausible outputs based on massive data, but they lack genuine cognition—no real understanding, no common sense beyond what's trained in, and they can hallucinate or falter on novel problems. Studies and real-world tests show they're more like sophisticated parrots than thinkers. So, if you're building Agentic AI, expect to integrate LLMs, but don't rely on them for deep reasoning without safeguards.

The key to success? Treat it like any major project: break it down into phases with clear goals across short-term (quick wins), mid-term (scaling up), and long-term (transformation) horizons. Define your path, measure success not just in tech metrics but in business ROI—like cost savings, revenue growth, or improved user satisfaction. Below, I'll walk you through a practical thinking process to make this happen. This isn't theoretical fluff; it's a roadmap based on how real teams are approaching it.

1. Narrow Down on a Use Case and Get Your Basics Sorted

First things first: don't boil the ocean. Pick a specific use case where Agentic AI can shine—maybe automating report generation in finance or handling inventory predictions in retail. Once you've narrowed it, map out the essentials.

  • Sponsors and Actors: Identify who’s championing this (e.g., C-suite execs for funding) and who’s involved day-to-day (like developers, domain experts, or end-users). Get buy-in early to avoid roadblocks.
  • Scope: Define boundaries—what's in (e.g., handling customer queries) and what's out (e.g., no legal decisions). This keeps things focused.
  • Success Metrics: Set measurable goals. Short-term: Completion rate of tasks. Mid-term: Reduction in manual hours by 30%. Long-term: 20% ROI boost through efficiency gains.
  • Options and Players: Explore what's out there—open-source tools like LangChain for agent frameworks, or vendors like Microsoft’s AutoGen. Assess competitors or similar implementations.
  • Readiness and Assessment Methods: Audit your current setup. Do you have the data infrastructure? Use checklists or pilot audits to gauge gaps, like running a SWOT analysis on your tech stack.

This step is your foundation. Skip it, and you're building on sand.

2. Build vs. Buy vs. Co-Develop: Choose Wisely

Now, decide how to get your Agentic AI off the ground. Building from scratch gives customization but eats time and resources—think months of dev work integrating LLMs with custom logic. Buying off-the-shelf (e.g., pre-built agents from Hugging Face or enterprise solutions) is faster for short-term wins but might not fit perfectly, leading to integration headaches.

Co-developing? That's often the sweet spot—partner with a vendor or open-source community to tweak existing tools. Weigh costs: Building might cost $100K+ in dev salaries, buying could be a subscription at $10K/year, co-dev a hybrid. Factor in your team's skills and timeline. For mid-term goals, aim for something extensible; long-term, ensure it scales without vendor lock-in. Remember, since agents often hook into public LLMs, test for reliability—LLMs can change APIs or behaviors, so have fallbacks.

3. POC and Fine-Tune: Test the Waters

Jump into a Proof of Concept (POC) to validate your idea without going all-in. Start small: Build a prototype agent for your use case, say one that automates email responses using an LLM backbone.

Run experiments—feed it real data, simulate scenarios, and iterate. Fine-tune by adjusting prompts, adding rules-based logic where LLMs fall short (e.g., hard-coded checks for accuracy), or even training on domain-specific data if possible. Measure against your success metrics: Did it handle 80% of tasks correctly? Track failures—LLMs might "reason" wrong on edge cases, so log them.

This phase is short-term focused: Aim for a working POC in weeks, not months. Use it to gather feedback and refine, ensuring the agent isn't just flashy but actually useful.

4. Scale and Integrate: Make It Part of the Ecosystem

Once the POC shines, ramp up. Scale by deploying to more users or data volumes—move from testing 100 tasks to 10,000. Integrate with existing systems: Hook your agent into CRMs like Salesforce or databases via APIs.

Watch for bottlenecks—LLMs can be slow or costly at scale, so optimize with caching or hybrid models (rules + AI). Mid-term goals here: Seamless integration without disrupting ops, hitting metrics like 99% uptime. Plan for data flows: Ensure secure, compliant connections, especially if using public LLMs that might expose sensitive info.

5. Parallel Run for Ops: Safety Net Mode

Don't flip the switch abruptly. Run your Agentic AI in parallel with human processes—let it shadow real ops, comparing outputs. For example, have the agent suggest inventory orders while humans approve.

This builds confidence and catches issues, like LLM hallucinations leading to bad decisions. Short to mid-term: Monitor discrepancies, refine based on them. Success metrics? Alignment rate with human judgments, say 95%. It's your buffer to ensure reliability before full rollout, minimizing risks in critical areas.

6. Pick on Skills and Get Experience: Level Up Your Team

Agentic AI isn't plug-and-play; it needs skilled hands. Identify gaps—do you need prompt engineers, data scientists, or ethicists? Invest in training: Workshops on agent frameworks or LLM limitations.

Gain experience through hands-on projects. Start with internal hackathons or collaborate on open-source agents. Long-term: Build a center of excellence. This drives ROI by reducing dependency on externals—your team becomes the asset, turning one-off successes into repeatable wins.

7. Strengthen Guardrails and Improve Observability: Keep It Safe and Transparent

Here's where you address the lies head-on. Since LLMs lack true cognition, add guardrails: Rules to prevent harmful actions, bias checks, or human-in-the-loop for high-stakes decisions.

Boost observability with logging tools—track every agent decision, input/output, and LLM call. Use dashboards for real-time monitoring. Mid to long-term: Evolve these as threats emerge, like new LLM vulnerabilities. Metrics? Error rates below 1%, compliance scores. This isn't optional; it's what turns potential disasters into managed risks, ensuring sustainable success.

8. Takeover and Transform to Drive ROI: The Big Payoff

Finally, let the agent take over where it excels, phasing out manual processes. Transform your ops—reallocate humans to creative tasks, unlocking innovation.

Drive ROI by tracking hard numbers: Cost savings from automation, revenue from faster decisions. Long-term: Evolve the agent ecosystem, maybe chaining multiple agents for complex workflows. But stay vigilant—regular audits, updates for LLM advancements. If done right, this isn't just tech; it's a business multiplier.

In wrapping up, Agentic AI is powerful, but only if you ditch the hype and embrace the grind. By following this phased approach, you're not chasing magic—you're engineering success. Plan meticulously, measure relentlessly, and integrate wisely. The result? Real, tangible wins that boost your bottom line. If you're starting out, pick one use case and iterate from there. You've got this!


One-page checklist (print this)

  • Named sponsor & owner, signed success metrics

  • Golden set & evaluator defined

  • Guardrails v1 (policy + filters + approvals)

  • Shadow run results and incident log

  • Integration plan (APIs, data, auth, SLOs)

  • Observability dashboards live

  • Parallel-run exit criteria met

  • ROI model and phase-out of legacy step




HTH...

A Tech Artist ðŸŽ¨

Wednesday, October 1, 2025

Have you tested your strategy lately?


Ten timeless tests can help you kick the tires on your strategy, and kick up the level of strategic dialogue throughout your company.

 

Test 1: Will your strategy beat the market?

Test 2: Does your strategy tap a true source of advantage?

Test 3: Is your strategy granular about where to compete?

Test 4: Does your strategy put you ahead of trends?

Test 5: Does your strategy rest on privileged insights?

Test 6: Does your strategy embrace uncertainty?

Test 7: Does your strategy balance commitment and flexibility?

Test 8: Is your strategy contaminated by bias?

Test 9: Is there conviction to act on your strategy?

Test 10: Have you translated your strategy into an action plan?



HTH...

A Tech Artist ðŸŽ¨

Tuesday, September 2, 2025

So What Really Went Wrong with AI & GenAI Movement ? & How We Can Do Course Correction. (And an opp. to win $100)

 


As soon as MIT report came out few weeks ago suggesting 95% of GenAI pilots fail (The number is close to what Mckinsey suggested as 86% or something a long time ago) or IDC Lenovo report numbers suggesting it as 88%.

So in the world where everyone wants to hop onto the GenAI bandwagon, what really went wrong so far and is AI/GenAI actually a good solution for some specific problems? Let's examine:

Over the last 4 years (since the chat gpt movement in late 2021), I have spent good amount of time learning and practicing AI (perhaps 1000+ hours), beside talking to 1000+ people across industries (IT, Banking, Retail, Aviation etc.)  &  across the roles (Leadership, Mid Management, Employees and even HRs) as part of my market research.

My findings suggest (Of course you may share a different view):

1. Most leadership people have absolutely no clue about what AI can and can't do (They can't even tell the firm differences between AI & Automation beyond generic and over simplified examples). Because no one is keen to dive deep into the topic at hand and pretty much rely on media reports, what partners/OEMs tells them,  60 mins. online courses, what their peers or reportee project them as AI being silver bullet and what not.

Of course sooner or later, the FOMO kicks in very quickly beside getting a push from CEO or board to do something about AI without a clear direction.

2. So most organizations AI strategy is nothing but false narratives and FOMO on steroids. 

90% of employees in those organizations have no real access to AI/GenAI capabilities (because since there is no clear strategy & needs specific trainings available for them to leverage those beside costs comes in to the equation), shadow AI becomes prevalent and it often contributes to AI psychosis without people necessarily noticing it. 

This also put the organisation wide knowledge management at risk and no one is noticing it. Since everyone is told by their seniors - Did you chat gpt it to find the answers?

3. I haven't come across a single real world DS practitioner in my circle who would articulate AI/GenAI as silver bullet or magic. There are some good use cases that AI/GenAI can deliver, but the underlying foundation is build on clear - strategy, automation, cross teams collaboration and data with the firm use cases and roi beside an active innovation mindset. 

I love the way BCG folks articulated it with their basic framework for AI success pillars beside the HBR's Digital Maturity model.

And if you are still not convinced, I'll be giving a $100 to to show me how GenAI is completely transforming the landscape in any context when applied to complex problems. Just drop me a note (anetworkartist@gmail.com) to have a 30 mins. zoom meeting to surprise me. We will follow the BCG's GenAI Success Stairway to have a quick sanity check and determine the business impact.

Further Readings:














HTH...

A Tech Artist ðŸŽ¨