Zen and the Art of AI Adoption: Surviving the Scare Trade to Thrive in the Age of AI

“the AI scare trade is speeding up AI transformation by years at tens of thousands of businesses.” (Nate B. Jones, Feb 19, 2026, Why the Biggest AI Career Opportunity Just Appeared—and Almost Nobody Sees It.)

Anyone who is focused on enterprise AI adoption and heard this on Nate B Jones‘s AI News & Strategy Daily channel or read it on his substack will be concerned about how this all works out on several levels. Let’s look at a few key impacts:

  • Your Investments
  • Your Career
  • Your Business

That list is not in order of importance; it is in order of a sequence of events in history where each of these aspects has had similar drivers. I will be referring to quotes from Why the Biggest AI Career Opportunity Just Appeared—and Almost Nobody Sees It. throughout, and I recommend you watch it after reading this because it has a lot more to offer on other topics, too.

Investments

While I agree that last year was not the AI bubble, there is one coming soon. Anyone that was part of the world of tech in 2000 will see the echo here:

Meanwhile, AI startups, regardless of whether they’re good or not, look relatively more attractive to everybody. You stick AI in the name, and magical things happen right now. Magically, more capital will flow to the AI company that has AI in the name and releases an AI press release than to anybody else. (14:23)

In 1999, companies with “.com” in their name were guaranteed a huge IPO, even if all they had to offer was a cool sock puppet as their mascot. Today, a similar narrative is forming where Wall Street hype and Media hype are creating a hype-vortex, drawing in capital regardless of actual value. When the Dot Bomb blew up the Dot Com Bubble, I watched friends near retirement age having to change their life plans with no choice because it wasn’t just tech stocks that took the hit, and it took the market 15 years to rise back up to the peak of the previous century. Diversification was not a watchword in those days; even for those that were diversified, the impact rippled.

Your Career

…the CFO pulling forward cost cuts to demonstrate to investors that management does take this transition seriously. Stock drop doesn’t just reflect reality, it creates reality. A company whose stock craters on AI fears is going to start behaving as if AI is an existential threat. Even if the actual tech is years away from threatening its core business, defensive postures get adopted right away. (4:58 to 5:19)

Lots of people have already had their lives upended by companies determined to show investors that they are becoming more efficient, even (especially?) if they aren’t. This reactive stance often ignores the fundamental reality that human behavior does not move at the speed of a GPU (see Your AI-Driven Digital Transformation is Impeded by Behavioral Challenges).

There are lots of takeaways from the video on this, so, again, watch it after reading this. Meanwhile, a recurring theme on the channel is that if you aren’t thinking about how you can learn to use AI and then figure out how to become more efficient with AI, your career is going to be in trouble. This will come later for some industries as a whole, and for individual businesses because, contrary to media hype, not everyone is going to be making the shift at the same time and certainly not at the same pace. But it is coming, and the speed it is coming at keeps increasing.

There were a lot of people who changed careers in the late 90s into a track that was tied to the .com boom. Only a few survived and then thrived in the rebuild. Of the rest, the lucky ones were able to resurrect their former skills and return to their previous work. The rest often drifted from job to job, sometimes finding a new track, and sometimes not.

Your Business

And because of the prominence of the American stock market, there are boards all over the world looking at this. Now visibility like this is what turns a slow trend into an urgent capital reallocation in favor of AI. I’m not kidding when I say the AI scare trade is speeding up AI transformation by years at tens of thousands of businesses. The scare trade is a transfer of career capital from the people who treated AI as somebody else’s problem to the people who have been invested in understanding it. (27:59 to 28:24)

The above quote is what inspired this post. I’ve been advocating for a while now that most businesses have been going all wrong about how they are adopting AI. They are buying the tools and then trying to figure out how to use them. This approach often fails to account for the competitive disadvantage smaller firms face when racing against the scale of industry giants (see Why Bigger Companies Move Faster than You in the AI Adoption Race).

Companies are looking at how others are using tools and assuming it is the tool that is making it work, so they start mimicking the behavior they can see and failing miserably because what makes the tools work is not the results, it is the process of adoption and growth. That is why for every Chase or Walmart example, there are 10 AWS or Replit incidents. Many of those don’t go reported because they happened to a business the size of yours, rather than one that is currently getting media focus on a regular basis.

It’s clear that some businesses are going to rush into their AI adoption approach. Some already have. Some will be like children that touch the hot stove (healing in a couple of weeks and exercising caution), and a few may even become great chefs. Others will be more like the fictional inventor from The Expanse, but without the rest of the world benefiting from his demise:

Lesson learned? Faster is better when you build speed, not when you jump straight from 0 to splat.

Post title inspired by both Zen and the Art of the Internet: A Beginner’s Guide and Zen in the Art of Archery. While the title Zen and the Art of Motorcycle Maintenance is the inspiration for the former, Zen in the Art of Archery is much more along the lines of this post and the AI adoption, though truthfully the adoption of AI, like your situation, is unique unto itself.

Looking for help adopting AI in your organization? Let’s talk. Tag me in a comment or reach out directly with a connection request.

If you found this interesting, please share.

© Scott S. Nelson
The New Digital Divide is Analog

Your AI-Driven Digital Transformation is Impeded by Behavioral Challenges

The recent article by CT Crooker, Why Everything You Know is Probably Wrong, is filled with hard truths that everyone in IT needs to consider. It starts by pointing out the evidence supporting the thesis that things are going to be very different.

“Going to be” is the one level where I depart from a lot of recent articles by really brilliant people. When discussing the unprecedented acceleration of new and improved capabilities that come under the media definition of AI, these experts are not only correct in their assessments of the rate of change; they understand the details of those changes better than most.

However, they often present these shifts as a present-tense reality for the masses. For the vast majority of organizations, these changes are still in the “going to be” phase because the experts are focusing on a very active and very small minority.

Then there are people.

  • Most CI pipelines aren’t really continuous and don’t truly integrate.

  • Teams hold stand-ups and manage backlogs that aren’t the least bit Agile.

  • Enterprise CRM systems are treated as glorified address books while the predictive analytics and automation features sit dormant.

  • Smartphones are used for scrolling while the powerful sensors and computing power in our pockets remain largely untouched.

The main impedance to technical solutions is rarely a technical problem. The real culprits are process and culture challenges that act as a silent brake on innovation. This resistance to change usually stems from a deep-seated fear of the unknown or a perceived threat to the status quo.

When a new capability arrives, it doesn’t just offer a faster way to work; it threatens the established hierarchy, the “way we’ve always done it,” and the specialized knowledge that individuals have spent years protecting. These psychological hurdles are the biggest obstacles to adding and improving technical capabilities. It will take significant time before these new tools make it into mainstream IT departments because human behavior does not move at the speed of a GPU.


A Challenge by any Other Name is…Entirely Different

This brings me to the point of my only contention with the article. I disagree with the suggestion that “transformation impedance” is a better way to think about these shifts than “epistemic flexibility under inversion.” While I find the shift in terminology problematic, Crooker’s post is otherwise incredibly thought-provoking and accurate; it is really valuable that he raised these points because they are essential to consider.

He explains “epistemic flexibility under inversion” as a capability characteristic of both systems and people to adapt to rapid changes and then adopt new approaches as a result. He goes on to suggest that “transformation impedance” may be a better way to think about it.

But branding is more important than most realize. People who take up the call of “transformation impedance” will be more likely to focus on the impedance side, which leads to conflicts between those who think everyone should reduce the impedance versus those who want to lower it. I’ll admit there is some room for collaboration on the rate of lowering impedance, but then again, there are still a lot of those CI pipelines that are still neither.

First, I will admit that I had to look up the definition of “epistemic flexibility under inversion” to fully digest it:

“Epistemic flexibility under inversion” is a specialized concept often found at the intersection of Bayesian statistics, cognitive science, and information theory. It refers to a system’s (or a mind’s) ability to maintain a coherent understanding of reality even when the “direction” of information flow or the relationship between cause and effect is flipped.

Once I had this better understanding, I had the same reaction to using “transformation impedance” as an alternative as I do to changing “issue” to “challenge.” (There is a lot more to that definition, of course, and I suggest you talk with your favorite Generative AI LLM to get the rest of the picture.)

The Utility of the Negative

Media tells us we should always be positive and pursue higher goals. We buy into this because the truth is that the method of using the negative to drive action, specifically addressing an “issue,” is much more likely to succeed than the message of chasing a dream. That’s another hard truth.

I like “issue” better than “challenge” because people will deal with an issue so it will go away. A challenge makes them feel good about pursuing it, and since the pursuit is the reward, completing it removes the reward and thus the incentive. If it is an issue, the incentive needs to be to correct it.

While “epistemic flexibility under inversion” may be harder to understand, it keeps the focus on how we need to change our approach to deal with the changes approaching us. “Transformation impedance,” on the other hand, is a label describing a phenomenon and doesn’t necessitate action until it is too late.

We need to flip our approach and find ways to catch up with change and not be left behind or run over. We should begin thinking about what problems need to be solved for our businesses, and even our lives, that for whatever reason we thought were too hard before, and then come up with new solutions taking advantage of the AI. To do that, we must be willing to set aside the old frameworks that impede our ability to do so.

If you found this interesting, please share.

© Scott S. Nelson

A Silver Bullet should Break Golden Chains

A recent exploration into the “frictionless trap” addressed how taking the easy path can weaken personal abilities and lower the collective capabilities of humanity. This observation did not imply that the opposite, such as excessive and grinding labor, is desirable. A BBC article serves as a stark reminder that in certain sectors, people are working far too much. While this trend has been reported on previously, a Slashdot post providing commentary on that specific article served as the final catalyst for this rant.


The Philosophy of Ease

Under the title of this blog is the statement: Technology should make things easier. The original driver behind my blog was to simplify complex tasks so that a struggle encountered the first time would be much easier the next. My focus was on tasks requiring repetition but occurring too infrequently to become muscle memory, which is a concept similar to the Second Brain framework later branded by Tiago Forte.

Under its original name and domain, my blog caught the eye of an editor at Developer.com. This led to me writing how-to articles on processes that were time-consuming to figure out but simple to execute once all the steps were gathered and sequenced. Generous copyright rules allowed me to republish these pieces after a holding period with proper attribution.

This shifted the focus of my blog toward making things easier for others. The beauty of making a task easier is that it frees people up to spend that time on more productive, interesting, and creative pursuits.


Historical Leverage: The Promise of Progress

History reveals a series of technological leaps designed to trade mechanical effort for human potential. The transition from foraging to settled agriculture allowed humanity to move beyond the daily search for calories. This newfound surplus of time provided the foundation for the birth of philosophy, mathematics, and complex governance.

During the Industrial Revolution, steam and steel began to replace human and animal muscle. Tasks that once required an entire village to complete over several weeks were suddenly finished in mere hours. This shift was theoretically intended to liberate the worker from the most back-breaking forms of labor.

By the 20th century, the “electric servant” arrived in the form of home appliances. Washing machines, vacuums, and ovens were marketed as the ultimate liberators of the domestic sphere. These tools promised to turn hours of physical toil into the simple push of a button, reclaiming life from routine chores.

The digital age followed with the promise of the paperless office and instant data processing via computers. Spreadsheets replaced rooms full of ledger-keepers, and word processors eliminated the need to re-type entire manuscripts. In every era, the pitch remained the same: efficiency would set the individual free.


The Darker Side: The Persistence of Burden

Despite these advances, the time saved has often been redirected into new forms of systemic entrapment. The agricultural revolution, while providing stability, was frequently accompanied by the rise of feudalism and organized slavery. In these systems, the efficiency of the land was not used to grant leisure to the tiller but to consolidate power and wealth for the few at the top.

The industrial era followed a similar pattern of redirected effort. Rather than creating a world of leisure, the introduction of the machine often birthed the sweat shop. Workers were required to labor for 16 hours a day in dangerous conditions just to maximize the output of the new technology. In the modern consumer age, this burden evolved into planned obsolescence, forcing individuals to work longer hours simply to maintain or replace items intentionally designed to fail.

Today, the digital version of this burden has manifested as the 72-hour work week. The efficiency of the computer has not actually shortened the workday for many; instead, it has been used to increase the speed and incline of the productivity treadmill. We have built tools that cut effort ten-fold, but the saved time is often swallowed by a demand for even higher volumes of output.


The Modern Silver Bullet

The conversation around these saved hours has reached a fever pitch with the advent of AI. A recent discussion explores the idea that we may finally have a “silver bullet” for software development. This technology attacks both accidental complexity (the mechanics of coding) and essential complexity (the logic of what to build) by leveraging decades of established patterns. However, the warning remains: while the silver bullet exists, the real bottleneck is no longer the code, but the management. If leadership fails to aim this tool correctly, the result is not liberation, but a “heck of a kick” that could lead to catastrophic failure or even more grueling hours for those involved.


Breaking the Cycle of Diminished Returns

There is no inherent opposition to working hard or putting in long hours. However, there is a strong stance against working to the point of diminished returns. This occurs when the final 42 hours of a marathon week produce less value than the first 30.

There is a fundamental lack of logic in developing a tool that cuts effort ten-fold only to use it 15 times as much rather than using the saved time to improve the lives of people. For business leaders, the goal should be to divide saved resources between improving work-life balance and enhancing the capabilities of the organization.

Working smarter on interesting tasks produces results far superior to grinding for the sake of volume. Technology can improve shareholder value without requiring the sacrifice of human well-being at the altar of effort.

The Question for Leadership: Is accelerating the ROI of AI initiatives worth the cost of driving people to work twice as much?


If you found this interesting, please share.

© Scott S. Nelson
Utopia or Dystopia

The Frictionless Trap: AI’s Greatest Benefit is also a Hidden Risk

I’m a big fan of classic science fiction. I generally avoid dystopian themes, but some are just too good to ignore, from A Boy and his Dog to Hunger Games. When ChatGPT started getting all that popular press a few years back, I was looking forward to finally living in that shiny future promised by Heinlein, Asimov, Clarke, and Roddenberry finally coming true, maybe even a flying car (the current prototypes still aren’t there yet, BTW). But the news of the last few years has had more Brave New World and 1984 vibes.

So when I read a recent NPR report on AI in schools, it felt like another example of how we are engineering frustration out of the human experience. The report describes software that is so sensitive to a student’s frustration that it pivots the curriculum before they even have a chance to get annoyed. On paper, it is a triumph of user experience; in practice, it might be a silent deletion of the very thing that makes a mind grow.

The Lesson of the Eloi

When H.G. Wells sent his Time Traveller into the year 802,701, he didn’t find a high-tech utopia or a charred wasteland. He found the Eloi: beautiful, peaceful, and intellectually vacant creatures living in a world of total automation.

Wells’ speculation in his passage on [suspicious link removed] hits quite close to home in the age of generative AI:

“Strength is the outcome of need; security sets a premium on feebleness.”

The Eloi weren’t born “slow” because of biology. They were essentially optimized into that state by an environment that removed every possible hurdle. They had won the game of civilization so thoroughly that they lost the ability to play it.

The parallel to AI-driven education isn’t that the technology is failing, but that it is succeeding too well. If the machine handles every productive struggle (sensing your confusion and immediately smoothing the path), it isn’t just teaching you. It is doing the mental heavy lifting on your behalf. You don’t get stronger by watching your trainer lift the weights, even if the trainer is a hyper-personalized LLM.

The Mirror of “Useful” Atrophy

It isn’t just about the classroom; AI is becoming a universal solvent for friction. History suggests that when we remove friction, we usually lose the muscle that was meant to overcome it.

  • The GPS Effect: We traded the frustration of paper maps for a blue dot that tells us where to turn. The result is that our internal spatial awareness is basically a legacy system. We can get anywhere, but we often have no idea where we are.

  • The Calculator Trade-off: We offloaded long division to a chip. This was a fair trade for most, but it established the precedent: if a machine can do it, the human brain is officially off the clock for that specific skill.

  • The Infinite Search: We stopped memorizing facts because we treat our devices as an external hard drive for our personalities.

Not all of that has been a bad thing, unless we get to live one of those post-EMP stories (which I avoid reading to avoid remembering it isn’t that far-fetched). I, for one, am glad that Einstein said “Never memorize something that you can look up,” because rote memorization is a struggle for me, but I really do enjoy exercising mental muscle memory. Which is where using AI the wrong way will lead to an atrophy that doesn’t need a major solar event to make us realize things went too far. It doesn’t just provide answers; it simulates the thinking.

The Verdict: Designing for Resistance

We should be optimistic about AI’s potential to amplify us, but we have to be wary of the passenger mindset. If we use these tools to abolish difficulty, we aren’t empowering ourselves. Instead, we are prepping for a very comfortable life as Eloi.

The challenge for educators, and for anyone using an AI “intern” in their daily workflow, is to intentionally design productive friction back into the system. We need AI that makes the work more meaningful and not just more invisible.

Mastery requires resistance. If the road is perfectly flat and the bike pedals itself, you aren’t traveling; you are just being delivered.

If you found this interesting, please share.

© Scott S. Nelson

3 Lies They’re Telling Us About AI

As a writer who has (rarely) been paid for the craft, I find attribution to be not only important, but a moral obligation. Especially when the quote resurfaces in my thinking time and again.

I first read “There are three kinds of lies: lies, damned lies, and statistics” in a Robert Heinlein novel, where he (through his character) attributed it to Mark Twain. Samuel Clemens actually attributed this truism to someone else, though that providence has not been agreed upon by scholars (per Wikipedia), so I will leave this instance of noblesse oblige as having made my best effort. But I digress (a specialty of mine).

Despite my failed attempt at correct attribution, it’s still true. And it is being proven once again, in the various reports and claims around AI. Allow me to categorize them as I see them today:

1. Lies

…are that AI will be able to do X by N. No one really knows how long. Extrapolating future data based on similar but different prior data has an unknown margin of error. This is why every honest company states at the beginning of their demo or presentation or pitch that you should not base purchase decisions on functionality not currently released (and I would add not proven for your specific use case).

2. Damned Lies

…are about what people are accomplishing now. The app vibe coded in a weekend is either going to be feature frozen in the near future, suffer a major failure through bad coding or bad actors, or (most rare) be the result of extensive planning prior to the weekend of wonder.

“The ‘vibe-coded’ apps that fail are not failures of code, but failures of craft.”

3. Statistics

…are the worst of the three, because anyone with half a brain and less scruples can get the same numbers to say entirely different things, like:

Lies Aside

What we call AI (usually) isn’t actually AI. What everyone is calling AI is still the biggest paradigm shift in Information Technology since computers shrunk from needing a room to sitting on a desk (and, not long after, a lap).

Impacts to humanity are a physics phenomenon in that every positive improvement has an equal and opposite potential that will eventually be realized. Whenever the many benefit, a very few benefit a lot more. Some deserve those benefits, and some don’t.

The Mirror and the Machine

We find ourselves at a peculiar crossroads where the technology is accelerating faster than our ability to tell the truth about it. We are told AI is a magic wand, a job-thief, or a savior, but the “Lies, Damned Lies, and Statistics” reveal a simpler reality:

“AI is not a destination; it is a mirror.”

The statistics that show executives “saving eight hours a week” while workers save none aren’t an indictment of the technology—they are an indictment of how we value human time. When we strip away the marketing gloss and the manipulated ROI reports, we are left with the same struggle that has defined every Information Technology shift since the first mainframe: the tension between efficiency and agency.

The Paradox of Progress

If my “physics phenomenon” theory holds true, the equal and opposite reaction to the AI boom will be a renewed, premium demand for the irreplaceably human.

“AI makes it remarkably easy to be mediocre at scale.”

The paradigm shift isn’t just about computers moving from desks to our pockets, or from pockets to our cognitive workflows. It can generate the lies, the damned lies, and the statistics for us in seconds. But it cannot provide the providence of a thought. It cannot feel the moral obligation of attribution. It cannot understand why a quote from a Heinlein novel matters to a writer’s soul.

The New Bottom Line

We should stop asking when AI will “arrive” or if the pilot programs will finally hit their 100% success rate. Instead, we should ask: Who is being served by the current narrative?

If the many are to benefit—and not just the few—we must look past the “weekend of wonder” and the padded ROI spreadsheets. We must demand a version of progress that doesn’t just prioritize the speed of the output, but the integrity of the outcome.

The lies will continue to evolve, and the statistics will continue to shift. But the truth remains:

“A tool is only as profound as the intent of the person wielding it.”

The question isn’t what AI can do for you by 2030; the question is what you are willing to stand for today while everyone else is busy chasing the vibe.


Would you like me to generate a SEO-optimized title tag and meta description to accompany this finalized draft?

If you found this interesting, please share.

© Scott S. Nelson