All about AI -- Aye Yi Yi
If you've breathed air in the last two years then then you've no doubt heard about and likely encountered AI and how pervasive it's become in our world today.
So in this article we'll discuss all things AI. What it is, what it means to you, should you use it, and if so, how to make the best of it.
The AI universe is incredibly complex, populated with some of the smartest software engineers on earth. I'm not an AI expert but I know enough about it to share some salient, useful information with you.
There's a lot to discuss. But, as usual, I'll avoid getting too deep in the weeds and will define any overly technical terms clearly. Hopefully you'll leave my article with a decent understanding of AI and its impact. That's all I want!
Pop quiz question! When did the earliest "AI" of any importance come out?
Meet Eliza
Eliza was born in the mid-1960s. Her father was Joseph Weizenbaum, an MIT computing researcher. She was a precocious little thing, learning how to interact with people, almost from her birth.
But Eliza* wasn't a little girl. She was instead a computer program -- one of the earliest programs capable of natural language conversation. She was a chatbot.
Yes, that's right. A chatbot, in the mid-1960s no less! Only we didn't have that term back then.
* Eliza (the program) was named for Eliza Doolittle, the Cockney English flower girl in My Fair Lady, being taught to speak "Proper English".
Eliza's algorithms, an advanced idea at the time, were pretty simple by today's standards. Eliza combined pattern recognition by parsing typed-in sentences for noun and verb structure along with asking open-ended questions, to produce a realistic, seemingly human interaction.
Here's a sample of how a conversation with Eliza might go:
Person: I feel stressed at work.
Eliza: Why do you feel stressed at work?
My boss gives me too many tasks.
Tell me more about how your boss gives you too many tasks.
Well, I don’t have enough time to finish them.
Why do you think you don’t have enough time?
Because there are only so many hours in a day!
Can you elaborate on that?
Never mind, this isn’t helping.
Why do you say this isn’t helping?
heh heh... Eliza could go on like this all day if you allowed it.
In the 1960s, you might have said it's an example of AI, however crude.
To be sure, this was not AI. Eliza knew absolutely nothing. Not even 2+2. There was no knowledge base from which to draw answers.
This may seem whimsical, and I suppose it is, but make no mistake, the idea of a natural language processor (NLP) like Eliza was foundational to the advancement of AI research that followed.
Now meet Siri, Alexa, and Cortana
These three gals are quite advanced compared to Eliza -- they're all pretty smart. But even this threesome are not good examples of AI.
Some might argue these are examples of narrow AI, limited to specific tasks. But that's too generous. To me, to earn an "AI badge", a system should be capable of answering disparate classes of questions by examining its training data and in ways not anticipated by the system designers. Siri, Alexa, and Cortana aren't there (yet). Cortana may never get there. She was put to pasture, at least in the consumer space.
These are more aptly called digital or virtual assistants. You can ask them to perform a number of tasks, but only tasks that have been preprogrammed. They are capable of NLP (Natural Language Processing) but that's not the same as full AI.
ChatGPT lands with a boom
In November 2022, OpenAI released their first publicly available version of ChatGPT. And, wow, what a thing that was. ChatGPT rapidly shot to the top becoming one of the most popular apps downloaded.
For the first time, everyday people could ask questions and get useful, actionable answers on the spot. Not just Google-like search results.
This is what I consider AI.
How does it work?
That's is a huge, complex question, with technical answers far beyond the scope of this article. I'll keep it to a useful executive overview.
ChatGPT is built on self-attention (not self-awareness!) mechanisms, allowing it to "understand" context, relationships between words, and generate coherent responses. It learns grammar, reasoning patterns, and styles but does not (usually) rely on direct access to live data or real-time internet. That would be far too slow.
Instead, it's trained on a huge amount* of data that comes from the internet. That data includes millions of books covering countless topics, newspaper articles, academic papers, websites, social media, and whatever else its web crawlers can scrape** up. That data is analyzed and optimized according to the training model and used by ChatGPT to provide input when generating its responses.
* ChatGPT's training dataset is about 45 terabytes of data. In terms of type written pages, that would be something like 13 billion pages, or a stack of paper about 800 miles tall. Crazy.
** Scraping/crawling means bulk, wholesale scanning of millions of websites to gather up data. For an AI crawler, that data goes into the training model. For a search engine crawler, the data goes into a database that is indexed for fast search and retrieval, like Google search.
Let's parse the word "ChatGPT":
-
Chat: It's a chatbot. It uses NLP (Natural Language Processing) to "understand" your input (called "prompts" in AI parlance) and to create human understandable responses.
-
G=Generative: Responses are generated on the spot specifically for your prompt. They are not canned or preprogrammed.
-
P=Pretrained: The vast data reservoir, collected mostly from the internet, imported into the AI engine.
-
T=Transformer: The process used by a particular AI model by which a response is generated. There are several.
The amount of processing power required by ChatGPT, and other AI systems based on models similar to ChatGPT, is truly enormous. Depending on who you ask, a ChatGPT prompt might consume upward 25x more processing power (electricity) than a Google search.
Existing AI systems are also eye wateringly expensive. Not just in the power used (which is enormous), but in the computing infrastructure needed to make it all work. The top Big Tech firms have cumulatively spent hundreds of billions of dollars on AI initiatives in just the last two years alone.
AI vs ChatGPT
Is "AI" and "ChatGPT" the same thing?
ChatGPT is just one implementation of AI architecture, but it's the one most people know. It's highly visible because it stands on its own as a retail (consumer facing) chatbot that you can use on your computer or phone to ask questions of, or to just shoot the breeze with.
AI is an all-encompassing term, a convenient "handle", if you will, to label the big, wide world of artificial intelligence initiatives. And, boy, are there a bunch of them.
We're starting to see AI everywhere these days. We'll cover some of that farther down.
AI vs Search
What's really the difference here?
We'll use Google as the example.
When you do a search in Google, the search engine very rapidly examines an enormous database that was previously compiled and is constantly updated, containing most of the surface web's* information.
Google isn't "searching the web" as you type in a search query -- that would be impossibly slow. It already has all that data stored and indexed for extremely rapid retrieval. That's why Google is so fast.
* The surface web is content that is ready available, at no cost, no access blocks, simply by browsing to that site, like my website that you are reading right now (did I say thanks?). Contrast that with the deep web which is information that's not readily available, such as data contained in databases, behind paywalls, or behind corporate and government firewalls.
But the key point here is that Google just returns links to web pages that (probably) contain more info on what you seek. If your search is a popular one, like how old is Biden or Trump, Google may return the result as a direct response, rather than pointing you to web pages that contain the answer. But, usually, it's a page full of links that you have to sort through.
With an AI-based approach, like ChatGPT, it's more analogous to having a helper sitting next to you that you can ask questions of. Your helper then does all the scut work of searching and collecting information on your behalf and compiling a useful, personalized answer.
You could think of ChatGPT as an "automated super google" that not only searches, but also presents the results in a more immediately useful form.
I ain't gonna lie. That's mighty useful.
Truth and Trust
But there's a downside, too.
We all know the internet is full of misinformation, disinformation, rumors, conspiracies, lies, and damn lies. When you do a Google search manually, if you are reasonably savvy and incredulous, you can often suss out the wheat from the chaff. You see exactly what's going into the sausage, so to speak. Google isn't making value* judgements on the results.
* Google does have algorithms which helps to rank pages. But that has more to do with page and site reputation, based on the rank of backlinks (referring pages) that link to the page in question or to other pages on the site. Not so much on whether a particular page is truthful or not.
But AI systems, like a sweet summer child, are more easily fooled. Unlike choosing from that screen full of Google search results, you don't have the opportunity to vet the data in the training model that contributes to the response.
However, there is way to get ChatGPT to show their work: After receiving a response to your prompt, reply with "cite sources" and ChatGPT will do just that. In most cases, it should return at least one relevant source supporting its previous response. Not as good as a screen full of raw links, but at least it gives you a place to start. It's better than nothing. I would strongly recommend doing with when prompting ChatGPT for hard facts.
Well, it is AI after all, shouldn't it be able to tell?
Sort of. We'll get more into this further down when we discuss what it means to be human.
But yes, to the extent that credible sources are saying one thing and non-credible sources are saying another, AI models that are paying attention will avoid scraping non-credible sites. Or, at least, labeling such collected data as less reliable, so that such data would be used in proper context. In a general knowledge retail chatbot like ChatGPT, the model is less likely to gather data from known unreliable websites.
That can help, but it doesn't mean AI still won't hallucinate.
When and How to use ChatGPT
You've noticed I've focused more on ChatGPT as a specific example of AI whereas AI itself is a much larger topic. Again, that's because ChatGPT is generally top of mind when people think about AI. Most people's useful exposure to AI is through ChatGPT or similar.
You'd be well advised not to rely on ChatGPT, or any AI chatbot, for authoritatively correct responses to anything. Just ask the several lawyers who got into trouble citing completely fabricated case law in some of their filings to various courts.
There's plenty of hilarious examples of people being professionally embarrassed or getting into real trouble because they relied on ChatGPT to provide accurate answers and didn't bother to verify them.
But that's not to say that ChatGPT can't be really useful. e.g. I use ChatGPT to help me write my articles on this website. But my use is limited to making some of my sentences easier to read or giving me some ideas on a topic that I'll further develop on my own. I'd never just copy and paste whole paragraphs. These are "soft" uses of ChatGPT where there's really no solid right or wrong. And I always verify any hard facts given to me.
Sometimes I'll ask ChatGPT for responses on a topic that I'm already familiar with, just to see how it does. It generally does well, but I've caught it out a few times, too. These are called hallucinations. When I've typed in a correction, ChatGPT immediately recognized its error.
n.b. ChatGPT does not update its training model based on your corrections. It may remember for you for next time, but that's all.
I also write quite a bit of computer code on my various projects. I've used ChatGPT to write code snippets for me as a time saver, especially in programming languages that I'm not fluent in. I often have to tweak the code, but it's considerably faster than developing it from scratch.
AI is everywhere
Now let's talk about the wider world of AI. Companies all across the business spectrum are rushing to introduce AI components into their products. I swear, half the sponsorship spots I hear on public radio these days includes company x, y, or z mentioning its AI.
Big Tech is forcing AI on us and often not letting us opt-out or disable it. Microsoft's CoPilot AI is pervasive inside of their products. Amazon's Rufus (really? Rufus?) was following me around like a lovelorn puppy until I killed it using one of my browser extensions. Google searches now often include an AI overview, further separating me from the webpages that I seek.
The familiar "chat with us" icon that we see on so many websites is now often AI-powered. Companies are starting to implement AI voice chatbots in their phone menu systems, making it even more difficult to reach a human. Terrific.
The medical profession is experimenting with AI to help diagnose disease, ailment, examine x rays, or other patient complaint. To the extent it's used to help surface certain conditions faster that a human doctor could then verify, that's not a bad thing.
But regardless of what form you encounter these AI assistants, be aware that all of them are subject to hallucinations, at least for now. This is pretty much why I don't trust nor wish to use AI-enhanced products in these disparate settings.
DeepSeek rocks the AI world
DeepSeek represents a newer approach to AI that promises to be far less costly and, potentially, more accurate.
DeepSeek uses a "Mixture of Experts" architecture as opposed to the "single expert" model used by ChatGPT. The MoE is a collection of smaller expert models that each focus on a specific range of topics. Also, by using SFT (Supervised Fine Tuning), certain AI model training data can be validated (labeled) by human reviewers, leading to fewer hallucinations.
It's not unlike having a room full of subject matter experts, each a master in their respective fields, instead of a single expert that "knows everything under the sun". You could think of it as a type of distributed processing.
When DeepSeek receives a prompt, it is gated (evaluated then dispatched) to one or more of these smaller experts to resolve.
DeekSeek R1 also utilizes reinforcement learning, a technique that can improve accuracy.
And, significantly, DeepSeek manages to do this using few, older, less expensive AI chips by nVidia. A massive cost savings.
This is important because it represents a democratization of AI, allowing less wealthy companies to launch their own AI initiatives. Up until now, AI research was the province of ultra wealthy tech firms, like Microsoft, Apple, Amazon, Facebook, and Elon Musk's various companies.
What makes a human?
For a thing without a single microchip, the human brain has a remarkable storage capacity. It has been estimated that a human brain can hold some 2.5 petabytes of data. That's equivalent to around 2,500 high-end iPhones.
Um, what is a "peta" byte?
Yeah, you've probably never seen the prefix "peta" before. Peta comes after tera, which you probably have heard of.
Let's examine the first few prefixes*, shall we?
1 byte = 1
1 kilobyte = 10^3 = 1,000
1 megabyte = 10^6 = 1,000,000
1 gigabyte = 10^9 = 1,000,000,000
1 terabyte = 10^12 = 1,000,000,000,000
1 petabyte = 10^15 = 1,000,000,000,000,000
human brain = 2,500,000,000,000,000
Each prefix is 1,000 times higher, three orders (^3) of magnitude, than the prefix that comes before. These numbers get really big, really fast.
* These are called SI units (Système International d'Unités).
But the human brain is more than just a huge storage reservoir. What separates us from AI are the qualities inherent and intrinsic to the biological and analog nature of our brains.
Humans possess innate capabilities such as intuition, consciousness, sentience, self-awareness, and emotions. We can also organically generate new knowledge through a distinctly biological process. Humans can truly understand words and meanings in a way that computers cannot.
In short, the human brain is far more than the sum of its stored knowledge and experiences. AI engines absolutely are not. Nor, do I believe they ever will be, even when quantum computing becomes mainstream.
Pizza Pizza
Here's a particularly fun example of an AI fail:
When prompted with "cheese not sticking to pizza", Google's search AI Overview feature suggested mixing in 1/8 cup of glue to give the sauce "more tackiness". (Google has since fixed that)
Without further contextual analysis, yeah, using glue on something to make it stick, might sound ok.
But there's so much wrong with adding glue to pizza sauce beyond simply the first obvious reason.
-
First and foremost, we don't use glue on food*
-
The moist nature of tomato sauce would prevent it from being glued in place
-
The high heat of cooking could ruin the glue
-
The glue would likely impart unpleasant flavors and texture
-
And certainly other reasons...
This is where a human brain would succeed where AI fails. Even if the idea of "glue" and "food" never came together in a person's mind, it would still be plainly obvious that doing so is probably not a good idea.
The additional bullets point out some reasons why it would not work. Any particular person may not know these additional facts. But just the idea of using glue on food is wrong enough, all by itself, to disqualify that as a useful suggestion.
* OK, foodies, I'm aware there are methods of "gluing" foods by using an egg mix, starches, etc. But that's not what we're discussing here.
AI Pushback
We're starting to see pushback building against AI as it becomes ubiquitous in everyday life. It's being sloppily shoehorned into every day consumer experiences without adequate thought given to implementation and the overall UX (User eXperience).
An an I.T. guy that works for small business and residential clients, I see this a lot. And I see it in my own life, too.
It's also pissing off website owners who are tired of their IP (Intellectual Property) begin scraped by AI crawlers. There's also a considerable cost to all those websites being scraped.
Cost to the websites? How so?
Visiting a website may be free to you, but not to the site owner. Aside from the cost of designing the website, there's a small, incremental cost to the site owner every time a page is fetched. Quite tiny, in fact, but not nothing.
Many site owners report the (sometimes vast) majority of page fetches are not regular people browsing, but instead are AI and search engine crawlers, gulping up everything they can find on a site. That can amount to real money cost to the site owner.
Webmasters can limit this by putting up a "no trespassing sign" (a robots.txt file) on their home page, specifying what sort of automated crawling they'll allow and by whom. But compliance is voluntary -- and routinely ignored.
Some site owners are resorting to guerrilla tactics to disincentivize this antisocial behavior by booby trapping their sites in ways that only affect AI crawlers and not regular people browsing.
Such "booby traps" include tar-pitting programs called Nepenthes* and another called Iocaine** as a countermeasure. Tar-pitting programs trap the AI crawler, sending them chasing down endless links with no way out, and returning non-sense, bogus data in an effort to poison the AI model.
* Nepenthes: A genus of carnivorous plants, also known as tropical pitcher plants, or monkey cups.
** Iocaine: A fictional deadly poison from the film The Princess Bride.
The Future of AI
AI isn't going away. It'll become ever more pervasive in our lives, just as streaming TVs, connected cars that track where you go and when, Wi-Fi enabled appliances with subscriptions, smartphones full of mostly stupid apps, and all the other tech we deal with that is fairly new to civilization.
Cheerleaders of AI, mostly tech companies funnily enough, say that AI will free people in the work place from having to perform boring, mundane tasks so they can focus on more creative endeavors. You don't often hear from AI proponents about the jobs that'll be displaced. But it's coming. Hell, it's already begun. How could it not?
Some of those proponents take it a step further, promising a sort of future utopia where many people won't have to work because AI will do it. Those people would be free to pursue whatever they enjoy. Problem is, people need money. They have to work. We don't have a Universal Income that would allow people to pursue non-occupational endeavors.
Help not Wanted?
But at least one CEO is being more forthright about his intentions for AI. Sebastian Siemiatkowski, the CEO of Swedish BNPL (buy now pay later) startup Klarna, readily admits that AI will allow him to operate his company with significantly fewer employees. If this bothers you, then you can choose to not use Klarna. But, alas, this kind of onesie-twosie boycotting isn't effective.
Klarna won't be the only one. As AI systems develop, especially to the extent that hallucinations can be significantly reduced, it will displace workers in a wide variety of companies. This is the ultimate reason companies are embracing AI!
Some types of jobs that could be eliminated by AI:
-
Data entry clerks, receptionists, transcriptionists
-
Call center agents, help desk, online chat CSRs
-
Cashiers, some sales associates, order clerks
-
Certain assembly line workers, machine operators
-
Truck drivers, delivery drivers, taxi/Uber drivers, if autonomous driving ever takes hold
-
News reporters and writers, copy editors
-
Paralegals, researchers, contract writing and reviewing
-
Bank tellers, loan officers, accountants, tax-preparers (heh, that last one wouldn't be so bad)
Most occupations where you sit at a desk using a computer could be subject to outsourcing to AI.
Here are some jobs that are less likely to be displaced by AI. Note that AI could play an assist role in some of these, but the job itself would probably be safe.
-
Human-centered, requiring emotional intelligence and empathy
Therapists, counselors, psychologists, psychiatrists
Social and outreach workers
Clergy and spiritual leaders -
Highly creative fields that require original thought
Writers, poets, artists, filmmakers, playwrights -
Skilled trades
Construction, electricians, plumbers, roofers, drywall installers, painters, auto mechanics, and others -
Unskilled, non-repetitive manual labor
Landscaping laborers (yard work, hauling materials, planting)
Construction assistants, go-getters -
Moving and transportation
Furniture, etc., junk removal
Package delivery (driving could become autonomous, but not package hand-off) -
First responders
Police, firefighters, paramedics, tow truck drivers -
Hospitality and entertainment
Restaurant workers, stagehands, roadies, festival workers -
Janitorial in residential and light commercial areas
-
Certainly lots more I'm not thinking of...
Essentially, any job that is physical, that has a hands-on component, and variable in nature should be safe. So, not including highly controlled, repetitive factory jobs, especially at larger companies that can afford the upfront costs of automation.
And, as mentioned above, jobs that require qualities that only a human brain possesses should be safe.

Stylized drawing of a human brain connected to a microchip