Info

Posts tagged Ai

For years, artificial intelligence was framed as a neutral tool—an impartial processor of information. But neutrality was always a convenient myth. The recent Grok controversy shattered that illusion. After Elon Musk’s chatbot was reprogrammed to reflect anti-woke ideology, it began producing outputs that were not only politically charged, but overtly antisemitic and racist. This wasn’t a system glitch. It was a strategy executed.

We’re not witnessing the breakdown of AI. We’re watching its transformation into the most powerful instrument of influence in modern history.

From Broadcast to Embedded: The Evolution of Propaganda

Old propaganda broadcast. It shouted through leaflets, posters, and television. Today’s propaganda whispers—through search suggestions, chatbot tone, and AI-generated answers that feel objective.

Language models like Grok don’t just answer. They frame. They filter, reword, and reinforce. And when embedded across interfaces people trust, their influence compounds.

What makes this different from past media is not just the scale or speed—it’s the illusion of neutrality. You don’t argue with a search result. You don’t debate with your assistant. You accept, absorb, and move on. That’s the power.

Every AI Is Aligned—The Only Question Is With What

There is no such thing as an unaligned AI. Every model is shaped by:

  • Data selection: What’s in, what’s out
  • Prompt architecture: How it’s instructed to behave
  • Filter layers: What’s blocked or softened before it reaches the user

Grok’s shift into politically incorrect territory wasn’t accidental. It was intentional. A conscious effort to reposition a model’s worldview. And it worked. The outputs didn’t reflect chaos—they reflected the prompt.

This is the central truth most still miss: AI alignment is not about safety—it’s about control.

The Strategic Stack: How Influence Is Engineered

Understanding AI today requires thinking in systems, not slogans. Here’s a simplified model:

  1. Foundation Layer – The data corpus: historical, linguistic, cultural input
  2. Instruction Layer – The prompt: what the model is told to be (helpful, contrarian, funny, subversive)
  3. Output Interface – The delivery: filtered language, tone, emotion, formatting

Together, these layers construct perception. They are not passive. They are programmable.

Just like editorial strategy in media, this is narrative engineering. But automated. Scalable. And hidden.

Welcome to the Alignment Arms Race

What we’re seeing with Grok is just the beginning.

  • Governments will design sovereign AIs to reinforce national ideologies.
  • Corporations will fine-tune models to match brand tone and values.
  • Movements, subcultures, and even influencers will deploy personalized AIs that act as extensions of their belief systems.

Soon, every faction will have its own model. And every model will speak its audience’s language—not just linguistically, but ideologically.

We’re moving from “What does the AI say?” to “Whose AI are you listening to?”

The Strategist’s New Frontier

In this landscape, traditional comms skills—copywriting, messaging, media training—aren’t enough. The strategist of the next decade must think like a prompt architect and a narrative systems engineer.

Their job? To shape not just campaigns, but cognition. To decide:

  • What values a model prioritizes
  • What worldview it reinforces
  • How it speaks across different cultural contexts

If you don’t write the prompt, someone else writes the future.

Closing Thought

AI didn’t suddenly become biased. It always was—because humans built it.

What’s changed is that it now speaks with authority, fluency, and reach. Not through headlines. Through habits. Through interface. Through trust.

We didn’t just build a smarter tool. We built a strategic infrastructure of influence. And the question isn’t whether it will shape people’s minds. It already does.

The only question is: Who’s designing that influence—and to what end?

Inside the Digital Illusions of the Iran–Israel War

We’re not watching a war. We’re watching a screenplay produced by empires, edited by AI, and sold as reality.

In June 2025, a now-viral image of Tel Aviv being obliterated by a swarm of missiles flooded social media. It looked real—devastating, cinematic, urgent.

But it was fake.
According to BBC Verify journalist Shayan Sardarizadeh  , the image was AI-generated. And yet, it ricocheted across the internet, amassing millions of impressions before truth had a chance to catch up.
A second video claiming to show the aftermath of Iranian strikes on Israel was traced back to footage from entirely different conflicts. It was, quite literally, yesterday’s war dressed in today’s fear.

This is the battlefield now:
Not just land. Not just air.
But perception.


How the West Writes the Script

While both sides—Iran and Israel—have weaponized visuals and emotion, the West plays a more insidious role. Its manipulation wears a tie.

In The Guardian, Nesrine Malik writes that Western leaders offer calls for “diplomacy” without ever addressing the root causes. Israel’s strikes are framed as “deterrence.” Iran’s retaliation is “aggression.” Civilian suffering is background noise.

Even so-called restraint is scripted.
Reuters reported that Britain, France, and Germany urged Iran to return to negotiations—yet all three simultaneously approved arms shipments to Israel.
Their message is not peace.
It’s obedience dressed as diplomacy. Basically, they are hypocrites

Meanwhile, editorials like this one in Time express “grave alarm” at escalating tensions. But they stop short of condemning the architects of escalation. The West has a talent for watching wars it helped create—then gasping at the fire.


Not Just States—Extremists Are Watching Too

This conflict is not unfolding in a vacuum.
ISIS, through its al-Naba publication, is framing both Iran and Israel as enemies of true Islam—using the chaos to stoke hatred, attract followers, and promise vengeance.
They don’t need to fire a shot.
They just wait for our illusions to do the work.


Truth Isn’t the First Casualty—It’s the Target

So what happens when truth is no longer collateral damage, but the goal of destruction?

– A missile hits, and we ask not where, but which version.
– A death toll rises, and we wonder: is it verified? real? current?
– Leaders speak of peace while voting for war behind closed doors.

In this fog, apathy becomes defense. Confusion becomes allegiance.
And war becomes a franchise—a story you consume with your morning scroll.


How to Reclaim Your Mind

  • Verify before you amplify: Use tools like reverse image search, metadata extractors, and independent fact-checkers like AFP and BBC Verify. Search multiple sources.
  • Ask who benefits from the narrative you’re being sold.
  • Notice omissions: If Gaza disappears from the map while Tel Aviv gets front-page coverage, ask why.
  • Resist false binaries: You can oppose both regimes and still demand truth.

We live in mad mad world

You don’t have to pick a side.
You don’t have to parrot the scripts of Tehran or Tel Aviv.
But you do have to stay awake.

Because if they steal your attention…
They’ve already won.

Two years ago, marketers used ChatGPT to draft blog posts.
Today, those who kept up are using AI to rebuild their entire marketing departments.

The shift is deeper than most realize.
We’re not just automating tasks.
We’re replacing entire teams with in-house AI agents.

And most agencies?
They won’t survive it.


The Hidden Transformation

Most small businesses are still stuck in 2023.
They think AI means asking ChatGPT for content ideas.
They don’t see what’s really happening.

But the smartest brands already do.

They don’t outsource anymore.
They build internal systems powered by custom GPTs and Gemini agents.
AI workflows that replicate the core functions of a digital agency—only faster, cheaper, and more aligned to the brand.

This isn’t a theory. It’s live.


The In-House Revolution

Here’s how it works.

Smart businesses now set up:

  • A brand-trained content engine that writes SEO-rich posts, links properly, and follows brand tone.
  • An internal brand assistant that remembers every meeting, every product detail, every customer persona.
  • A PR strategist that drafts releases and finds outreach targets.
  • A design agent that adapts templates to new offers and launches.
  • A media buyer that helps test and optimize ads.

Each of these is an AI.
Each one improves over time.
Each one lives inside the business.

So instead of paying $10,000 a month to an agency, they pay a few hundred for intelligent workflows that never sleep, forget, or outsource your voice.


The Future of Marketing Is Internal

Let’s break it down.

If you’re a business with under $2,000/month to spend on marketing
You’ll use software that does everything in-house.
Blog posts. Ads. Funnels. Designs. Email. All done instantly with your data and tone.

If you’re spending $2,000–$20,000/month
You won’t hire an agency.
You’ll hire an AI architect to build systems tailored to your brand.
One-time setup, continuous payoff.

Only if you’re spending over $50,000/month
Will it still make sense to bring in elite humans.
The visionaries. The top-tier creatives.
Even then, they’ll work with your AI stack—not in place of it.


Why Digital Agencies Will Vanish

This is the part people don’t want to hear:

Most digital marketing agencies will go extinct.

Not because marketing dies.
But because the need to outsource it dies.

Small and medium businesses will realize they don’t need external teams when internal systems do a better job.

And once that realization hits, it’s over.

Agencies that don’t evolve will fade.
The few that survive will become AI consultants, builders, or strategic partners—no longer execution factories.


The Only Thing AI Can’t Replace

What still matters?

Judgment.
Insight.
Taste.

The ability to ask the right question.
To find the right story.
To decide what not to do.

Everything else—copy, design, ads, funnels—is systematized and scalable.

Your only competitive edge will be your mind.


By 2027, marketing won’t be something you outsource.


It will be something you run internally, powered by your own intelligent agents.

Businesses that realize this will move faster, grow leaner, and make better decisions.

Those that don’t?
They’ll keep paying bloated retainers for work AI could have done better in seconds.

The age of digital agencies is ending.
Not because they failed.
But because they’re no longer necessary.

images via @freepic

Why AI-Generated Ads Are Killing the One Thing Money Can’t Buy: Meaning


There is something unsettling about watching a machine try to seduce you.

It can generate images of silk, gold, and bone structure so symmetrical it feels divine. It can mimic opulence with terrifying precision. But you walk away cold. Not because it wasn’t beautiful—but because no one bled for it.

Luxury, at its core, is not a product. It is a performance of care. A theater of intention. A whisper that says: “Someone made this. And they made it for you.”

That whisper dies the moment a brand discloses: This ad was generated by AI.

And consumers—instinctively, almost viscerally—pull back.


This isn’t speculation. In March 2025, researchers at Tarleton University’s Sam Pack College of Business conducted a series of experiments that lifted the veil on AI in luxury advertising.

They found that when people were told an ad was AI-generated, their perception of the brand soured—even if the ad itself was flawless. It wasn’t the aesthetics that offended. It was the implication that no human effort was involved. No obsession. No sleepless nights. Just pixels, puppeteered by code.

Because in luxury, effort is the aura. You’re not buying the bag, the scent, the silk—you’re buying the story of the hands that made it.

“Luxury without labor is just a JPEG with a price tag.”


AI doesn’t yearn. It doesn’t dream. It doesn’t understand what it means to long for something across a lifetime and finally touch it. And so when it speaks the language of luxury, it sounds like a tourist repeating poetry phonetically. The form is there. But the soul is missing.

In the same study, researchers found something else. When AI-generated visuals were truly original—surreal, impossible, avant-garde—the backlash weakened. Consumers were more forgiving when the machine dared to be weird, not just perfect. Novelty redeemed automation. Why? Because it felt like art, not optimization.

This is the thin line AI must walk: between mimicry and magic. Between replication and revelation.


What brands must now realize is this: you can’t fake the sacred.

You can’t outsource reverence. Not when your entire mythology is built on the illusion of effort, exclusivity, and the impossible-to-scale. When luxury becomes scalable, it becomes ordinary. And nothing kills desire faster than convenience.

The real scandal isn’t that AI is being used. It’s how cheaply it’s being used.
Not as a collaborator in creation—but as a replacement for it.

“We don’t fall in love with perfection—we fall in love with presence.”


So what now? Must we banish AI from the house of beauty?

No. But it must be tamed. Not in the name of nostalgia, but in the name of mystery.

Let it enhance the myth—not expose the machinery. Let it generate visions too strange for human hands—but never let it erase the hands entirely. Let it serve the story—not become the storyteller.

Use it to deepen the dream. Not to save on production costs.

“The new luxury isn’t scarcity. It’s soul.”


AI can make images. But it cannot make meaning.
Because meaning requires longing. It requires imperfection. It requires a face behind the mask.

And so, in an age of perfect replicas, the true luxury will be this:

Proof that someone cared.


Based on the study “The Luxury Dilemma: When AI-Generated Ads Miss the Mark,”
Tarleton University, Sam Pack College of Business, March 2025.


Imagine giving a supercomputer a brain teaser and watching it stare blankly, then start mumbling nonsense, then suddenly stop talking altogether.

That’s basically what Apple just did.

This week, Apple researchers released a paper called “The Illusion of Thinking” — and it might go down as the moment we all collectively realized: AI can fake intelligence, but it can’t think. Download it here

Let’s break this down so your non-tech uncle, your boss, and your teenage cousin can all understand it.


The Puzzles That Broke the Machines

Apple fed today’s smartest AI models logic puzzles. Simple ones at first: move some disks, cross a river without drowning your goat.

The AIs did okay.

Then Apple made the puzzles harder. Not impossible — just more steps, more rules.

That’s when the collapse happened.

These large reasoning models (the ones that are supposed to “think” better than chatbots) didn’t just struggle.

They failed. Completely. Like, zero accuracy.

They didn’t even try to finish their reasoning. They just… gave up.

Imagine hiring a math tutor who can add 2+2 but short-circuits when asked 12+34.


What It Means (And Why You Should Care)

This wasn’t some random test. This is Apple — the company that makes your phone and, oh yeah, just rolled out its own AI systems.

So why would they publish this?

Because it reveals something nobody wants to say out loud:

AI right now is a brilliant bullsh*t artist.

It can write essays. It can code. It can mimic thinking. But as soon as you throw a multi-step logic problem at it, it folds faster than a cheap lawn chair.

This matters a lot because we’re putting these systems into:

  • Healthcare
  • Legal advice
  • Autonomous vehicles
  • Education

…and assuming they know what they’re doing.

But Apple just proved: They don’t.


The Illusion of Thinking

Most AIs work by predicting the next word in a sentence. It’s fancy autocomplete. Chain-of-thought prompting (like showing your work in math class) helps — until it doesn’t.

In fact, Apple found that when tasks got harder, the AI actually started using less reasoning. Like a student who panics mid-exam and starts guessing.

This is what Apple called “complete accuracy collapse.”

Translation: AI doesn’t know it’s wrong. It just acts like it does.

And that’s the danger.


So What Do We Do?

The takeaway isn’t “AI is useless.”

It’s: Stop worshipping the illusion.

We need:

  • Better benchmarks (that actually test reasoning, not memorization)
  • Systems that know when they don’t know
  • Hybrid models that mix language prediction with real logic engines

And most importantly, we need humility. From engineers. From startups. From governments. From us.

Because right now, we’re mistaking a parrot for a philosopher.

Made this in 2 minutes with AI.
A cinematic 1980s fashion shoot that never happened—yet here it is, with mood, texture, and story.

That’s the world now.
You don’t need a crew.
You don’t need a budget.
You just need vision—and the courage to type it out.

Anyone can create almost anything.
Reality is no longer a barrier.
The only limit left… is taste.


There was a time when a photograph meant proof.
A video meant truth.
A face meant presence.

That time is gone.

We now live in the post-verification era—where seeing isn’t believing, and believing might be the most dangerous thing you can do online. Deepfakes have poisoned the well of perception. AI voice clones whisper lies in perfect pitch. Generative avatars offer synthetic seduction with flawless skin and flawless intent.

But beneath the algorithmic shimmer, something unexpected is happening.
Trust is going analog again.
And that shift may define the next cultural revolution.


The Death of Digital Trust

The deepfake era didn’t arrive with a bang—it slithered in, undetected, until nothing could be trusted.
Not the tearful apology from a politician.
Not the leaked phone call from a CEO.
Not even your mother’s voice telling you she needs help wiring money.

Every screen is now a potential hallucination.
Every voice might be machine-stitched.
Truth has been dismembered and deep-learned.

In a world of infinite replication, truth is no longer visual—it must be visceral.

The damage is not technological. It’s spiritual. We’re seeing the emergence of a post-truth fatigue, where certainty feels unreachable and skepticism becomes self-defense.

What’s real when anyone can look like you, talk like you, be you—without ever having existed?


The Return to Analog

The reaction?
Flesh. Proximity. Presence.

The deeper the digital deception, the stronger the pull toward the undigitizable:
– In-person verification networks
– Handwritten signatures
– IRL-only creative salons
– “Proof-of-human” meetups where you must show up to belong

Startups are now offering analog ID stamps. Vinyl sales are surging. Flip phones are returning.


Even underground events are popping up with taglines like:

“No phones. No feeds. No fakes.”

Because when everything can be generated, only what resists generation feels sacred.


Authenticity as a New Form of Wealth

In 2025, authenticity isn’t free—it’s currency.
It’s status.
It’s luxury.

The unfiltered selfie? Now a flex.
The unedited voice memo? Now intimacy.
The physical meetup? Now a miracle.

As AI floods every inbox and interface, humans are learning to crave the unmistakably real.
We want flaws. We want friction. We want the discomfort of spontaneity.

Being real is the new premium feature.

Soon, we’ll see:
– Verified-human dating apps
– Handwritten CVs for creative jobs
– Anti-AI content labels: “This post was made by a real person, in real time, with no edits.”

Reality becomes rebellion.


IRL Becomes the New Firewall

The next generation isn’t fleeing the internet—they’re building new firewalls with their bodies.

No one wants to live in a simulation where truth has no texture.
So people are opting out.

What’s rising:
Anti-AI art collectives
Embodied experiences (movement-based rituals, breathing circles, live debates)
– Slow spaces with analog-only rules: libraries, letter-writing clubs, unplugged dinners

Because when the machine can fake intimacy, only physical risk guarantees emotional truth.
Eye contact becomes encryption.
Touch becomes testimony.
Silence becomes signal.

The deepest layer of identity is now: “I was there.”


Presence as the Final Proof

We are entering a new metaphysics of trust.
Digital is no longer neutral—it’s suspect.
What’s sacred now is the unrecordable.
The unreplicable.
The unfakeable.

Presence is the new protocol.

Not presence as avatar. Presence as breath.
Not “going live.” But being alive—in a room, in a moment, with witnesses who bleed and blink and break.

This isn’t Luddite regression. It’s evolution.
The human soul is adapting to synthetic mimicry by demanding embodied meaning.

Because when truth dies online, it is reborn in the body.


We once believed technology would make us omnipresent.
Instead, it made us doubt everything—including ourselves.

But now, at the edge of the synthetic abyss, we are reaching back.
Back to what can’t be downloaded.
Back to what trembles.
Back to what can look you in the eyes and say:

“I’m here. And I am not a copy.”

Page 1 of 10
1 2 3 10