Smoke, Mirrors, and Fake News About AI


plasmaglyphs
 
Edited

Subject: Smoke, Mirrors, and Fake News About AI
Date: 2023-03-23 09:13
From: Gab <GabNews@...>

Smoke, Mirrors, and Fake News About AI

By: Fosco Marotto [1], CTO Gab.com

There used to be a meme where you'd start a sentence on a mobile phone
that would suggest words above the keyboard, and you keep picking the
middle suggestion until you completed a sentence. This was an admittedly
simple AI, generating text, predicting what the next word could be. We
are not much further along than this today, although the AI does have a
better vocabulary this time around.

While it can certainly be useful, the current generation of "AI" is not
remotely intelligent. It doesn't think, it doesn't remember anything,
and it doesn't have a personality. Much of the coverage of it is as fake
as the propaganda that led to the Iraq War.

The fear mongering around the capabilities of AI and potential dangers
is based on false assumptions and is done intentionally to gain power.
There's a lot of smoke and mirrors happening and the people in power
really don't want you to look behind the curtain.

THE ISSUE THEY HAVE WITH AI CREATING “DISINFORMATION” IS NO
DIFFERENT THAN THE ISSUE THEY HAVE WITH YOU DIRECTLY. No one is allowed
to share an opinion-or even a thought-that they don't agree with. That
is what drives this. It has nothing to do with the AI creating the text.


The leading "AI" conversation tools from "Big Tech" (ChatGPT, Sidney,
and Bard [2], etc.) are selling deception. The impression is that your
input is provided to "the one model," and then you're given the output
from "the one model." This is not how it works at all.

When you ask a language model a question, you aren't actually asking a
language model a question. What you prompt a Big-Tech bot "AI" with is
analyzed by many different models, to classify what or who it might be
about, what the emotional sentiment is, and more.

That information is used to construct a new prompt, with extra context
and formatting, which is fed into a language model to generate more text
at the end. This newly generated text is then taken, analyzed and
altered by yet more models, checking it and putting it into a
conversational response. Along the way, it could trip some sort of
"safety" detection and instead use more models to create a response
refusing to discuss it and admonish you for asking, or attempting to
re-educate you.

This complicated pipeline is responsible for the high quality of the
answers, yet also the main source of bias and censorship at play within
these systems.

Language models will generate text regardless of the input. You could
prompt a language model with just one word "The " and it will generate
as much text as you want, and without context, the results are
essentially random noise. You can prompt it repeatedly with a structured
setup like "Q: Do you prefer cats or dogs? A: ", and every time you ask
you'll get a different answer. From loving to hating either dogs or
cats, and anything in-between. It will also start asking its own
questions, because after-all it is only generating text. It has no idea
it's supposed to stop after one question, it doesn't know what a
question is!

"When you have a hammer, everything looks like a nail." This applies to
AI as everything is a "token" to an AI. All it does is decide the next
token in the sequence. Special logic is used to determine when to stop
generating more tokens.

The people in power in Silicon Valley want you to be scared of AI and
free thought, so they can push through laws and regulation about AI all
while consolidating more power to control you. They can only do this
through these straw man arguments. Their attempts will ultimately fail,
as regulating AI is like regulating math or programming in general. It's
stupid, and it's not possible.

IT’S STILL EARLY, BUT BASED AI IS INEVITABLE.

At Gab we're joining the large community of engineers who are working on
Alt-Tech solutions without guardrails and without a forced ideology.
We've already built an image generating AI, Gabby [3], and a movie
generating AI, Mel [4]. We are currently exploring different language
models for a text generating AI and in the process of doing this I came
to the realizations I mentioned above.

Whether the text is generated by a person or a language model, it is
just words anyway. You're still the one responsible for using or sharing
the text, whether you wrote it or not. You're technically also the one
who generated it, by creating the initial input and using the model.

Using AI as a tool is based. Delegating decisions to AI is cringe.

Don't fall for the smoke and mirrors.

Click here to view this email as a blog post and please share it! [5]

What's On Gab TV Today?

WATCH: Why isn't the media covering the mass police exodus across the US? [6]

WATCH: Globalism's Founding Myth [7]

Just For Fun: Throwback Gaming, Goldeneye [8]

Help Gab Keep Building The Parallel Economy and Home Of Free Speech Online

Go GabPRO Today! [9]
Grab a hat, shirt, stickers, and more on The Gab Shop [10]

-------------------------

Gab AI, Inc.
700 N. State Street
Clarks Summit, PA 18411

Gab.com Privacy Policy [13]

Links:
------
[1] https://gab.com/fosco
[2] https://summit.news/2023/03/22/googles-new-bard-ai-is-riddled-with-political-bias/
[3] https://gab.com/ai
[4] https://gab.com/movie
[5] https://news.gab.com/2023/03/smoke-mirrors-and-fake-news-about-ai/
[6] https://tv.gab.com/watch?v=641bb5c4ba1e2d232e8ceee6
[7] https://tv.gab.com/watch?v=6414ac5783f954eaaefdbee1
[8] https://tv.gab.com/watch?v=6417898fdc134cb9966d0879
[9] https://pro.gab.com
[10] https://shop.dissenter.com
[11] https://gab.com
----

-------- Original Message --------
Subject: Smoke, Mirrors, and Fake News About AI
Date: 2023-03-23 09:13
From: 🐸 Gab <GabNews@...>

Smoke, Mirrors, and Fake News About AI

By: Fosco Marotto [1], CTO Gab.com

There used to be a meme where you'd start a sentence on a mobile phone
that would suggest words above the keyboard, and you keep picking the
middle suggestion until you completed a sentence. This was an admittedly
simple AI, generating text, predicting what the next word could be. We
are not much further along than this today, although the AI does have a
better vocabulary this time around.

While it can certainly be useful, the current generation of "AI" is not
remotely intelligent. It doesn't think, it doesn't remember anything,
and it doesn't have a personality. Much of the coverage of it is as fake
as the propaganda that led to the Iraq War.

The fear mongering around the capabilities of AI and potential dangers
is based on false assumptions and is done intentionally to gain power.
There's a lot of smoke and mirrors happening and the people in power
really don't want you to look behind the curtain.

THE ISSUE THEY HAVE WITH AI CREATING “DISINFORMATION” IS NO
DIFFERENT THAN THE ISSUE THEY HAVE WITH YOU DIRECTLY. No one is allowed
to share an opinion-or even a thought-that they don't agree with. That
is what drives this. It has nothing to do with the AI creating the text.


The leading "AI" conversation tools from "Big Tech" (ChatGPT, Sidney,
and Bard [2], etc.) are selling deception. The impression is that your
input is provided to "the one model," and then you're given the output
from "the one model." This is not how it works at all.

When you ask a language model a question, you aren't actually asking a
language model a question. What you prompt a Big-Tech bot "AI" with is
analyzed by many different models, to classify what or who it might be
about, what the emotional sentiment is, and more.

That information is used to construct a new prompt, with extra context
and formatting, which is fed into a language model to generate more text
at the end. This newly generated text is then taken, analyzed and
altered by yet more models, checking it and putting it into a
conversational response. Along the way, it could trip some sort of
"safety" detection and instead use more models to create a response
refusing to discuss it and admonish you for asking, or attempting to
re-educate you.

This complicated pipeline is responsible for the high quality of the
answers, yet also the main source of bias and censorship at play within
these systems.

Language models will generate text regardless of the input. You could
prompt a language model with just one word "The " and it will generate
as much text as you want, and without context, the results are
essentially random noise. You can prompt it repeatedly with a structured
setup like "Q: Do you prefer cats or dogs? A: ", and every time you ask
you'll get a different answer. From loving to hating either dogs or
cats, and anything in-between. It will also start asking its own
questions, because after-all it is only generating text. It has no idea
it's supposed to stop after one question, it doesn't know what a
question is!

"When you have a hammer, everything looks like a nail." This applies to
AI as everything is a "token" to an AI. All it does is decide the next
token in the sequence. Special logic is used to determine when to stop
generating more tokens.

The people in power in Silicon Valley want you to be scared of AI and
free thought, so they can push through laws and regulation about AI all
while consolidating more power to control you. They can only do this
through these straw man arguments. Their attempts will ultimately fail,
as regulating AI is like regulating math or programming in general. It's
stupid, and it's not possible.

IT’S STILL EARLY, BUT BASED AI IS INEVITABLE.

At Gab we're joining the large community of engineers who are working on
Alt-Tech solutions without guardrails and without a forced ideology.
We've already built an image generating AI, Gabby [3], and a movie
generating AI, Mel [4]. We are currently exploring different language
models for a text generating AI and in the process of doing this I came
to the realizations I mentioned above.

Whether the text is generated by a person or a language model, it is
just words anyway. You're still the one responsible for using or sharing
the text, whether you wrote it or not. You're technically also the one
who generated it, by creating the initial input and using the model.

Using AI as a tool is based. Delegating decisions to AI is cringe.

Don't fall for the smoke and mirrors.

Click here to view this email as a blog post and please share it! [5]

What's On Gab TV Today?

[6]

WATCH: Why isn't the media covering the mass police exodus across the
US? [6]

[7]

WATCH: Globalism's Founding Myth [7]

[8]

Just For Fun: Throwback Gaming, Goldeneye [8]

Help Gab Keep Building The Parallel Economy and Home Of Free Speech
Online

Go GabPRO Today! [9]

[9]

Grab a hat, shirt, stickers, and more on The Gab Shop [10]

[10]

-------------------------




Gab AI, Inc.
700 N. State Street
Clarks Summit, PA 18411

Gab.com Privacy Policy [13]

Links:
------
[1] https://gab.com/fosco
[2] https://summit.news/2023/03/22/googles-new-bard-ai-is-riddled-with-political-bias/
[3] https://gab.com/ai
[4] https://gab.com/movie
[5] https://news.gab.com/2023/03/smoke-mirrors-and-fake-news-about-ai/
[6] https://tv.gab.com/watch?v=641bb5c4ba1e2d232e8ceee6
[7] https://tv.gab.com/watch?v=6414ac5783f954eaaefdbee1
[8] https://tv.gab.com/watch?v=6417898fdc134cb9966d0879
[9] https://pro.gab.com
[10] https://shop.dissenter.com
[11] https://gab.com


----