lynx--- via nettime-l on Wed, 8 Nov 2023 15:05:48 +0100 (CET)


[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]

Re: <nettime> It's Time to Fight for Open Source Again (fwd)


This manifesto in favor of open source AI contains some difficult
statements and non-trivial axiomatic assumptions that I'd like to
point out for discussion, taking the inability of civilization to
govern itself into account.

Daniel Jeffries <danieljeffries@substack.com> writes:
> Open source AI is under siege. 
> 
> A loosely knit group of AI opponents want to crush it completely. 
> They want to make sure you never get to download a powerful model.  If they
> win, you'll be forced to access models trapped behind a steel cage of
> corporate firewalls, the models safely castrated so you can't do anything
> they don't want you to do or think or say.
> 
> Instead of the open web, where anyone can publish anything without
> intermediaries, you'll have a filtered web where big models parrot official
> talking points only and keep you from "misinformation," a vaguely defined
> term that always means "whatever the folks in power don't like at the
> moment."

With the rising dominance of misinformation on the Internet thanks to
suitable infrastructures and unsuitable regulation, here's an
increasingly popular assumption that there is no such thing as
scientific facts. Well, I'd like to challenge that and say that, despite
some hindrances, the concept of 'scientific consensus' still exists and
there still is, in the majority of situations, ways to differentiate
facts from misinformation.

And I'd like to emphasize how misinformation is indeed becoming a
massive threat to the original intentions of the Internet, now that its
main role is the undermining of the population's ability to form
fact-based opinions, instead fostering radicalization of minds based on
false assumptions and data. The Internet is undermining democracy and
the open source model may have a role in that.

Instead we are seeing a presumed conspiracy of "the folks in power"
here. A conspiracy theory is legit only when the number of people
involved is reasonably small. All the folks in power cannot be in a
conspiracy, but they may have some interests in common. They may have a
common interest in fighting misinformation, and we should be glad for as
long as our leaderships are still majority-wise in favor of fact-based
information, because the Internet is happily bringing people into power
that do not share that value.

> You see, you're too stupid to be trusted with powerful models.  You might
> hurt yourself or others.  After all AI is just like a nuclear weapon they
> tell us.  It's too dangerous to let into the hands of mere peasants like
> you. 

My theory is that artificial intelligence is harmless if fed with only
ethically legitimate data. It's the collection of private and/or
democracy-destabilizing data which is threating what's left of
democratic civilization, with or without AI applied to it. That's where
regulation should be enacted, because the individuals who "grant" usage
of their data are factually neither competent nor legitimized to
relinquish personal data that in fact affects not only themselves but
their entire social neighborhood and the social structures of democracy
as a whole. GDPR has been misconceived (in best intentions as we all
know) and is acting as a legalization of the threats to democratic
society.

> To understand why, you just need to understand a little about where open
> source came from, where it is today and why it's the most important software
> in the world by a massive margin.

Just because the market power of open source was strong enough not to
pay for proprietary unices (but still not enough to crush Windows and
MacOS) and eliminate them doesn't imply that open source is per se
ethically better. Windows and MacOS engulfed some of the open-source
into their proprietary systems and Google even developed a worldwide
surveillance operating system on the shoulders of thousands of
developers who believed the naive idea of open source being
automatically for the good even if abandoned into the hands of big
proprietary players who can take 99% of it, put a 1%-sized proprietary
shell around it and effectively cut out humanity from the financial and
ethical surplus they created by writing open source technology in the
first place. If Android is a threat to humankind today, was the open
source ideology a good idea? How much longer until policymakers
understand that proprietary operating systems are a threat to human
civilization and its ability to address more important issues like
planetary survival?

> Open source AI will make the world safer and it will be one of the
> dominant forces for good in the world. 

That's the same style of simplification that led everyone in 1994 to
believe that Internet was going to inevitably improve democracy on a
worldwide scale. Open source hasn't made the world safer, it is thus
naive and unlikely that open source AI will automatically do so. It may
simply be the case, that it won't generate much ulterior damage if my
theory about data holds true. IMHO the lack of regulation of proprietary
systems is a bigger issue than AI - and it's an elephant that has been
standing in the room for decades and hardly anyone is seeing it.

> Frankly, it's bizarre that it's under attack.  I though the battle for open
> source was over forever.  How could it not be?  Open source won.  It won
> massively.

It's the same with capitalism. It won over socialist models because it
was economically more powerful to ruthlessly dismantle Earth's resources
and monetize on the inability of customers to make rational choices
about how they spend their money. And capitalism works anarchically, it
does not need any central coordination - that's how it won over
socialism, but how stupid were we in 1989 to believe that it won because
it was better? It is more efficient, but what's better on an ethical
scale is a whole different question.

It's like saying Caesar, Napoleon and Bismarck were good folks because
they won some wars, ignoring that they shouldn't have started the war
in the first place.

> Open source is the basis for 90% of the world's software today!  Read that
> again slowly.  90% of the world's software.

Because in our capitalist architecture of society you can't afford to
do things differently. Is that an achievement? Is that good? That's
an axiomatic assumption here.

> Linux runs every major cloud, almost every supercomputer on the planet, all

I see all the promoters of BSD deeply offended by you not mentioning
the large market share it has due to MacOS, haha.

> Open source levels the playing field in life.  It gives everyone the same
> building blocks.  You get to use the same software as mega-corporations with
> 10s of billions of dollars in revenue.  You get to use the same software as
> super powerful governments around the world.  So do charities, small
> businesses just getting started, universities, grade schools, hobbyists,
> tinkerers and more.

That's not always true. Half of my devices disallow me to put my own
free operating system onto them. I have been running Gentoo-based
laptops for over a decade. I installed Linux the first time in 1992. And
yet today I'm forced to use Android devices that thrive financially on
observing my digital habits, just because I'm socially forced to be
active on Instagram, which only works properly as an app. I could run an
Android emulator in a Linux, but for some obscure possibly political
reasons I still see that as a not very commonplace solution - also it
would not enable me to post stories on location from a mobile phone like
my social neighborhood does, allowing Meta and Alphabet to gather
everybody's whereabouts as they do. And solving the issue just for
myself doesn't address the societal problem of it.

> Go ahead and grab Linux for whatever project you dream up. Grab one of the
> 10s of millions of other open source projects for everything from running a
> website, to training an AI model, to running a blog, or to power a Ham radio
> You don't have to ask anyone's permission or pass a loyalty test.

That's how the Chinese government can implement a system for keeping
score of citizen's behaviors and despite it most likely (facts welcome)
being built on top of plenty of open source technology the citizen are
under totalitarian control and cannot escape from it. No reverse
engineering. No resetting their score by running their own instance. No
forking of the source code.

> With that kind of reach and usefulness I never saw it as even remotely
> possible that someone would see open source as a bad thing or something that
> must be stopped ever again.

Oh wow, the road to hell being paved by good intentions and convictions
here. Not even remotely being able to see the dark side of things. So
easily adamant to praise the simple solutions.

> Even Microsoft loves Linux now.

And Apple and Google and everyone who makes lots of money using code
that others spent years on.

> But I was wrong.  Here we are again.  The battle is not over.  It's starting
> anew.

Open source never won, because it can be integrated and proprietarized.
Back in the late '80s and early '90s there were plenty of free software
and source code releases that came with licenses that were forbidding
commercial or military abuse. The open source ideology has established
the term "non-free" for such ethical software. Shame on open source!
What a miserable brain-wash!

> They didn't listen to their own experts because too often governments
> fall prey to the fear mongering and doomsday scenarios and a hatred of Big
> Tech companies.

It's true that policymakers have utterly failed to regulate everything
digital from the beginning. The freedom of technology is not the freedom
of people affected or using it. By leaving technology unregulated it has
been allowed to evolve anarcho-capitalistically rather than in the
interest of humankind.

> This is a disaster for society, for innovation and for transparency. 
> Openness is the foundation of modern democratic societies.  Openness has led
> to some of the most powerful breakthroughs in history. 

Like the careless exhaustion of CO2 into the atmosphere and the
polluting of seas with plastic. It's the same disregard for the
long-term effects of allowing everyone to operate in anarchic-capitalist
openness. There are never simple solutions to complex problems. Calling
"openness" a panacea in every context is ideological, especially if it
means that everybody should be given the tools to harm the planet or
society. Maybe AI actually won't, but "open source" by its current legal
definition certainly does.

> When it comes to software, when something is know to a wide group of people
> it means anyone can try their hand at fixing problems with those systems.

Not in this case, because Google only lets you tinker with your Android
emulator (or expensive special hardware that only has a reduced number
of proprietary blobs in it), not with the billions of smartphones out there.

> You never know where innovation is going to come from but you do know that
> an open and level playing field is the best way to maximize the possibility
> that innovation will happen.

That I would agree with, but we are decades away from the last time we
had a level play-field and at least a century away from the times when
the biggest innovations weren't endangering sustainability of living
species on the planet.

> They’re well intentioned extremists.

Yes we are, and that includes you. Now we could start reasoning
collectively, but even then we are subject to a system of governance
which doesn't foster collective rationality so we are heading for doom
anyhow, with or without AI.

> Less is more when it comes to crafting tight, clear legislation that works.

Another simplification that doesn't scale up to the complexity of
reality. "Less is more" is a boomer phrase from the '80s that never
proved true, IMHO. Simplifying complexity with brute force usually leads
to even more problems.

"Simple" legislation like the radical approach I am suggesting could
work if there is the comprehension of urgency of taking such a radical
approach rather than trying to make every imaginable lobbyist happy. It
would take the ability to explain to all the stakeholders that they are
actually better off if the market is a level and fair playfield like
back in 1995, unless of course you are one of the few monopolists.

> What is the actual probability that AI will kill us all?  Trick question! 
> Any answer someone gives is just totally made up bullshit based on
> absolutely nothing more than someone's feelings. 

Indeed it is ridiculous to attribute such power to AI given that we have
already enacted all the anarcho-capitalism and lack of regulation on a
worldwide scale that is leading us to a 95% probable doom. And we're
taking all of fauna and flora with us - the result will be an
inhabitable planet for nearly all species.

The inability of public attention to focus and priotize the issues that
are indeed leading to the demise of human society is itself a big
problem. The chances of AI being just another technological hop speeding
us up on our insane trajectory may be of secondary relevance. At least I
value it less bad than blockchain currencies.

> If you believe in an AI apocalypse that's fine but you don't get to make the
> rules for the rest of us.

That's ideological on both sides of the fence. If they are convinced of
the AI apocalypse of course they need to fight against you being allowed
to use it and if you are convinced that open source AI is the answer it
could be considered disrespectful if you get to make the rules against
them just because of that. In theory we should have fact-based policy-
making, checking the argumentations of all extremists on all sides.
Alas, in practice...

> That's life.  It's a risk at every step and the beauty is in the risk. 
> Nothing worth having comes without risk.

That's how we undertook the path of industrialization that is now
threatening our existence on the planet. Some doomsayers back in the
1910s were worried about the long-term consequences of ejecting CO2
into the atmosphere. What if their opinion had seriously influenced
policymaking at the time? Industrialization could have taken place
at a path of sustainability, allowing a wealthier life for every
being on the planet.

We could have skipped globalization entirely, because it never made
sense to ship onions from China to Europe while the residues of burnt
heavy fuel oil is ejected into the atmosphere.

Throwing ourselves into risks just because the short-term advantages
are tempting, now that's short-sighted! Especially in a society that
is notoriously unlikely to agree on correcting the trajectory once the
gun is shot.

> The arguments for the AI apocalypse fall squarely into the Münchhausen
> trilemma.

Whenever people are convinced of things based on unproven or
fallacious axioms... being against AI is just as questionable as
being in undisputed favor of "open source".

> To fix problems we have to put AI into the real world.  That's how every
> technology works, throughout all time.

And then it becomes impossible to take it back on a worldwide scale
because if you regulate it in the EU or the US the remaining countries
will want to cash in on the capitalist advantage of not being affected
by such regulation. On a worldwide scale we are living in an anarcho-
capitalist system. Anarchic structures like the UN can regulate them-
selves sometimes, but only when all national leaderships agree. If it
isn't happening for the noble cause of saving our existence on the
planet, why should it work for much else? Ah right, we did save the
ozone layer. At least one thing where short-term financial and
political profit weren't luring enough.

It's amazing how wise we can be individually, yet utterly stupid on a
collective level.

> In this story in Fortune [ https://substack.com/redirect/4f85e73c-54dc-4a62-bb20-9127dfa7cf52?j=eyJ1IjoiN2N2MHMifQ.YtcUop6k1Xss2moIExwcy0mVj0mtto86Q9WaZCPRkdc
> ] , OpenAI said exactly that.  “[Their] biggest fear was that people would
> use GPT-3 to generate political disinformation. But that fear proved
> unfounded; instead, [their CTO, Mira Murati] says, the most prevalent
> malicious use was people churning out advertising spam."

When a new technology comes out, first it is the cool people that start
using it for disruptive yet ethically valid purposes. Then come the ones
that use it for porn. Next come those that use it for commercial gain,
which frequently boils down to spam. Only then come the ultimate smart
and macchiavellic ones, that figure out how to use it for political
manipulation. Alright, that's a theory of mine. But wouldn't you say it
applies to at least all social media innovations? Even makes sense for
e-mail and USEnet.

> You're probably reading this on something that's built on open source,
> whether that's your phone or your laptop.  It doesn't matter if your phone
> is an Android phone or an iPhone, there is some open source in there.  The
> kernel of both MacOS and iOS are open source and Android has a Linux kernel
> in it. Yes they are wrapped in proprietary code but that's the beauty of
> open source.  If you're a big, rich tech company, you can build flashy
> proprietary things on top of it.

The "beauty of open source" ? OMG, that's such a boomer thing to say.
Will it take a generational change for people to realize how wrong open
source has been by enabling proprietary technologies and keeping them in
a position of dominance? I think we made a mistake by fighting against
software patents, because I am pretty sure we would be living in a
better world if proprietary technologies were entirely illegal and all
the earnings for creative people came from patents on brilliant but free
algorithms that have been societally proven worthy of being patented
in a suitable way, not in an inhibitory way, but distributing profit to
all the ones that actually contributed to it. Our smartphones would be
truly free and everyone who contributed to the software stack running in
them would be earning a fair share for their pension. Why does it take
me to say such a thing and envision a better world of truly free
software? The Internet should be full of people seeing things this way.

> It's time to fight for open source again.

It would be a better world if proprietary technology and the open source
ideology that empowers it were outlawed. And then there are
possibilities to allow proprietary software to run in sandboxes like
Android already provides, but it is a major harm for human society that
the foundational operating systems aren't actually free - we are just
brainwashed to think that it is cool that they are based on software
that once was open.

Let's fight for free software that no longer subscribes to the open
source ideology. Let's regulate proprietary software. No piece of
hardware or software should be allowed to be proprietary if it is
involved in handling any data of essential importance to our society or
of constitutional important to our democracies. If it is treating a
private photograph or a personal message, nothing involved in doing so
must be proprietary. And in no way is it enough, that it once was "open
source".

Software patents offer enough ways to monetize software, we don't need
to allow living our lives off of food whose ingredients we are not
allowed to inspect.

> It's time to fight for open societies.  It's time to fight for open software
> and the free exchange of ideas.

It's time to fight for open software that cannot be used against us.

> Trust in open.  Never trust in closed. Nothing good ever comes from
> extremist polices throughout all history. When we let extremists come to
> power we get disaster.

Oh really? When safety belts in cars were introduced, most car drivers
yelled at the policymakers, calling them extremists. Take murder as an
example: is it open or closed to disallow murder by legislation? I
would say it is "closed", and yet it is foundational for civil society.
It allows me to walk in parks at night safely, at least in my country
with reasonable gun control.

> AI fear is just a con job, designed to fool you.

Another improbable conspiracy assumed here?
-- 
# distributed via <nettime>: no commercial use without permission
# <nettime> is a moderated mailing list for net criticism,
# collaborative text filtering and cultural politics of the nets
# more info: https://www.nettime.org
# contact: nettime-l-owner@lists.nettime.org