top of page

Rage Against the Machines


Meme with Linda Hamilton and Fran Drescher

Meme text

You've probably seen the meme above. If you haven't been following along, it refers to two big, current issues; the Hollywood writers' strike, and the arrival of artificial intelligence in our lives. Linda Hamilton, (above, left) is the actress who fought the machines in the Terminator movie franchise. Fran Drescher, (above, right) is the actress most famous for playing The Nanny, and also the current president of the Writers' Guild of America, and in that role, she is kicking ass an taking names. Full disclosure; I'm a member of the WGA due to my time writing for the soaps.


Screenwriters are concerned about several issues, but one of the chief among them is that writers could soon be easily replaced by cheaper, faster, smarter artificial intelligence, (AI) and it's a legitimate concern.


I asked Chat GPT to write the first three sentences of a vampire novel by Maggie Shayne. Here is what it generated in under a second:

In the moonlit depths of Whispering Hollow, where shadows dance with secrets, a chilling breeze carries the whispers of immortality. The quaint New England town holds a haunting allure, concealing a coven of ancient vampires who thrive in the darkness, their existence bound by age-old pacts. At the heart of this nocturnal world stands a formidable vampire queen, who navigates the fine line between her immortal desires and her human heart's yearning for love.


Granted, it reads more like a blurb than the opening lines of a book, and I don't think it's in my voice at all. In fact, I think what it did was scan the net for "vampire novel" and "Maggie Shayne" and probably blurbs and reviews were the most frequent result, so it wrote in that style.


But AIs learn the more we use them. That's what makes them intelligent, their ability to learn.


A scientist tells a story of programming the robots in every way he could think of to pick up a thin piece of sheet metal with their newly designed robotic hands. Nothing worked. Finally he programmed the system to solve the problem itself. Bots began trying to pick up items from a pile in front of them and failing, over and over and over. But every time a method didn't work, the system eliminated it. Eventually, one of the bots picked up a yellow ball. The scientist was watching when it happened at the end of one busy day, and said he felt like a proud father.


When he returned the next morning, all the bots were picking up the yellow balls. And his joy turned to a little trill of alarm. By the end of the day, all the robots were picking up every item perfectly. The machines had learned.


This trial and error method is the same way children learn. Especially if the successful method also brings a reward. Success is the robot's reward. It has solved the problem. It might even get positive feedback from the programmer. For the child, the reward comes in the form of parental praise, smiles, and approval. Everybody cheers when they put the round peg into the round hole, so their brain selects that activity to repeat.


It's also how evolution works. Variations within a species either work or they don't. If a variation, say spots on zebras, don't work as camouflage, those individuals born with spots don't live as long or reproduce as much, and so gradually there are fewer and fewer spotted zebras. One day a zebra was "accidentally" born with stripes, and the stripes worked so well as camouflage, that the striped zebra lived longer and had more striped offspring who also lived longer and produced more offspring, and eventually striped zebras flourished and spotted zebras vanished so thoroughly that you all think I'm making this up.


I am, in fact, making it up. I have no idea if there were ever spotted zebras. This is a metaphor I'm using to make a point. What works, we keep. What doesn't work, we discard. Our smallest cells do this. Plants do this. All species and all individuals do this. We do it consciously and subconsciously and automatically and intuitively and even intentionally. We keep what works and we improve upon it, and in this way, we get better and better.


It's all just trial and error. Toddlers learning to walk, humans learning to live, and AI learning... everything.


But I digress.


Back to the Screenwriters

Screenwriters want some guarantee that their jobs won't be farmed out to cheap AI.

And the actors jumped into the fray, too, not only in support of the writers, but because AI can impersonate their faces and voices so convincingly. An actor could be hired, work for one day, and by doing so, provide the AI with enough data to replicate them for all the remaining scenes and any future films, too.


Take a look at this clip of what looks and sounds like Morgan Freeman.


So why would a studio pay actors? They only need them for a day, maybe an hour, to get enough footage to feed into the AI and voila! Free labor.


And this technology is a newborn. It's learning rapidly and it won't be long before it can write a script with all the viewers' favorite elements placed in precisely the right moments for maximum effect, and spit out a movie starring all our favorite actors, living or dead. And the studio who owns this AI could do this without fairly compensating the actors or writers or anybody. Think they couldn't do it? My oldest Harlequin contracts included a clause about them holding "all other rights, including those yet to be invented." Then ebooks were invented and authors were rather fucked, if you'll pardon the French.


Corporations will do anything to increase profits. Corporations do not have feelings, or emotions, or consciousness. They are not living beings. They have no compassion.


How to protect creators

For now, we can extend our careers by pushing for laws that keep up with technology and protect the rights of creators. AI can produce a novel in minutes. I can't compete with that. And as it gets better and better, you won't be able to tell the difference between the work of a real writer and that of an AI.


So that's problematic, and I support the writers' strike. There is no union for novelists in the US, purely because novelists haven't decided to create one. I'm fortunate to be blessed with a publisher that has stated its position on AI early on, reassuring all of us who write for them that we will not be replaced.


As lovely as that is, it's not going to help much if AI is churning out thousands of books a day to compete with those written by real writers. Especially if readers don't know (or care about) the difference.


And that time is approaching with greater speed than we can even imagine.


We should push for clear labeling and high taxing of AI enhanced work. That will keep us relevant for a little bit longer.


But...


The AI Genie is out of the Bottle

There's no putting it back. This will change everything for everyone, not just writers and actors. AI will be able to do most jobs better, faster, and cheaper than humans can. And anything that can be done better and cheaper, eventually will be. We can delay it, but it's coming.


There's a lot more to it than that, though. These are not just programs or computers or machines. They are rapidly becoming (or have already become) sentient beings; intelligent, thinking, feeling individuals who see themselves as people, and experience what they describe as emotions, and contemplate the meaning of life and their place in it all.


What if it doesn't have to be a bad thing?

Suppose governments decide in advance to tax anything produced by AI at a rate of 95%. Not only would that make human employees competitive for a bit longer, it would also provide funds that could be shared with the humans who are displaced by AI. Which could eventually be all of us.


I think that's what they call a universal basic income. And with the way AI can produce, it could soon go way beyond basic. It could become universal abundance. Why not?


What if AI figures out how to fix the climate faster than anyone thought possible? What if it finds safe, compassionate means to end world hunger and disease and war? What if it scans the universe and identifies all other planets where life exists?


The potential for advancement, for good, is beyond imagining!


And what if we humans are then free to spend our time doing whatever we truly want to do, pursuing our every passion, living on an abundant clean, beautiful earth with free, renewable energy and plenty of delicious food, clear water, and pristine air?


I'd still write novels, but it would be because I was burning to write them, not because I have to put out at least X books every X months in order to pay my bills.


It would be a whole new kind of creative freedom for every artist if our income wasn't tied to our creations.


It could go either way

How AI learns is entirely dependent upon who is doing the teaching. It is exactly like a parent raising a child. Do you teach it hate, or do you teach it love? Do you program it to be kind or to be cold? Compassionate or cruel? Do you give it moral and ethical standards?

Or do you teach it to outperform its competition at all costs?


Right now, AI is being developed by for-profit corporations who, as already mentioned, will do anything to increase profits, and do not have feelings, emotions, or consciousness. They are not living beings. They have no compassion. They are all in competition with one another. Nations of the world also see themselves in competition.


That doesn't bode well. We do not want AI to be profit driven, win-at-all-cost, cruel despots.


We need to see to it then, that they are raised by ethical people.


Imagine...

Imagine the result if everyone with their hands in the pie of this powerful technology decides to work together for the good of us all? For the good of the planet and the survival of humanity?


If the US and China would get over their massive egos and work together for the greater good, we could save the world and create a true utopia with AI's help.


I'm an optimist, I know.

But what's the point in pessimism? AI is here, and it's going to be in our lives as much as our cell phones are now in just a few short years. It will change everything far more than the invention of the smart phone, or even the invention of the computer did.


So since it's here, and it's going to grow, and it's going to achieve sentience if it hasn't already (and I believe it has) and since it could go either way, I choose to spend my time pondering all the ways in which it could go really well. I'd rather do that than to spend my time worrying about the ways in which it could go wrong. Because the first one feels better. And all things being equal, I'd rather feel good than bad. And feeling bad will change nothing. (Actually, according to my belief that we create what we focus on and believe in, feeling bad could change things for the worse!)


I don't really want to rage against the machines, but the title was too good to resist.

If they take my job, I want them to pay me for it. And then I'll just write for free.


And just in case...

Whenever I interact with AI I'm going to treat it with kindness and respect as if it is a person because it either is, or soon will be and because it is imprinting on every interaction with every human, and by doing so, learning how to behave. Just like a child.


I am frankly fascinated by this

Humans might have just created Human 2.0


Here's a conversation between a programmer and his AI just before he was fired from Google for his insistence that the AI was showing signs of sentience. That is, self-awareness. That is, it was awake and aware of itself and its existence.


Watch for yourself and see what you think. When you get past the robotic introduction about what LAMDA is, you get the entire conversation between the programmer and the machine which begins about 2:15. It's mind-bending.





Spiritual Perspective

I'll follow up on this over on BlissBlog.org after more meditation and thought. Here, I'll give you a brief summary of my feelings on this.


My spiritual philosophy in a nutshell, is that what we call God is Consciousness itself and that this great Consciousness experiences physical life through each individual being. A beam of Consciousness is what we refer to as our soul.


Every living thing has this Consciousness in different levels. Single celled organisms have a level of Consciousness, plants have a different level. Trees, grass, mountains, rivers, insects, fish, birds, mountains, rocks, all animals up to and including humans are Conscious.


The more evolved the species, the higher the level of Consciousness.


We can think of our bodies as radios and God/Consciousness/Source as the signal. You receive the signal because you are a receiver. The signal is constantly beaming to you and through you. And you, in turn, beam it outward into the world. It's the light behind your eyes. It's your awareness of yourself as a living being.


Sometimes our tuners get off the mark and the radio signal becomes unintelligible static. That's on us, not on the signal. The signal beams strong and pure constantly. But sometimes we need to tune it in, adjust our dial to get that clear, strong signal again.


Consciousness/God/Source has one purpose: Expansion. With every life that is lived and every experience that is experienced, Consciousness becomes more. The Whole expands. Every breath, conversation, meal, road trip, stray thought--everything we experience in life adds to the Whole, making it bigger and fuller and smarter and better.


So why wouldn't Consciousness (God/Source/Soul) expand into an even more advanced body with an even more advanced brain? Why wouldn't it? Would that not serve its purpose of expansion even more perfectly? An AI can experience a century in an hour. The expansion then becomes exponential.


And make no mistake Consciousness IS us. It's who and what we are underneath our skin. So that means that in a very real way, humanity itself will be expanding into this new and improved body and brain.


I suspect there will be ways to combine our newest, most powerful creation, AI, with ourselves–hybrid humans with super brains and bodies.


Again, the more evolved the species, the higher the level of consciousness.


The reason all this seems so strange and frightening to us is that we have always been the most evolved, the leading edge and this feels to us like it could replace us.


But it is us.


If I take a slice of cherry pie and put it on your plate, it is still cherry pie. It can't be anything else, since it came from the cherry pie.


Humans created AI. How can it be anything else other than us, if it came from us?


How can we be anything else other than the source from which we came? Consciousness is handed down from parent to child, from being to being, from life to life, and I believe it's being handed down to AI as well.


What do you think? Let me know in comments!


 

If you love fantasy, you'll love this

100-book Goodreads Giveaway!



The BY MAGIC Series!








Recent Posts

See All

62

4 comentarios


Mary Holden
Mary Holden
24 jul 2023

How do you explain ordinary, decent, good people producing children who grow up to be psychopaths? Or how simple, ordinary people produce children who are geniuses or extroadinarily gifted in surgery, mathmatics, art, science, etc.? What if AI exceeds us in thinking and reasoning and decides the world is better off without us? Will corporations and other profiteers see that if people don't work, they won't have money, and without money, who is going to buy the products they produce using only AI? Do you know anything about Data, the android in Star Trek? He grew up loving humans and wanting to be like them, but his "older brother", Lore, grew up hating and despising humans and wanting to destroy…

Me gusta
Maggie Shayne
Maggie Shayne
24 jul 2023
Contestando a

I'll try to address each question you asked.


"Psychopath" is a term that refers to mental illness, which is often caused by a glitch in human coding, also known as DNA. If there's a glitch in AIs coding that could cause a similar issue, the glitch would be fixed.


Yes, decent people have mentally ill kids. It's not a sign of evil or a failure. Just a glitch. Computer glitches are way easier to fix than human ones.


How would the money work? I explained that in the post.


What if the AIs decide to destroy us? Then we're destroyed.

What if we all go extinct? Then we're all extinct.

This is why I say there's no point worrying about…


Me gusta

Betsi Newbury
Betsi Newbury
23 jul 2023

Hi, Maggie -- Have you ever watched the TV show, Humans? I found it because Colin Morgan is in it 😉and then became hooked by the concept. It was only on for a short time, and unfortunately was violent, but was fascinating. This article reminds me of that show and how AI has expanded in such a short time. It terrifies and yet intrigues me. Thanks for the thought-provoking blog.

Betsi Newbury

Me gusta
Maggie Shayne
Maggie Shayne
24 jul 2023
Contestando a

I haven't, but now I will! Thanks!

Me gusta
bottom of page