aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorXe Iaso <me@xeiaso.net>2024-01-17 20:55:38 -0500
committerXe Iaso <me@xeiaso.net>2024-01-17 20:55:38 -0500
commitf84bfbd2155a11163025bbe6b8b29733c0020ae8 (patch)
tree4d56b7e4f62e318232505cd8c1b98c7e58d14257
parent09507d34435ef566a1da86ec26291219e49aa4a5 (diff)
downloadxesite-f84bfbd2155a11163025bbe6b8b29733c0020ae8.tar.xz
xesite-f84bfbd2155a11163025bbe6b8b29733c0020ae8.zip
talks/2024: add AI ethics talk
Signed-off-by: Xe Iaso <me@xeiaso.net>
-rw-r--r--lume/src/talks/2024/_data.yml4
-rw-r--r--lume/src/talks/2024/prepare-unforeseen-consequences.mdx754
2 files changed, 758 insertions, 0 deletions
diff --git a/lume/src/talks/2024/_data.yml b/lume/src/talks/2024/_data.yml
new file mode 100644
index 0000000..de57dc0
--- /dev/null
+++ b/lume/src/talks/2024/_data.yml
@@ -0,0 +1,4 @@
+layout: talk.njk
+type: talk
+index: true
+year: 2024
diff --git a/lume/src/talks/2024/prepare-unforeseen-consequences.mdx b/lume/src/talks/2024/prepare-unforeseen-consequences.mdx
new file mode 100644
index 0000000..cdf1374
--- /dev/null
+++ b/lume/src/talks/2024/prepare-unforeseen-consequences.mdx
@@ -0,0 +1,754 @@
+---
+title: "AI: the not-so-good parts"
+date: 2025-01-16
+tags:
+ - ai
+ - ethics
+ - philosophy
+---
+
+Hey, if you normally read the written form of my talks, I highly
+suggest watching or listening to the video for this one. The topic I'm
+covering is something I'm quite passionate about and I don't think
+that my tone is conveyed in text the same way it is in voice. If the
+version on XeDN doesn't load for you for whatever reason, please
+[contact me](/contact/) with the output of
+[cdn.xeiaso.net/cgi-cdn/wtf](https://cdn.xeiaso.net/cgi-cdn/wtf) and I
+will figure out what is wrong so I can fix it.
+
+You can find the YouTube version of this talk [here](https://youtu.be/EfAjITmLP50?feature=shared).
+
+<XeblogVideo path="talks/2024/ai-ethics" />
+
+---
+
+<XeblogSlide name="2024/ai-ethics/001" essential />
+
+Hi, I'm Xe Iaso and before we get started, I want to start by talking
+about what this talk is and is not. This talk isn't going to be the
+kind of high signal AI research that I'd really love to be giving
+right now. This talk is about actions and consequences.
+
+<XeblogSlide name="2024/ai-ethics/002" essential />
+
+What impacts will our projects have on the real world where people
+have to take objects like this and exchange them for food and shelter?
+
+I'm sorry to say that this talk is going to be a bit of a wet blanket.
+I'm so sorry for Yacine because all that stuff with local AI inference
+in browsers was really cool. And that dogfooding of
+[dingboard](https://dingboard.com/) for a presentation about how
+dingboard works was cool as hell.
+
+<XeblogSlide name="2024/ai-ethics/004" essential />
+
+All the best things in life come with disclaimers, as I'm sure you
+know, and these words are my own. I'm not speaking on behalf of my
+employer, past employers, or if you're watching the recording and I've
+changed employers, any future employers. I am speaking for myself, not
+other people.
+
+Before we get into this, let's cover my background, some stuff about
+me, what I do, and all this AI stuff has benefited and harmed me
+personally. As Hai [the organizer of the AI meetup that asked me to
+speak there] mentioned, I'm a somewhat avid blogger. I've only got
+like 400 articles or something. I write for the love of writing and
+I've got like maybe four 3D printed save icons of text available on my
+blog for anyone to learn with any topic from like programming,
+spirituality, semiotics, AI, etc. My writing is loved by the developer
+community and it's the reason why I get hired.
+
+<XeblogSlide name="2024/ai-ethics/007" essential />
+
+Regardless of anything I say in this talk, please make a blog,
+document what you've learned, document what works, document what
+fails, just get out there and write. You'll get good at it, just keep
+at it. This is genuine advice.
+
+<XeblogSlide name="2024/ai-ethics/008" essential />
+
+However, as a reward for making my blog a high-quality thing, it's
+part of the ChatGPT training dataset. Somewhere in some data center,
+my blog's information is sitting there tokenized, waiting to get
+massaged into floating point weights by unfeeling automatons used to
+make unimaginable amounts of money that I will never see a penny of.
+This is the punishment I get for pouring the heart, soul and love into
+my craft as a blogger.
+
+I get turned into ChatGPT.
+
+<XeblogSlide name="2024/ai-ethics/009" essential />
+
+Now in our system of law, things are generally lawful unless there's
+some law or precedent that says it's not. At the time of me speaking
+this, we aren't sure if training AI models on copyrighted
+data is fair use or not. The courts and lawmakers need to battle this
+out (if they'll be allowed to because there is a lot of money behind
+the AI industry right now).
+
+This is technology that is so new, it's making Bitcoin look like Stone
+Age, 8-bit computing back when you couldn't count above 255 without
+major hacks.
+
+<XeblogSlide name="2024/ai-ethics/010" essential />
+
+And mind you, I'm just one blogger. I'm just one person. I don't have
+that big of a platform, all things considered. Sure in the genre of
+technology bloggers, I'm probably fairly high up there, but I'm not
+like front page on New York Times big. I'm just a person who likes
+talking about computers and how they should work. I'm just someone
+that gazed into the void too much and now people to pay me to gaze
+into the damn void.
+
+<XeblogSlide name="2024/ai-ethics/011" essential />
+
+So how do we understand all this?
+
+How do we figure out how to peel back all the layers of terminology
+bullshit that keep us from having a clear understanding of what people
+are even saying?
+
+If we take all the drama and interplay involved in our society, we can
+boil it down to two basic things, actions and consequences. Actions
+are the things that we do and consequences are the things that result.
+
+So let's say you cut a tree down to make a fire, but that tree was
+used by animals to shelter them from the winter and now those animals
+have a harder time finding shelter in the winter.
+
+You take actions and something or someone else has to deal with the
+consequences.
+
+Most of the time our actions serve to make us better off and shield us
+from the consequences. We see this happen with that tree that got cut
+down. We will see this happen with ChatGPT and we will keep seeing
+this happen time immemorial as society keeps repeating.
+
+As exciting as all of this AI technology is, as a science fiction
+writer, I can't help but see the same actions and consequences and
+analyses for how we're using it today.
+
+<XeblogSlide name="2024/ai-ethics/016" essential />
+
+Now your pitchforks can go down, I see you out there, you holding them
+up, I'm not trying to be a contrarian or decry AI as wrongthink. I've
+been using AI for my own stuff and I genuinely think that there's a
+lot of really exciting things here.
+
+I'm mostly worried about how the existing powers that be are going to
+use this surplus of cheap labor and have those actions have massive
+consequences on us all.
+
+<XeblogSlide name="2024/ai-ethics/017" essential />
+
+One of the things I'm trying to get across here is not all "Capitalism
+bad! Let's get back the bread lines, baby!" There's plenty of places
+to see those arguments and I don't want this to be one of those. I
+more want to inspire you to see what the consequences of your actions
+with AI stuff could be so that we can make the world a more equitable
+place.
+
+Of course, this is made even more fun by the concept of unforeseen
+consequences or downstream consequences that you couldn't have
+possibly seen coming when you were experimenting with things.
+
+<XeblogSlide name="2024/ai-ethics/018" essential />
+
+As an example, for a long time people thought swans were white. Swans
+became symbols of literary purity or something like that and it was so
+common that there was an English idiom of a black swan being an
+impossible thing.
+
+As this photo proves, swans can be black.
+
+And now the term "black swan event" describes something that should
+have been obvious in hindsight but something that we couldn't possibly
+have foreseen at the time.
+
+(Begin sarcastic tone)
+
+Just like that unmentionable-on-YouTube viral pandemic that happened a
+few years ago that our society will never really recover from!
+Scientists were warning us for years that we'd be totally screwed by a
+viral pandemic but no, we didn't take them seriously.
+
+(End sarcastic tone)
+
+<XeblogSlide name="2024/ai-ethics/020" essential />
+
+Whenever anyone takes actions and there are consequences or impacts,
+you can usually model them as on yourself, your friends or the world
+at large. I haven't found a good way to model the impact risk of a
+given field very well, but I like triangles so I made this triangle
+called the impact triangle to show what all of the factors in the
+computer science industry are.
+
+In terms of access, anybody can become good at coding and start
+working at a company or creating a company to solve a problem that
+they have in their lives. I'm pretty sure that this basic thing, the
+computer industry is open access to anybody is basically why everybody
+in this room is here today.
+
+Personally, I'm a college dropout.
+
+Without the industry allowing just about anyone to walk in the door
+and start being successful, yeah, I'd still be in the Seattle area
+probably working minimum wage at a fast food place. I wouldn't be able
+to dream of immigrating to Canada and I probably would have never met
+my husband who is so thankfully recording this for me.
+
+There's also no professional certification or license required to
+practice computer science or software development or whatever we call
+ourselves now. And basically anybody off the street without
+certification can make an impact on the world scale if they get lucky.
+
+And then in terms of limits, our industry measures results in small
+units of times like individual financial quarters. In aggregate, our
+industry only cares about what we do to make the capitalism line go up
+for next quarter and there's no ethical or professional guidelines
+that prevent people from making bad things or even defining what good
+and bad is in the first place. In an ideal world, the thought is that
+the market should sort everything out and realistically, with the GDPR
+and the like, there are some laws that enable, that force people to
+comply but as long as you have good lawyers, you can get away with
+killing murder.
+
+For most other professions in the job market, our industry looks
+incredibly reckless. Like, accountants need to be licensed and pass
+certifications. If you want to call yourself a surgeon, you need to
+have surgery practice, you need to have a license in surgery, and you
+need to keep yourself up with the profession.
+
+We don't have such barriers to entry.
+
+As an example of this, consider Facebook. They have a billion users.
+That is nine significant figures, a billion with a B as in bat. When
+they made Facebook, the thought was that they could make everybody
+better by reducing the social distance and that could make everybody
+like happier and live more fulfilled lives.
+
+An unimaginable amount of photos, video and text posts are made to
+Facebook every day. Some measurable fraction of these violate
+Facebook's community guidelines and are full at the very least and are
+fully legal at the most. Many trivial cases can be handled by machine
+learning algorithms but there's always that bit that needs to be
+judged by a human.
+
+Speaking as a recovering IRC op, content moderation is impossible at
+small scales and the level of impossibility only grows as the number
+of people involved in a thing grows. I am fairly certain that it is
+like actually entirely impossible to moderate Facebook at this point
+because there's just too many people. You have to have some machine
+algorithm in there at some point and there are going to be things that
+the algorithm can't handle.
+
+So then you go and you use humans to rate that.
+
+You contract out a company who very wisely decides to subcontract that
+out because they don't have to deal with the fallout and finally it
+ends up on the desks of people that are tortured day and night by the
+things they are forced to witness to make rent.
+
+For the action of creating Facebook and all of the systems that let
+Mark Zuckerberg make a bunker on Hawaii, raise his own cattle, make
+his own beer, and smoke those meats, he doesn't have to see those
+images and things that the content moderators have to see.
+
+He just lays back and watches his bank account number go up and maybe
+does CEO things if he has to.
+
+The human cost is totally discounted from the equation because the
+only limit is what makes the capitalism line go up. The people doing
+the actions almost never see the consequences because the CEO of Uber
+never got his job replaced by an Uber driver. The CEO of Google never
+suffered the algorithm locking him out of his entire digital life for
+good with no way to get it all back. And the people doing the actions
+and making the decisions are not affected by any of the consequences,
+foreseen or unforeseen.
+
+The last time I spoke here, I spoke about a work of satire called
+[Automuse](/videos/2023/ai-hackathon/). Automuse is a tool that uses
+large language models to recreate the normal novel writing process
+using large language models and a good dose of stochastic randomness
+to make some amusing outputs.
+
+When I made it, I really just wanted to throw ink to the canvas to see
+what would happen, then write [a satirical scientific
+paper](https://cdn.xeiaso.net/file/christine-static/video/2023/ai-hackathon/automuse-2.pdf).
+
+<XeblogSlide name="2024/ai-ethics/031" essential />
+
+To my horror, I won the hackathon with a shitpost about the publishing
+industry that was inspired my fear of what could happen if things like
+Automuse were more widespread.
+
+When I gave my talk at the hackathon, I had a five minute slot and
+there was something that I had on my script that I cut out as I was
+speaking.
+
+Not sure why I did, it just felt right at the time.
+
+The part that I left out was inspired by this quote from the
+philosopher SammyClassicSonicFan:
+
+<XeblogSlide name="2024/ai-ethics/033" essential />
+
+When will you **learn**?
+When will you learn that your **actions** have **consequences**?
+
+I made Automuse precisely because I understand how impractical such a
+thing is. The output quality of Automuse will never compare to what a
+human can write no matter what large language model you throw at it.
+
+Okay, yes. I did my research, there's actually a rather large market
+for low quality pleasure reading that something like Automuse could
+fill. There's a surprisingly large number of people that enjoy reading
+formulaic things about good winning out over evil or old people
+reading romance novels to feel the passion of being young again or
+whatever. Not to mention doing something like that as a company would
+leave me an excellent moat because most AI companies want to focus on
+the high quality super output and here I am, the trash vendor going
+in, yeah, I'd basically be invincible.
+
+But I don't know if I could live with myself if I turned Automuse
+into a product.
+
+When I made Automuse, I knew that this was a potentially high impact
+thing, so I crippled it.
+
+I made it difficult for anyone to use, even me.
+
+I made it rely on a private NPM dependency that is on a server that
+only I have the API token to and it just so happens to be the thing
+that generates random plots.
+
+I also made it in a way that requires massive human intervention and
+filtering in order to get decent results and every so often I get a
+message from somebody that asks me:
+
+<BlockQuote>Hey, how can I set up Automuse on my stuff?</BlockQuote>
+
+And they're surprised when I quote them a five figure number to get
+them to go away. And some are even angry and curse me out because a
+person making open source software on the internet would want to be
+paid for their time.
+
+I can't understand that actually.
+
+But above all, the reason why I really don't want to productize it or
+make it available for mass consumption in any form is the problem of
+book spam. Automuse would make the problem of book spam worse.
+
+The Book Spam problem is where people upload nonsense to the Kindle
+store and make boatloads of money doing it. This problem has been
+accelerated by ChatGPT and is getting to the point where Amazon's book
+vending thing actually had to implement rate limits for uploading
+books.
+
+I don't think I could live with myself if I made and released an easy
+to use product that made that problem worse.
+
+It's bad enough that whenever I get around to finishing my novel
+Spellblade (I couldn't find the cover I commissioned, so I just put
+the name on the slide), I'm almost certainly just going to release it
+on itch.io or to my patrons for very cheap. In theory, the Kindle
+store would be the best place for that kind of high signal original
+fiction but I just don't want it to get flooded out in a wave of AI
+generated mushroom foraging books.
+
+I don't think that anyone at OpenAI anticipated that people would use
+ChatGPT to make the book spam problem worse. I have a friend that
+works there and generally from what I've seen, the research side of
+OpenAI really has their head screwed on the right way.
+
+The problem is the capitalism side of OpenAI getting that sweet, sweet
+return to an investment by making a product that nobody else can
+provide and then charging for the output.
+
+<XeblogSlide name="2024/ai-ethics/039" essential />
+
+Above all, the part that really confuses me is why we're automating
+away art and writing instead of like snow blowing or something
+actually useful. There's a subtle part of me that's really concerned
+for the future of our industry and I really think we need to be aware
+of it before it all bites us and like getting rid of everybody that
+has aesthetic knowledge really seems like a bad idea for an industry
+that focuses so much on design.
+
+<XeblogSlide name="2024/ai-ethics/040" essential />
+
+With the Industrial Revolution came factories. Factories allowed us to
+produce objects on scales like never before. Raw materials go in at
+one end, human labor goes in the middle, finished products come out
+the end. This has allowed us to become the kind of species we are
+today. You can circumnavigate the globe in 100 hours while playing a
+contrived game show about travel. You can head to an entirely
+different continent in like what, 12 hours and this has led us to
+discoveries that have made us healthier, lived longer lives and
+overall it's been a boon for the human race.
+
+<XeblogSlide name="2024/ai-ethics/041" essential />
+
+However, this is a modern assembly line for cars. Look what you don't
+see here, people. All of those robot arms and the like represent jobs
+that were done by humans, operating the crane to lower the truck body
+onto the chassis, all of that stuff. With every new model year there's
+more automation at play and less room for human jobs.
+
+Sure, we can make more cars per hour but like every job that's not
+done by a human is another family that can't make rent. It's another
+child that can't grow up and you know actually cure cancer or
+something. And I just feel like it's another way for the ownership
+class to scrape more off the top.
+
+With that in mind, I want you to consider this:
+
+<XeblogSlide name="2024/ai-ethics/042" essential />
+
+These are our factories, the open office environment. Instead of wool
+or wood or water as input, we have user stories, electricity and
+coffee. Many of the companies out there are really just assembly lines
+for code features or Kubernetes configurations. I think the ultimate
+dream of this lies in the idea of the T-shaped developer that I've
+seen many management people talk about when they're trying to
+reorganize their companies.
+
+<XeblogSlide name="2024/ai-ethics/043" essential />
+
+The core idea of the T-shaped developer is that you have really good
+competency in one field and enough broad knowledge in other fields
+that you can basically be put anywhere in a project and be useful.
+This is why you see things like ephemeral teams or decrees from on
+high that thou must write in JavaScript for all things.
+
+And in theory, it makes it a lot easier to move people around and
+place them wherever the company needs in order to make the process
+more adaptable to the circumstances. Not to mention, if everyone's
+just a T-shaped developer, that makes it really easy to get people off
+of the street and into the job in days so you don't have to spend the
+months training them on how you messed up Jenkins this time.
+
+Ever notice that every job opportunity is only for senior roles?
+
+This is why.
+
+Usually by the time you convince companies to give you a title that
+starts with the word "Senior", you've already been molded into a
+T-shaped engineer and you can slot in just about anywhere.
+
+This is our assembly line, created in the fear that if we don't do
+this, the wrong line will trend in the wrong way and investors won't
+give us as much money as freely.
+
+Like, okay, I realize I'm doing some doom and gloom stuff here.
+
+It's probably going to be a while until AI is actually able to replace
+our jobs. Right now, there isn't a magic button that product teams can
+use to "just implement that feature" based on the textual description.
+That's probably a long ways off and it'll probably require a different
+fundamental architecture than attention window transformer models.
+
+But with that in mind, there's a segment of people that already have
+the magic "just implement it" button today:
+
+Artists.
+
+Stable diffusion, mid-journey, and Dall-E 3 have gotten to the point
+where the output is not just good.
+
+It's good enough.
+
+For the vast majority of people, as long as there's nothing obviously
+wrong with the hands, you won't be able to tell an image that is AI
+generated.
+
+However, artists can tell instantly when you have an AI generated
+illustration.
+
+<XeblogSlide name="2024/ai-ethics/049" essential />
+
+Just look at this one I used earlier in this talk. It's so bad. Look
+at the stem on that flower. That is not how stems work. The brush at
+the bottom is just blending into the easel in ways that physically
+separate objects don't work. The flower that the robot is holding is
+inconsistent. It looks like the light is coming from both forward and
+backward at the same time. The antennae are melting into the shoulders
+of the robot.
+
+It's totally passable at first glance.
+
+I'm pretty sure that before I mentioned all those stuff and put all
+the arrows on the slide, you wouldn't have seen any of it. But when
+you start critically analyzing it, it just falls to pieces.
+
+I guess the better question here is why would you want to use an AI
+generated image for something?
+
+One of the big places you want to use an AI image is for the cover
+image on your blog post because we've come to expect that blog posts
+need cover images for some reason.
+
+There's more desire for people to have cheap filler art that meets a
+certain criteria than there are artists willing to work for
+unrealistically low prices with incredibly quick turnaround times. Art
+is everywhere and yet it's commoditized so much that it's worthless in
+a day and age where rent and food prices keep going up.
+
+So we end up with something like this:
+
+<XeblogSlide name="2024/ai-ethics/051" essential />
+
+You get an AI generated of assembly line of robots painting flowers.
+
+This is really why I didn't want to develop Automuse into a company. I
+just fear that action would have too many consequences and my friends
+and fellow artists would suffer. This is why I did so much detailed
+math about how much it would cost per word, how the quality would be
+seen in the market, and what impact such a technology would have if it
+churned out hundreds of books per hour.
+
+Outside of the systems we live in, yeah, this AI stuff is great. It's
+fantastic tech that allows us to do any number of things we couldn't
+do before.
+
+But inside the systems we live in, I can't say the help, let's see
+this is yet another way that human labor is being displaced without a
+good replacement.
+
+And we wonder why we can't call ourselves engineers in Ontario. Do we
+really engineer anything or are we just making the line go up?
+
+When will we learn that our actions have consequences?
+
+Until then I guess we need to prepare for unforeseen consequences.
+
+Thank you all for watching this and I hope it gives you some things to
+think about. I hope I didn't break too many taboos about the industry
+in the process but who am I kidding? I just broke all of them.
+
+<XeblogSlide name="2024/ai-ethics/061" essential />
+
+Thanks to everyone on this list for inspiring me to take action and
+pushing towards the presentation I gave tonight. Special thanks to
+Mystes and Layl for really grinding hard into this, ripping in half
+and telling me where I'm full of shit. Extra special thanks to my
+husband for recording this for me and thank you for watching.
+
+<XeblogSlide name="2024/ai-ethics/062" essential />
+
+I recognize that this is like really a heavy talk. It'll probably take
+you some time to surface some good questions about it but if you
+happen to have them right now please feel free to ask. I will be happy
+to answer but if it takes you a while to come up with it just email
+[unforeseenconsequences@xeserv.us](mailto:unforeseenconsequences@xeserv.us).
+It'll get to my inbox and I promise you I will reply. Have a good
+evening and does anyone have any questions?
+
+## Q&A
+
+<BlockQuote>What was the sigil you displayed at the beginning of your
+talk?</BlockQuote>
+
+That was the sigil of Baphomet, one of the names for Satan as
+celebrated in Satanism.
+
+<BlockQuote>Do you see a future where AI technology can equitably help
+humanity thrive?</BlockQuote>
+
+I do see a future where it can be used to benefit us all. The problem
+is the intersection of what could be, what is, and the tools in the
+process where you get the real interesting stuff and there's probably
+at least five good sci-fi novels you could write about this.
+
+You could write a really compelling one about just what happened with
+OpenAI and especially what's happened with the e/acc people. I wrote
+the plot outline for a bad science fiction novel about the madness
+that is e/acc.
+
+<BlockQuote>What do you think we should do about this
+problem?</BlockQuote>
+
+Just be aware that your actions don't exist in a vacuum.
+
+If you build something that could replace jobs, then you need to be
+cognizant of the people that you're going to make unable to pay rent
+because if you make something that replaces knowledge work labor, you
+price them out of being able to eat. When people aren't able to afford
+to eat, they especially can't afford to retrain themselves to work in
+another industry that hasn't been taken over by infinite cheap labor.
+
+<BlockQuote>
+First, thank you very much for the presentation. I'm not debating
+here. I'm very open for these type of discussions, but you show the
+industrial revolution and the next slide was all the people who were
+impoverished. I don't see it as a linear change though, so industrial
+revolution and all those workers working in those situations by itself
+was not a necessary, better situation than those workers in those
+dangerous situations being replaced by robots on the other side. As we
+move on, we never had any occasions that we needed to get rid of a
+bunch of populations because we didn't have jobs for them, but we
+eventually came up with solutions, new jobs, some sort of a solution.
+So the main question is how do you see that change exactly from
+industrial revolution to industrial revolution?
+</BlockQuote>
+
+At some level, this stuff is going to happen regardless, and if it's
+going to happen, there should be some societal support mechanism, like
+universal basic income (which no matter what study is made to prove it
+doesn't work, actually does work) to replace the income that we're
+losing to machines taking over jobs that were previously done by
+humans. Something like universal basic income would probably help a
+lot here, but I don't know.
+
+I don't have any solutions.
+
+I'm more trying to blow the whistle that there's a problem before it
+gets bad enough that things become irreparable.
+
+<BlockQuote>
+All right, I'd like to commend you first on your courage to do this.
+It's obviously difficult to come into a room and say the opposite. At
+the same time, I'll give you the opposite and the pit that was out of
+the pit. You know, one of the things that, to act your way a little
+bit, automation is known to increase the standard of living. So we
+have all great things we can do because of automation. So AI is
+automation's superpower. Now to say there's no consequences of AI
+being abused, there definitely will be, but looking at the greater
+impact of it all, and I think that's the reason we're at all here, is
+because we know that they're [unintelligible], but truly down, we know
+that bringing abundance to the world is far greater and needs to be
+substantial in that event.
+</BlockQuote>
+
+I mean, yes, congratulations. You actually got the point of the talk.
+The point of the talk is to get you to think critically about what
+these tools are, what's going on, and what the benefits could be as
+well as what the downsides could be. I just don't know if our current
+system of distributing wealth and resources is really going to be able
+to adapt to that in time without some major cataclysm forcing the
+measure.
+
+<BlockQuote>
+I just wanted to ask you.
+You said you're not sure if this system of wealth distribution is the
+right system that should be, you know, that should have this kind of
+AI in place for moving forward. So what kind of system do you think is
+more practical for that?
+</BlockQuote>
+
+
+So I think one of the more ideal outcomes would be if people that
+whose work is in the training set of ChatGPT end up getting royalties
+from OpenAI for their data being used to make unimaginable amounts of
+money.
+
+Like, I have been transformed into ChatGPT. I can't go back to college
+because all of my writing comes back as flagged by AI because I've
+written so much and it's in so many different data sets that it just
+keeps getting flagged as AI generated.
+
+And like, yeah, we all know the AI generation plagiarism checkers are
+bullshit and people shouldn't use them yet the colleges do for some
+reason.
+
+So like, what can you do?
+
+Really the best possible way to get equity here would be to basically
+make it so that if you research AI with copyrighted materials, that's
+fine. But when it comes to putting the money generator in the mix,
+hold up, maybe you actually need to pay royalties because those blog
+posts and the like, they don't just come out of nowhere for free.
+Like, you know, you have to train to be an artist. Like, this photo of
+this log that I got off of Pexels, a public domain image stock image
+sit, you have to have some like skill in photography to know the rule
+of thirds and you know, like be able to configure your camera to
+capture the exact moment of this log falling like this. There are
+actual skills that don't look like skills that still require a lot of
+time, energy, and frankly, remuneration to compensate for.
+
+I think one of the best ways would be to make the concept of an open
+source model that is only just the weights without any of the training
+data or any of the training methodology involved an unterm.
+
+Like, that is not open source, that is open access. Open source would
+be providing all of the code you used for training, all of the data
+that you used for training, and a summary of where you got the data
+from.
+
+That would be closer to what open source actually is and anything
+close to the definition of open source back when the GPL was the
+dominant definition of open source.
+
+Generally open source AI stuff is really cool. There's a lot of stuff
+you can do with it. I'm just really concerned about the intersection
+between that and, you know, the capitalism system that we're all
+forced to live under.
+
+<BlockQuote>How do we combat abuse or data that isn't labeled as AI
+generated? Are we in the death of the Information Age because of
+this?</BlockQuote>
+
+Oh. I have no idea.
+
+On my blog I've been tracking AI-generated content farms and the tools
+that they use to do it because it's kind of horrifying how easy it is
+to get ChatGPT to hallucinate something about how to make soap with
+radishes.
+
+By the way, don't do that. It'll kill you. It will actually kill you
+dead. Do not do that. No, I'm actually serious here.
+
+The worst part is how this intersects with content farms, those random
+websites you find on Google with negative amounts of information and
+ads everywhere. I've already seen ChatGPT make that problem worse.
+
+Hell, there was this SEO heist a while ago where this person basically
+fed Google Trends results into ChatGPT, SEO heisted by rewriting their
+competitor's website entirely from scratch and stole all their traffic
+and made a whole bunch of ad money contributing nothing to society.
+
+I don't really know how this is all going to work out, but I really
+hope we're not in the death of the information age because that's what
+pays my bills. But if things keep going the way they're going, I can't
+help but agree that we may be on the decline of everything getting
+drowned in pages of trivia and celebrity bullshit.
+
+<BlockQuote>
+Thanks for talking. You see that, you know, that the technology
+naturally democratizes people's access to information. Won't more
+access to information make things better for everyone?
+</BlockQuote>
+
+I'm very glad that inference is getting so much cheaper, like, hell,
+this MacBook right here (I would lift it up, but it's hooked up via
+USB and I don't want to disconnect it). It can run Mixtral [a model
+considered roughly equivalent to GPT-3.5, the model used for ChatGPT]
+and it's just a random MacBook off the shelf.
+Looking back, I kind of regret not getting as much RAM because I
+didn't think I would be doing all this, but, you know, c'est la vie
+[Canadian idiom meaning "that's life"].
+
+I have been thinking about doing an experiment of using Q-LoRA to
+train the ultimate recommendation engine based off of posts that I've
+either commented on or upvoted on Hacker News and using that as input
+with the classification of like or dislike. And because I downvote or
+flag a fair number of posts there, I can use that to create a somewhat
+rough aggregate of things that I would be interested in. And that
+would be something that I see could be a really interesting
+application of all this.
+
+Like I said, though, the open source AI stuff is really cool, but the
+intersection between that and the system and the powers that be today,
+I don't know how that's going to happen and I'm just afraid that it
+won't end up good for all of us.
+
+But thank you for all the questions. I am really happy that I was able
+to get you to be engaged with this topic and really start thinking
+because I don't know what's going to happen either.
+
+Thank you so much. Good night all! Drive home safely! The roads are
+wild.
+
+
+
+