Therapeutic Outlet - AI Will (Probably) Not Replace You

Posted on Saturday, Nov 15, 2025 | Mental Health, Art, Creativity, Mental Illness, Art Therapy, Creativity, Generative AI
Alex returns to the thorny subject of generative AI and creativity reiterating his case for AI as a valid part of the creative process, but acknowledging the ethical and moral dilemmas it poses. He discusses the current, precarious state of the AI industry and speculates of what’s to come as the AI bubble is stretched to its limits.

Transcript

Nobody can miss them for me That’s reason why I try both This is all that’s old, I can see a thousand times of it This is all that’s old, I can see a thousand times of it I’ve been watching Northern Exposure, the 90s TV series for a while now It’s now available again on Amazon and I used to love this TV show when I was a kid and I always really basically hero worshiped Chris, the radio DJ sometimes known as Chris in the morning, who is not only a very handsome fellow but says some pretty profound stuff and I always sort of pictured myself maybe at some point in the future I could be like a radio DJ because this was before podcasts and even actually really in any meaningful sense the internet existed and so I could sit there and yammer on about whatever was in my brain and people would listen and so I guess I got my wish to a degree whether anyone’s listening or not is an open question but here I am pretending to be Chris in the morning nowhere near as easy on the eye and maybe I’m not as easy on the ear either but it is what it is, I am what I am and I’m here and you got me With all of that, what am I going to talk about today? I’m going to talk about AI and I’ve spoken about this several times before in the past there’s mentions of it on quite a few episodes but I believe I’ve only ever done one episode dedicated to it and that was in July 2024, so last year a year and a quarter ago I guess and that might not seem that long a time but in the age of AI it’s quite ridiculous, it’s a lot and I mean a lot has changed since then not just in the world of AI but in my world as well and so I wanted to sort of revisit it I should note before I get into this that I may make some statements like I did on my last AI episode which was called AI will not replace you and I should point out that I’m not referring to everyone or anyone in any part of their life will not be replaced in any way by generative AI I recognise that for all its faults as a technology set it does do certain things reasonably well and that might put pressure on certain industries and certain skill sets and people will rightly or wrongly lose their jobs I am talking about this in my capacity as an artist and a podcaster rather than making any proclamations from a wider business perspective although I might allude to some of that so if you are under pressure from generative AI on your day job you have my sympathies and I recognise that there’s more going on than just the stuff that I’m going to talk about here anyway what did I say a year and a bit ago well I said that generative AI as it was then is pretty good at some stuff but it’s not that good I said that ultimately machines can’t create culture and that artists being humans will always be in demand because people connect with people not machines I also said that I consider it completely legitimate for people to use AI as part of their creative process and that indeed it’s completely legitimate to use AI entirely as your creative process because in my mind it’s still art, it just isn’t necessarily going to be good art and I would argue that in most cases it isn’t but either way getting an AI to produce something of even vague value is not quite as easy as it might first seem and you’ll often end up producing lots and lots of versions of something before you find one that achieves whatever goal you’re trying to achieve and that itself is an artistic choice, a qualitative choice and therefore there is some creative element that goes on in that process now things really have moved on since then I should just reiterate that as I said before in the prior episodes my day job to a degree is generative AI or AI and machine learning, data science I’m a data scientist for my day job when I can get the work doing it and that involves both the usage and the creation of forms of AI and there are quite a few and I in my past roles up to well over a decade ago was experimenting with the early versions of what we would now call large language models like chat, GPT and Claude and so I’m pretty familiar with both the process and technology required to create these models as well as how to use them to the point where I actually have another podcast it’s called the confusion matrix wherever you get your podcasts if that sort of thing interests you which looks at things from a purely AI perspective with probably more of a focus on language models and how they’re deployed and used and how we maintain quality and accuracy and so on on these things and I take quite a critical view there as well and so I’m quite deeply embedded in this stuff and I have lots of opinions but those opinions have to sort of be adjusted all the time because things are changing so quickly although I’d argue that over the last say six months they’ve changed a lot less frequently, things have moved on we seem to meet started to head towards a plateau where the progression of the technology is not quite at the pace that it has been but last year in July I talked about tools like Dali and Mid-Journey as visual creative tools and just to clarify what these things were they are tools where you can go in, type in a prompt, a set of words can you make me a picture of a squirrel dressed as a SAS gunning down a bunch of Easter eggs and it will attempt to produce that and in most cases would produce something and back then Dali which no longer exists has been replaced by OpenAI Sora Dali was OpenAI’s image generation model the stuff that was producing was not great you could get it to produce some interesting things that were often odd and didn’t quite do what you wanted them to but were quite funny but often completely mangled and just bad and since then various other models have emerged that are both from a static image creation perspective as well as a video generation perspective are much more effective and so the one that I’m probably most familiar with is Google’s image effects which is part of a suite of tools and it’s really good now I need to qualify that a little it produces extremely high quality images that are often quite close to what you might want if you use the right words to describe them you can get it to produce hyper realistic close ups of human face for example and they will look highly photographic and extremely realistic perhaps too realistic if such thing is possible but the sort of glossy, high quality, high definition images you might find on a fashion site or something that’s been photographed with a really good quality camera but then manipulated with Photoshop or something so these things are perfect so not really that realistic but like insanely good quality and it will produce this in a matter of seconds and let’s put the valid criticisms aside for a minute and the plagiarism and copyright issues that’s amazing that it can do that and I think this is an aspect of the generative AI revolution and the predictable and gathering steam backlash it’s worth pausing just for a minute because I would struggle to find someone who wouldn’t agree when they sat and thought about it for a minute that that wasn’t completely amazing that that’s possible at all and even a year ago something like that simply wasn’t conceivable and how amazing it is that these things can write large volumes of pretty high quality language or code programming languages and any other things that you may have criticisms about but the very fact that it can do it at all the very fact that I can ask it to build me a Python script that does something fairly complex and I have done many times the fact that it can just sort of do that again with problems, with faults but the fact that it can do it at all is mind blowing, right? and if you’d asked me a couple of years ago whether we were anywhere near being able to do stuff like that I would have laughed at you but here we are, this stuff, despite being completely spectacularly amazing that it exists at all, has become already a bit old hat and people are looking at it again, well that’s not very good is it? it doesn’t get this right, it doesn’t get that right and the criticisms are completely valid but still we seem to have forgotten about how amazing it is that this exists at all now I’m being very very careful here to not paint the picture that eclipses the problems that are created when, now that these things are out there being used in a sort of commercial arena I’m simply stating that the fact that it can do it at all is just so cool and from a person who comes from my industry and does what I do for a living and understands how these things work it’s especially nerdily amazing and I’m not going to stop thinking that, it blows my mind but we are all haters to a degree on the rest of it because I use generative AI a lot I have to do for my job to keep up with what’s going on I have to do for my podcast, I have to be using this stuff I have to understand it, I have to understand the differences between the models and what’s available and what’s out there and again I’m not an authority on this stuff but I’m pretty well versed on what’s available and I use it a lot and I use it in my day job in terms of, I write very little code anymore I use generative or large language model agents that are able to do multi-step processes for example, I think a lot of people will look at chat GPT as being the most advanced implementation of language models but it’s not by a very very long way there’s a lot more coming down the line if you’ve only ever experienced chat GPT which is pretty much a one-shot thing by which I mean you ask it a question or you give it a task and it does that one thing, it responds with tell me when Taylor Swift was born and I hope it might give you a right answer it might give you a wrong answer but again give you an answer with some extra words around it and then it stops and waits for you to do something else to ask it a different question like where was she born or something and you can have these long chat streams and because the whole stream is fed back into the model every time you make a query it can give contextual answers so if you’ve asked that question when was Taylor Swift born it will go whenever that is, I’m not going to look at that, I don’t care it will, and then you ask the question and where was she born without mentioning Taylor Swift again I certainly should recognise that’s what you were referring to and give you whatever answer it thinks is correct which may or may not be correct and so it can look like a sort of an interactive process it can look like a conversation is evolving it can look like the model knows something about you or at least knows something about the context of the current situation but that isn’t true at all, models are completely static they don’t learn at all everything that the model knows about you or the current conversation is sent to the model every single time you make a query and in this sense they’re sort of limited that particular facet has its values and usages but it can be quite limiting now we’re moving into a world, particularly if you’re already in technology where we have these things called AI agents which have some level of autonomy and control over something so you might say to it can you take all my photos of Taylor Swift in this folder rename them and put them in another folder on my computer file system I don’t know why I’m talking about Taylor Swift here I guess it’s because everyone’s always talking about her and therefore her name seems to pop into my mind quite a lot but I don’t really spend much time thinking about Taylor Swift otherwise or even listening to her music so I don’t know why you want to move your pictures of Taylor Swift but that’s a multi-step process to do it yourself you have to know where that folder is you have to open the folder, you have to list the files you have to control A, control C, whatever you’re going to do to copy them you have to go find this other folder or create it then you have to control V or to copy them into that new place and then if you wanted to rename them, so they all had say Taylor Swift at the beginning of the file name then you’re going to have to go through each of those now if there’s 2 or 3 of them, probably not a big issue if there’s 20, 30, 200, 300, that could be quite a time consuming process and you have to go through each of them and click on them and change the name or if you’re using a Unix or Linux based system it’s a little easier but still a little bit more time consuming whereas with an AI agent you can just type in the words in English can you move this folder of images to that folder and rename them all with the words Taylor Swift at the beginning and it will just go off and do it and that’s a multi-stage process and you’ve given it at that point access to your file system so it’s run multiple commands and done multiple actions and it’s done it all on its own, it might ask you for permission it might ask you to clarify, it might do any number of things but when it gets on with the thing it’s doing, it will just do it and that makes these things incredibly powerful where there is some level of repetitive process and something that can be done electronically that will otherwise require human intervention and be quite time consuming and potentially boring and of relatively low value so presuming you’ve got a good reason for moving those photos and maybe that’s going to make your life much easier for some reason it’s still relatively low value as a task it’s not going to change anyone’s life, it probably just makes your job a bit easier and maybe you wouldn’t have bothered doing it at all had you not had an agent there to do it for you and this is where the pressure comes from this is where I think a lot of people just see chat GPT as a proxy for search like a proxy for Google as a bit of a limited use product if you’ve never experienced an agent I can understand that but once you’ve seen these agents at work you start to realise how powerful this stuff can be now the agents suffer from exactly the same problem that the one shot language interfaces do which is that it can get things wrong and if something gets Taylor Swift’s birthday wrong then that’s annoying but if when moving all your files it actually used the wrong command to delete them permanently that’s considerably worse and if that wasn’t a set of fairly innocuous image files but some vital legal documents or something then this could be very problematic indeed and so these things aren’t a magic bullet they’re going to solve all our productivity problems in certain circumstances when monitored properly and you understand what they’re doing and how they’re doing it the right guardrails are put in place they can be really really incredible productivity tools and that’s great but those guardrails and your knowledge of what it’s doing and how it’s doing it is essential and you need to pay close attention to these things but when people are talking about jobs being taken they’re not saying that chat GPT takes their jobs they’re saying that one of these agencies and there’s lots and lots of products that call themselves AI agents and all the big companies Amazon, Google, OpenAI, Anthropic and so on all have their own versions of agent and if you’re using platforms like Microsoft VS Code as a developer then there’s lots of them available within that too so these things currently very much in the tech nerdery realm are going to start spilling out into everyday life and will do things like automation of processing of imagery and I use these things quite a lot I want a mosaic of images for like a social media share or a little leaflet poster thing I made to advertise my art portrait work so I wanted a little mosaic, a little grid of images that had some of my examples of my work and I could have just done it in something like Word or PowerPoint and dragged and repositioned and so on these images and done it by hand but I knew I had this need quite a lot so I just had an AI agent to write me a little program that given a set of images or a folder full of images will create a little mosaic or grid of those images based on certain parameters the amount of columns, the amount of images and so on and it’s incredibly useful and I wouldn’t have bothered doing that before it just wasn’t worth the effort but it took me five minutes to do and every time I think oh I could do with it being able to do this as well or that then I just go to the agent and go can you update this now so that it can take this parameter or I can change this about it or whatever resize the images, make the monochrome, whatever and so already I’ve got capability that I probably wouldn’t have had I could have created it, I could have probably found a piece of software that did it for me but I just built one that did it exactly to spec and I used that a lot and I think it’s going to start seeing more and more people realising this capability to manipulate images, change them, group them and this is just in terms of basic productivity I don’t see that that task in itself particularly affects anyone else it’s just made my life a bit easier and made my marketing perhaps a little bit more effective and so these are the types of use cases that are going to come up more and more but as you can see when you start to think about things in multi-step or being able to do things you couldn’t have done before enterprising people are going to start finding ways to do much more complex and exotic tasks with these now this is before we even get into the realm of of just being able to type a prompt into an image generation model to create whatever you want so I use generative AI constantly and I think I mentioned in the last episode I did on this that even at that point I was building out my own AI image generation tools and training my own models to create these gnarly, weird, spooky pictures that I created by constraining image generation models so they were not actually very good at doing that and they got images all mixed up and mangled up in really cool exotic ways and I’ve since moved on from that because what I realised was that the image generation models that we’ve got now things like ImageFX and Sora and others did something similar without me even having to intervene because much like the language models where they are getting stuff wrong the image generation models get stuff wrong too but as opposed to just giving me something that’s just incorrect they give me something that’s weird it too is incorrect but I might ask for a picture of someone sitting on a chair and they might give that person three arms which if what I want is a picture of someone sitting in a chair for an advert or something I want to create or for a reference for an artwork but what I really need is someone sitting in a chair with just the normal number of arms that’s highly problematic but if you’ve got a brain like mine who really really likes real weird, unexpected, serendipitous stuff then someone turning up in an image with three arms for me is beguilingly, fascinatingly brilliant and I love it and that’s a fairly facile example because these things tend to get things wrong in all sorts of weird ways and I find it incredibly entertaining and I also find it really really stimulating I find that the weirdnesses and the wrongnesses that are created by models will really fuel my creative juices and give me ideas and I think that’s no different from the ideation process that any artist who wants to create pictures of things that don’t exist in the real world would go through anyway you’re looking for things in the world that will trigger off an imaginative thread that you can pull on and sometimes it might seem like that stuff comes to you out of the ether but almost always it’s about something within the world that you’ve experienced will trigger that off and I find that the weirdnesses of these models are really good for my creative process and to the point where I love the stuff they produce so much that I just print it out and start sticking it on my canvases and use it as a baseline or a starting point for my artworks that’s not unusual for me I use a lot of collage in my art the collage comes from things I found on the web I use imagery of people and places and things that are out there and available and as part of the collage process that’s kind of how collage works and so I’m already repurposing stuff that is from other people’s work and some of the themes and the subjects of my artworks are real human beings like movie stars or so on and I’ve always been probably skated slightly too close to the copyright line on that which I’ve always been very uncomfortable with but given that I’m aphantasic, I have no mind’s eye I’m kind of constricted to depicting imagery from photographs because I can’t imagine and I love painting pictures of humans, that’s what makes me happy and I can’t imagine humans I can’t conjure up something in my brain and paint it or draw it because there’s nothing to conjure up so I’m kind of constrained to images of stuff and what generative AI allows me to do is create sort of generic images that I can use as a starting point or to work from as a reference now I’m aware that to get to that point the models have had to plagiarise lots of other people’s stuff and whether or not the image that’s spat out to me is a direct copy of someone else’s picture or someone else’s likeness is entirely opaque to me I’m not even sure if it’s currently possible to find out whether that’s the case and I’m absolutely certain knowing what I know about these models that the provenance is not something that the model creators are able to verify one way or the other, the models don’t work like that they are abstract mathematical entities so you can’t reach into it and say well which bit of code created this image because nothing created it, it was just a statistical, probabilistic, mathematical rendition based on some input text they’ve got no more idea of how I got to the output that I did than they do, right? than I do, sorry and so it’s copyright dubious but in some ways that ship has sailed I think that the creative fightback from the creative industries is picking up speed governments are slowly starting to react to this and things are being legislated for and the landscape will change because of this and I’m going to run with that, right? but at the moment I find image generation tools, particularly Google’s image effects to be so beguiling and fascinating and interesting and useful that I find it very, very difficult not to use it it’s very, very, very well suited to my needs and I hope I use it well but I’m aware that there could be criticisms to this and how I’m using it I’m open minded to hearing those but it’s hard for me to get past the reality that these things are useful for me and I was already using image generation AI of sorts before these things started to become really, really powerful I’ve experimented with the video side of things as well but video isn’t really my medium but I’ve done some quite cool stuff with that mainly just to see how it worked and I think we’re quite a long way away from people being able to generate long form video currently you can do like 4 or 8 seconds at a time with something like Google’s Flow or Sora and that’s cool because it will look amazing in some cases and I’ve produced some stuff that looks absolutely spectacular but then when you want to continue that it’s very hard to do it’s very hard to constrain the machine to do what you want and so the flip side of me using this to create weirdness is that if you want to create something that is very much to spec you have to write quite a lot of prompt and be very, very detailed in particular and specific to get something even vaguely like what you need because you need a very particular image of something or a video of something and then a lot of the time the model will just not do what you want if you don’t understand how that particular version of that model needs to hear what you need to tell it then it ain’t going to give you shit of any use it’s going to keep getting it wrong and you’re going to have to keep iterating you’re going to have to keep burning GPU and CPU time which is a lot of energy and a lot of water consumption these things are very bad for the environment and to get something right you might have to produce thousands of images before you’ll get one that probably just about maybe works to spec and I think this is a much bigger problem more often than not I’ll take whatever comes out of the machine I’ll have it iterate a couple of times to give me some options but my usage is actually relatively low and I’ve got, I’m very idiosyncratic I want the weird images, I want the messed up-ness but as most people don’t and so from a commercial perspective the image and video generation stuff I think despite what many people have claimed is far from coming of age probably famous last words here because things really have moved a lot faster than anyone should have expected so maybe doing long form video is only but a few weeks or months away but we’re not there yet but there are plenty of roles that are going to be put at risk by this type of stuff particularly in the creative industry apparently coders are losing their jobs I think anyone trusting an AI to code on mature code bases deserves everything they get and very little of that is going to be good but I feel like there’s a certain amount of plateau happening here the most recent language models iterations GPT-5 was not a disaster as such but it was a big PR nightmare as a lot of people who were using GPT one of the GPT-4 range of models freaked out because the personality of their personal assistant suddenly changed plus GPT-5 from a technical perspective is if anything a step backwards and GPT-5.1 just came out to almost no fanfare and the same goes for the various other model creators the public have shrugged collectively although many people are using these tools on a day to day basis no one’s particularly excited about it and the markets are starting to cool off and people are openly talking about there being an AI bubble and what they mean by that is that the AI companies are hugely overvalued or the perception is they’re hugely overvalued their share prices are way too high not one of the companies who are purely AI and we’re talking open AI and anthropic really here and a few other small players are in any way profitable or even close to it they have sucked up literally trillions of dollars of investment and created huge amounts of infrastructure bought hundreds of thousands if not millions of GPU chips which are the bits of computer hardware needed to efficiently run these models and they just simply haven’t translated that into a commercially viable business and the same goes for the major players who are trying to monetize generative AI the likes of Microsoft and Amazon, Google and so on simply have failed to meaningfully translate the AI elements of their product portfolios into profitable things now the AI bubble is very bizarre because usually to justify high valuations there has to be some clear path to profitability and thus far after several years of rapid evolution of these models still there’s no compelling consumer or business use case from a financial perspective and I think that the markets are only going to stay along for the ride for so long and at some point the share prices of these companies are going to collapse and several of them are going to go completely out of business which is going to send ripple effects through the global economy because something like three quarters of the growth in the American economy the biggest economy in the world by some distance over the past few years has been driven purely by technology companies and particularly around the prospect of generative AI so at some point that’s going to collapse the growth in the American economy will be taken down with it and since these companies are now borrowing money to drive their insatiable need for energy and hardware they may well default on debts which will further impact the wider economy and could cause something resembling a recession or a proper financial crash in the worst worst case scenario I’m not suggesting that’s inevitable and so there’s going to be a showdown at some point with reality the other side of this there will still be the models there will still be the technology the technology will still exist and the hardware will exist to maintain, grow, retrain and continue to deliver these models even if for example the price is likely to go up and the speed might go down but these things aren’t going away let’s be clear if the absolute worst of the most cynical pessimists case scenario the most AI hating people if their worst possible most cynical state of affairs happens we will still be left with the models we will still be left with the algorithms that created them the technology and the hardware to run them these things are not going away now the landscape might change who uses them and how they use them might change the investment might shift to other aspects and let’s be clear that these things do some amazing things for example the healthcare use cases are really quite compelling in terms of augmenting humans in the healthcare arena these things are incredibly powerful and will save lives so we don’t want them gone we just don’t want them destroying the environment increasing societal and economic disparity replacing people and making people jobless and perhaps even homeless or certainly in financial difficulties simply because greed the billionaires want to have slightly cheaper workforces so any number of things could come out the other side of this but what’s not going to change is the existence of these models again the constraints around things like copyright and plagiarism again lots to play out there but still still still still still these models will remain they will still be here they will still be doing things that may or may not be useful they will still be getting things wrong and people will still be using them and people like me and I don’t want them to go away I find them a really valuable, really exciting and interesting creative tool that I do not use AIs to create artworks in their own right I do not use them exclusively I do not take the output of generative AI and pass it off of something that I’ve created when I label up my artworks, if there’s an AI element I will add that as just one of the things like when I say that this is acrylic and mixed media on canvas I will say acrylic AI mixed media on canvas and I’m not going to shy away from that or apologize for it as I said earlier I will accept and engage in meaningful debate about the validity of what I do and the potential ethical and moral implications of it and I’m an open mind but I’ve got an open mind with regards to the fact that I consume meat which from a moral and ethical perspective is still dubious to me but I also really like meat and I recognize that that’s problematic and I’m open to people having different opinions to me on these things and maybe one day I’ll go vegetarian or vegan I’ve tried it before to obvious limited levels of success but you know this is all a journey and we are who we are so I don’t really have a conclusion to all of this I didn’t really see this as well as my usual here’s a point that I’m making here’s what I want you to take away from this I think this is more general outpouring of my current point of view on the generative AI arena and I’ve absolutely got a lot more to say on this and again if you’re interested in the more sort of business and technology focused edge of this then go listen to my other podcast The Confusion Matrix, wherever you get your podcasts where you’ll get a lot of my thoughts on this stuff but are relatively little if not very little reference to the creative realm this will not be the last time I monologue on this stuff I’m almost certainly going to stop recording in a minute and then think oh no I meant to say this, I meant to say that, I meant to say the other or fine when I edit this down a little bit that I’ve probably said a bunch of nonsense but if you’re listening to this I decided to publish it but yeah I hope that has been of some interest or value to you things are going to keep changing rapidly if anyone tells you that they know what’s going to happen on this front they are either an idiot or a liar the one thing I can say is that a lot is going to change so if the next time I do one of these it’s in a year’s time I fully expect to be saying wow that was an interesting year here’s what’s changed and it’s huge and that will also be affected by the wider social economic and political landscape of a very very weird and somewhat scary world that we live in that’s all I’ve got to say, I’m not going to say anything else and I’ll be back at some point with something else bye

Show Notes

Alex returns to the thorny subject of generative AI and creativity reiterating his case for AI as a valid part of the creative process, but acknowledging the ethical and moral dilemmas it poses. He discusses the current, precarious state of the AI industry and speculates of what’s to come as the AI bubble is stretched to its limits.

You might want to listen to the earlier episode on Gen AI and creativity first: https://aami.alexloveless.co.uk/therapeutic-outlet-ai-will-not-replace-you/

Also, remember to subscribe to my substack for even more content: https://artagainstmentalillness.substack.com/