OpenAI’s Mira Murati: “some creative jobs maybe will go away, but maybe they shouldn’t have been there in the first place” And you stole everything from creative people who provided free texts, images, forum answers, etc. To date, your company has refused to acknowledge any credit. Rich people truly live in their bubble and have zero sympathy for fellow human or their livelihood.
So, to be fair, the line after the quote refers specifically to those that produce “low quality” output.
So a charitable but not unreasonable read might be that she’s saying any creative role that’s easily replaced with AI isn’t really a loss. In some cases, when we’re talking about artists just trying to make a living, this is really some vile shit. But in the case of email monkeys in corporations and shitty designers and marketeers, maybe she’s got a point along the same lines as “bullshit jobs” logic.
On the other hand, the tech industry’s overriding presumption that disruption by tech is a natural good and that they’re correctly placed as the actuators of that “good” really needs a lot more mainstream push back. It’s why she felt comfortable declaring some people in industries she likely knows nothing about “shouldn’t exist” and why there weren’t snickers, laughter and immediate rebukes, especially given the lack (from the snippet I saw) of any concern for what the fuck happens when some shitty tech takes away people’s livelihoods.
If big tech’s track record were laid out, in terms of efficiency, cost, quality etc, in relation to the totality of the economy, not just the profits of its CEOs … I’m just not sure any of the hype cloud around it would look reasonable anymore. With that out of the way, then something so arrogant as this can be more easily ridiculed for the fart-sniffing hype that it so easily can be.
I get your overall point and don’t disagree.
The thing is about that specific bit - no job is a bullshit job when it’s what you rely on to pay your bills. Even if you don’t like the job, even if you aren’t the best at it, if it’s keeping a roof over your head, having it arbitrarily erased by some technology that didn’t exist 10 years ago is a pretty shitty thing. And even if that’s somehow an inescapable reality of progress (and I think there is a lot of discussion that could be had about that concept) it’s still shitty for her to portray that as no big deal. I don’t think context makes the comment much less shitty than the headline implies.
In addition to the fact that as has been pointed out repeatedly, AI learned how to do what it does from the output of the jobs it will now destroy, and without compensation of any kind to the people who created that output.
OK, so one caveat and one outright disagreement there.
The caveat is that she herself points out that nobody knows whether the jobs created will outnumber the jobs destroyed, or perhaps just be even and result in higher quality jobs. She points out there is no rigorous research on this, and she’s not wrong. There’s mostly either panic or giddy, greedy excitement.
The disagreement is that no, AI won’t destroy jobs it’s learning from. Absolutely no way. It’s nowhere near good enough for that. Weirdly, Murati is way more realistic about this than the average critic, who seems to mostly have bought into the hype from the average techbro almost completely.
Murati’s point is you can only replace jobs that are entirely repetitive. You can perhaps retopologize a mesh, code a loop, marginally improve on the current customer service bots.
The moment there is a decision to be made, an aesthetic choice or a bit of nuance you need a human. We have no proof that you will not need a human or that AI will get better and fill that blank. Technology doesn’t scale linearly.
Now, I concede that only applies if you want the quality of the product to stay consistent. We’ve all seen places where they don’t give a crap about that, so listicle peddlers now have one guy proofreading reams of AI generated garbage. And we’ve all noticed how bad that output is. And you’re not wrong in that the poor guy churning those out before AI did need that paycheck and will need a new job. But if anything that’s a good argument for conusming media that is… you know, good? From that perspective I almost see the “that job shouldn’t have existed” point, honestly.
Even if we take as settled the concept that more jobs will exist in aggregate, I’m doubtful that there’s a likely path for most of the first wave (at least) of people whose jobs are destroyed into one of those jobs “created” by AI. I have nothing to back this up but my gut, however in this case I feel pretty good about that assertion. My point is that their personal tragedy at losing their job is in most cases not going to be alleviated by the new jobs created by this advancement.
I’ve seen recent AI porn images and I saw what Deepdream was doing a few years ago. I don’t see a reason to think we can’t expect it to get better based on that. 🙂 I also acknowledge that these may be apples and oranges even more than I suspect they are.
As someone who works in IT (though as I’m sure you can tell I have no expertise whatsoever in machine learning), I still tend to strongly agree with this statement from @[email protected] :
Every industrial transition generates that, though. Forget the Industrial Revolution, these people love to be compared to that. Think of the first transition to data-driven businesses or the gig economy. Yeah, there’s a chunk of people caught in the middle that struggle to shift to the new model in time. That’s why you need strong safety nets to help people transition to new industries or at least to give them a dignified retirement out of the workforce. That’s neither here nor there, if it’s not AI it’ll be the next thing.
About the linear increase path, that reasoning is the same old Moore’s law trap. Every line going up keeps going up if you keep drawing it with the same slope forever. In nature and economics lines going up tend to flatten again at some point. The uncertainty is whether this line flattens out at “passable chatbots you can’t really trust” or it goes to the next step after that. Given what is out there about the pace of improvement and so on, I’d say we’re probably close to progress becoming incremental, but I don’t think anybody knows for sure yet.
And to be perfectly clear, this is not the same as saying that all tech disruption is good. Honestly, I don’t think tech disruption has any morality of any kind. Tech is tech. It defines a framework for enterprise, labor and economics. Every framework needs regulation and support to make it work acceptably because every framework has inequalities and misbehaviors. You can’t regulate data capitalism the way you did commodities capitalism and that needed a different framework than agrarian societies and so on. Genies don’t get put back in bottles, you just learn to regulate and manage the world they leave behind when they come out. And, if you catch it soon enough, maybe you get to it in time to ask for one wish that isn’t just some rich guy’s wet dream.
Just a clarification: the “gig economy” was not “new” in any way, just using new technology to skirt around labor laws and find loopholes in regulations in order to claw back profits that had been “lost” to things like pensions and health coverage.
Well, yeah, that’s what I’m talking about here, specifically. There was an application of technology that bypassed regulations put in place to manage a previous iteration of that technology and there was a period of lawlessness that then needed new regulation. The solutions were different in different places. Some banned the practice, some equated it with employees, some with contractors, some made custom legislation.
But ultimately the new framework needed regulation just like the old framework did. The fiction that the old version was inherently more protected is an illusion created by the fact that we were born after common sense guardrails were built for that version of things.
AI is the same. It changes some things, we’re gonna need new tools to deal with the things it changes. Not because it’s worse, but because it’s the same thing in a new wrapper.
Thank you for clarifying. Definitely agree on this. Especially with regards to the perceived guardrails.
I agree with most of what you wrote in this paragraph, but we have no such strong safety nets. I don’t think the fact that it has happened previously is justification for creating those circumstances again now (or in the future) without concern for how it impacts people. We’re supposed to be getting better as time goes by. (not that we are by many other metrics I can see on a daily basis, but as you say that’s another conversation)
I also agree with this.
But, I find there is plenty of justification to push back and try to slow the proliferation of AI in certain areas while our laws and morality try to catch up.
Your “we” and my “we” are probably not the same, I’m afraid. I’m not shocked that the difference in context would result in a difference of perception, but I’d argue that you guys would need an overhaul on the regulations and safety nets thing regardless.
Fair point, this ol’ nation needs a new set of spark plugs and a valve job, at a minimum. :)
Edit: DAMMIT how am I a moderator again @[email protected]? Removing myself again now.
I’m with you on that.
Thanks. I hate deliberately out of context quotes. Watching the entire interview is actually very interesting. Lots to agree and disagree with here without having to… you know, make things up.
On the jobs situation she later mentions that “the weigth of how many jobs are created, how many jobs are changed, how many jobs are destroyed, I don’t know. I don’t think anybody knows(…), because it’s not been rigorously studied, and I really think it should be”. That also comes after a comment about how jobs that are “entirely repetitive” (she repeats that multiple times) may be removed, but she clarifies that she means jobs where the human element “isn’t advancing anything”, which I think puts the creative jobs quote in context as well. I like how the intervewer immediately goes to “maybe we can cut QA” and you can see in her face that she goes “yeah, no, I’m gonna need those” before going for a compromise answer.
I don’t agree with the perspective she puts forward about how the tools are used, I think she’s being disingenuous about the long term impact and especially about the regulations and what they do to their competitors. But latching onto this out of context is missing the point.
That all makes sense. Thanks! I was tempted to watch the whole interview but realised I didn’t actually care that much. I figured something like what you describe was where she was going just from the clip I watched, and of course it was obvious that rage baiting was going on here.
And exactly what you said … many listening to the full interview will probably think she, and therefore AI generally, is actually “right” and will be valuable … except for all the shitty ways it’s gonna be used by corporations and all of the shitty and presumptive things OpenAI and others will do to win their new platform war.
I think there’s plenty of rightful criticism to the things she actually says, and plenty of things she says I wouldn’t take at face value because they’re effectively token corporate actions to dismiss genuine concerns.
She actually gets asked in the Q&A about the IP rights of creators included in training data, and she talks about some ideas to calculate contributions from people and compensate for them, but it’s all clearly not a priority and not a full solution. I’m not gonna get into my personal proposals for any of that, but I certainly don’t think they’re thinking about it the right way.
Also, if you REALLY want a chilling thing she says, go find the part where she says they may eventually allow people to customize the moral and political views of their chatbots on top of a standard framework, and she specifically mentions allowing churches to do that. That may be the most actually dystopian concept I’ve heard come out of this corner of the techbrosphere so far, even with all the caveats about locking down a common baseline of values she mentions.
eeeesh.