Reshaping the AI Industry

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry

Contents

1. Rationale

Why is it important? Why is it crucial? First. We, uh, need to make sure that if we figure alignment out people actually implement it. Like, imagine that tomorrow someone comes up with a clever hack that robustly solves the alignment problem… but it increases the compute necessary to train any given ML model by 10%, or it’s a bit tricky to implement, or something. Does the wider AI community universally adopt that solution? Or do they ignore it? Or do the industry leaders, after we extensively campaign, pinky-swear to use it the moment they start training models they feel might actually pose a threat, then predictably and fatally misjudge that? In other words: When the time comes, we’ll need to convince people that safety is important enough to fuss around a bit for its sake. But if we can’t convince them to do that now, what would change then? I suppose having a concrete solution, instead of vague prophecies of doom, would give us more credibility. But… would it, really? And what if we won’t have a concrete solution even then, just a bunch of weird heuristics that may nonetheless measurably improve our odds? The latter seems reasonably likely, too. As this excellent post points out, most of the contemporary deep-learning progress comes from "messy bottom-up atheoretical empirical tinkering". If AGI comes from DL, it’s plausible that, even if we arrive at the solution to alignment from mathematical foundations, the actual implementation will take the form of messy hacks. Ones that will probably need to be fine-tuned for any given model architecture. And given the no-fire-alarm principle[1], to be safe, we’ll need to ensure that any sufficiently big model is only run with these hacks built-in. If any given AI researcher is still not taking alignment seriously by then, how will we make them bother with all of that every time they run an experiment? How will we ensure they don’t half-ass it even once? Second. If that figure I quoted, the 49:1 ratio, is even remotely correct, there’s plenty of room to upscale our research efforts. Imagine if every researcher started spending 2% more of their time thinking about alignment. That’d double the researcher-hours spent on the problem! Which doesn’t directly translate into 2x progress, I’ll grant. Given the field’s pre-paradigmic status, the returns to scale might be relatively small… but by no means negligible. Even if we don’t necessarily have research directions clearly outlined, having much more people stumbling around in the dark still increases the chances of bumping into something useful. Another argument I’ve seen is that upscaling may increase the capabilities gain too much. I don’t find this convincing:

2. Existing Work

3. Types of Interventions

I would broadly categorize them into the following:

3.0. Effective Strategies for Changing Public Opinion

The titular paper is very relevant here. I’ll summarize a few points.

3.1. Straightforward Appeals to Insiders

As per above, we’d be fighting an uphill battle here. Researchers and managers are knowledgeable on the subject, have undoubtedly heard about AI risk already, and weren’t convinced. Arguments that recontextualize AI risk, AI, or existential risks in general, are likely to be more effective than attempts to tell them things they already know. They’re more likely to misprioritize safety, rather than be totally ignorant of the need for it. An in-depth overview of the best ways to craft an argument is beyond the scope of this post (though this might be a good place to look). Two important meta-principles to keep in mind:

There are two tacks to take here: macro-level and micro-level. Macro. Broad appeals to the entire industry, with the aim of changing the agreed-upon social reality, de-stigmatizing AI Safety, and so on. Concrete projects may look like this. Micro. Targeted efforts to convince industry leaders. As per, 50% of AI progress is made by fewer than 50% of the researchers; orders of magnitude fewer than that. Similarly, getting the leadership of DeepMind and OpenAI fully on our side would have an outsized impact. In theory, a project here may go all the way down to "find effective arguments to convince this specific very important person" levels of fidelity. I’m more optimistic about the second tack, and generally about activism that has precise focused short-term objectives whose success or failure can be clearly evaluated, and which we can quickly iterate on. One of the flaws of the "micro" approach is that our victories may be washed away by a paradigm shift. Most of the top GOFAI researchers didn’t keep their positions into the ML era, and the top ML researchers may not survive into the next one. I expect this isn’t much of a problem, though. If we manage to convince the leading researchers, their views should quickly trickle down to the rest of the field, and the field’s structure is likely to survive an upheaval.

3.2. Sideways Appeals to Insiders

There’s one dimension along which we can broaden our standards for persuasion. When trying to influence people — either individually or en masse — we usually argue that addressing existential risks is necessary because, duh, the looming end of humanity. The importance of that work should be self-evident to any moral person. They’d agree with us if we can only make them recognize the existential threat for the existential threat it is. No, it isn’t just sci-fi! Yes, working out these weird math problems really can save the world! But it’s not the only reason someone might decide to work on AI Safety. People’s career choices are motivated by all kinds of things:

3.3. Appeals to Outsiders

Any effective work along this dimension requires answering an exciting question: how do you put out a flame using a flamethrower? Perhaps that’s a bit harsh. Perhaps even counter-productively harsh, given my previous calls for treating audiences with respect. But let’s not kid ourselves: we’ve seen how the world handled COVID-19. An initiative that pushes for X might convince people or governments to do anti-X instead. If we convince them to do X after all, they might do extremely ineffective things that accomplish nothing, or even somehow do things that actually make anti-X happen. And conversely, activism completely unrelated to X might make it happen! Good news, though: COVID-19 had shown us just how badly things are broken. Keeping the Simulacra Levels and the autopsies of the failures in mind, it might be possible to find interventions that have the effects we want. That’s explicitly what we’d be doing, though: deciding what effect we want to cause, then searching for an action that would cause it, once propagated through the broken pathways of our civilization. For that reason, I’m not making the distinction between "straightforward" and "sideways" appeals here: surface-level efforts to achieve something aren’t strongly correlated with that thing happening, even given their surface-level success. All appeals are sideways appeals. Having a good model of realpolitik is a necessity here. The general principles of "know your audience" and "maintain epistemic hygiene" still apply, though. The inference gap is much larger, but that has its advantages: direct persuasion would be more effective, on average. Useful consequences in this area may include:

3.4. Joining the Winning Side

In some sense, the easiest way to accomplish our goal is not to try to change the AI industry’s incentive structures, but to ride them. The industry as a whole is agnostic with regards to alignment. It cares about:

3.5. Influencing the Research Culture

All of the other approaches attempt to influence the AI industry through intermediaries: through the research projects it pursues, through the people it’s implemented on, through the wider social environment it’s embedded in. But perhaps there is room for a more direct intervention? The industry is a social construct. The qualities that make a project a good one, the tastes the researchers have, the incentives they operate under — all of this is, to some extent, arbitrary. It has a ground-truth component, but the current configuration is not *uniquely determined *by the ground truth of the research subject. Rather, it’s defined by weights that this social construct currently assigns to different features of the ground truth. The current AI industry prefers tinkering to empiricism, and capabilities to safety. How can we shift this? There’s been two proposals that I’ve already mentioned:

I think something like 2) is worth implementing. I’m unclear on how to evaluate 1); I’m guessing mechanistic interpretability just hasn’t progressed that far yet. If we generalize from those two, though... We want to synthesize a construct C with the following properties:

4. What You Can Do

The logistics graph that leads to a superintelligent AI’s deployment has many bottlenecks, and controlling any one of them would be sufficient. Taking over the researcher supply, or the money supply, or the compute supply, or the research project supply, or the reputation supply, or the supply of any other crucial resource I’m not thinking of, would ensure excellent conditions for a safe advanced AI to emerge. But the path to this doesn’t look like a concentrated push along the corresponding dimension. As Pragmatic AI Safety points out, diversification is key. There are interdependencies everywhere: success at one thing affects the probabilities of success of all other projects. Finding an appealing research direction would make it easier to attract people our way. Putting social pressure on major AI labs would make safety-adjacent research directions more appealing. Shifting research tastes in a subfield would make it easier to change people’s minds. And so on. Moreover, it’s not obvious what bottleneck would be the easiest to gain control of, without the benefit of hindsight. Future events and novel discoveries may shift any part of the landscape in unpredictable ways, open or close doors for us. Improving AI Safety’s future position means pursuing a strategy that is robust to such random environmental fluctuations. It means maximizing our far-away action space. We need to have a diversified portfolio of plans; we need to be improving our position all across the board, always looking for what new opportunities have arisen. In theory, it would be great to have central coordination. Some organization or resource which tracks the feasibility of various interventions across the entire gameboard, and pursues/​recommends those that move the gameboard into the most advantageous states while spending the least resources, and also you should put me in charge. In practice, this sort of coordination is both difficult and fragile, with a single point of failure. We’re not a single organization, either, but a diverse conglomerate of organizations, movements, groups and individuals. But we can approximate central coordination. It’s often pointed out that impact in the modern world has a tail-heavy distribution. In some areas, it’s effective to have many separate groups putting their full strength behind diverse high-variance projects. Many of them will fail, but some will succeed massively. The project of advancing AI Safety is, to a large extent, one such area.[3] My general advice would be as follows:

5. Avoid Thermonuclear Ideas

You likely know what I’m talking about. The class of ideas that includes lying and manipulation as its most tame members, and expands to cover some much worse extremes. I know some of these ideas may seem very clever and Appropriately Drastic, and the stakes — literally astronomical — could not be higher. We’re accelerating directly into a wall, and our attempts to swerve away seem ineffectual. It may feel emotionally resonant to resolve to Stop Being Nice and Pull Out All the Stops and solve the problem in some gravely decisive fashion, By Any Means Necessary. But it will not work in the real world, outside fantasies. It will not solve the problem in the long term, and in the meantime it will crash and burn, and hurt people, and ruin our PR, and tank the chances of other, more productive and realistic approaches. Even if you think your idea will definitely succeed, you’re failing to think at scale. What would you expect to work better: a policy under which some of us pursue plans that blow up so hard they set us collectively back a few years, or a policy under which our plans only ever compound on each other’s successes? Following the first policy is a defection, not just against the rest of society, but against all our other risk-mitigation initiatives. We’re better than this. As a rule-of-thumb, you can use something like Shannon’s maxim. If whatever clever plan you’re considering and the entire causal chain that led to it became common knowledge, would it fail and destroy our credibility and our other plans? If yes, this is a radioactive plan, get it away. Things that seem like ruthless pragmatism are frequently not actually ruthlessly pragmatic. They’re just excuses to indulge your base instincts. Be cool, in general. Find ways to be cool about this mess. We have resources for that and everything.

6. The Thin Line

I concur with lc’s post and the people in that post’s comments: we have a slight taboo against the sort of full-scale activism I’m arguing for. It’s exemplified by this sort of sentiment. I suspect it’s a combination of two things:

7. Closing Thoughts

The recent months have seen increasing amounts of alarm and doom-saying in our circles. AI capabilities are advancing rapidly, while our attempts to align it proceed at a frustratingly slow pace. There are optimistic voices, but the general disposition seems quite grim. Well. If alignment is really so hard, maybe we should quit trying to solve it? In hindsight, I’m a bit baffled that field-building wasn’t our main focus this entire time. Getting the AI industry to take AI risk seriously is a necessary and sufficient condition for survival. Solving alignment by ourselves is… neither. If the technical problems are truly insurmountable in the time we have left — and I don’t yet know that they are, but I can certainly imagine it — we should just shift our focus to social-based solutions. The goal, I should note, is not outreach. Convincing a few, or many, AI researchers to switch to alignment won’t solve the problem where we have a multi-billion dollar industry stockpiling uranium in the hopes of spontaneously assembling a nuclear reactor. The aim should be to shift that status quo. Changing people’s minds is a fine instrumental goal, but the terminal one is to influence the robust agent-agnostic process itself. I’d like to suggest that there might be a snowball effect involved — that a 10% progress at this task would make the subsequent 90% easier, and so on. There might, indeed, be. I’m not that optimistic, though. I expect it’ll be an uphill battle all the while, because the sort of carefulness we’d like to cultivate has the tendency to rot away, as organizations become corrupted and people value-drift. It’s possible that this is also impossible. That we can’t change the AI industry in time, any more than we can independently solve alignment in time. But it seems less impossible to me. And if we keep looking for approaches that are less and less impossible, perhaps we’ll find one that isn’t impossible at all.

Comment

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=C9EX5QsnDcNHWvQmh

As someone who has really not been a fan of a lot of the recent conversations on LessWrong that you mentioned, I thought this was substantially better in an actually productive way with some really good analysis.

Also, if you or anyone else has a good concrete idea along these lines, feel free to reach out to me and I can help you get support, funding, etc. if I think the idea is a good one.

(Moderation note: added to the Alignment Forum from LessWrong.)

Comment

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=ia4asCY2bwf6QY7wh

I’d be curious to hear what your thoughts are on the other conversations, or at least specifically which conversations you’re not a fan of?

Comment

My guess is that Evan dislikes the apocalyptic /​panicky conversations that people are recently having on Lesswrong

Comment

That’s my guess also, but I’m more asking just in case that’s not the case, and he disagrees with (for example) the Pragmatic AI Safety sequence, in which case I’d like to know why.

Comment

I was referring to stuff like this, this, and this.

I haven’t finished it yet, but I’ve so far very much enjoyed the Pragmatic AI Safety sequence, though I certainly have disagreements with it.

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=wuoZcARmZgeBLwLLH

IMO prosaic alignment techniques (say, around improving supervision quality through RRM & debate type methods) are highly underrated by the ML research community, even if you ignore x-risk and just optimize for near-term usefulness and intellectual interestingness. I think this is due to a combination of (1) they haven’t been marketed well to the ML community, (2) lack of benchmarks and datasets, (3) need to use human subjects in experiments, (4) it takes a decent amount of compute, which was out of reach, perhaps until recently.

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=aL2pE7XXBa96EnpXC

Great post!

In hindsight, I’m a bit baffled that field-building wasn’t our main focus this entire time.

I have a sneaking suspicion that the implicit thought process here is something like:

I’m a smart computer guy, and this looks like a really important technical problem! Therefore, I should help out by doing what I’m best at, which is writing software and mathematical proofs!

I’m not really one of those management-type community organizers, so I’ll leave that part of the problem to someone else.

The problem being, of course, that smart technical people are way more likely to be convinced by x-risk arguments in the first place than management-type people, so if the nerds stay within their comfort zone, very little field-building will ever get done.

Comment

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=CjHrgPPxr4AAXzdXQ

I have a pretty confident understanding that that is what happened, not a sneaking suspicion.

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=ot9gCAr7KZutHsCcN

There are three books that I massively recommend for anyone who thinks the AI industry is easy to reshape or influence in any direction. These books are Mearshimer’s Tragedy of Great Power Politics and Nye’s Soft Power (2004). The third is basically any book that covers the military significance of AI, in any way whatsoever, such as how AI is mounted on nuclear stealth missiles. In addition, I highly recommend against trying to formulate (or even think about) AI policy without meeting a ton of people with experience with AI in the policy space. Trying to reinvent the wheel on this is a losing strategy, it’s time-inefficient at best, and at worst it can attract unwanted attention from extremely wealthy, powerful, and vicious people. If your proposals are good, and many of them are, it’s best to have them evaluated by experienced individuals who you know personally, not shoved in front of the eyes of as many strangers as possible.

Comment

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=WKjxP4CqaeTs8iMES

This sounds important. Could you say more?

Comment

Yes, books are a big investment, so it was rude of me to fail to explain why it is worth people’s time to look into getting them. Mearshimer’s Tragedy of Great Power Politics (Ch. 1 and 2): Explains in detail why governments and militaries keep doing all these horrible things, like gain-of-function research, or creating offensive nuclear stealth missiles that deliberately disguise their radar signiatures as computer glitches. Nye’s Soft Power (2004, Ch 1 and 4): Explains why governments take the media so seriously, and it gives one of the the best explanations I’ve seen for why massive, competent lies are critical for national security. Chapter 4 also gives a fantastic history of propaganda, including describing the nitty-gritty of how propaganda has become prevalent in modern media. Both of these books are absolutely critical for anyone trying to understand AI policy, and only a small fraction of each book needs to be read in order to get 95% of the neccesary information.

Comment

I didn’t mean to imply any rudeness on your part. Thank you for the recommendation and summary. Could you say in short what the reasons Mearhimer and Nye give and how/​why you think it impact on AI safety? I think it would be good to hear some different perspectives on the issue of A(G)I policy, especially less socially desirable/​cynical ones.

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=BmcEB2vLBARRibdLu

This is an excellent post. It clarified much of my thoughts on the subject; I also hadn’t stumbled upon the Pragmatic AI Safety sequence. Thank you very much for writing this.

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=L7Wf28eKfanNnxnNy

Getting the AI industry to take AI risk seriously is a necessary and sufficient condition for survival. I’m going to play devil’s advocate against this claim. First: the AI industry taking AI risk seriously is not obviously a sufficient condition for survival. In the long run, the hard technical problems would still have to be solved in order to safely use AI. And there would still be a timer before someone built an unsafe AI: FLOPs would presumably still keep getting cheaper, and publicly-available algorithms and insights would keep accumulating (even if somewhat less quickly). Even with the whole AI industry on board, sooner or later some hacker would build an unsafe AI in their basement. Getting the whole AI industry on board would buy time. It would not, in itself, be a win condition. Second: getting the AI industry to take AI risk seriously is not obviously a necessary condition. It is necessary that people working on alignment have a capabilities lead. However, as you mention in the post: Moreover, I don’t think alignment and capabilities are orthogonal. I think they’re very much positively correlated. It is true that today’s alignment researchers do not have any significant capabilities edge (or at least aren’t showing it). But today’s alignment researchers are also not even close to solving the alignment problem. I expect that an alignment research group which was able to solve the hard parts of alignment would also be far ahead of the mainstream on capabilities, because the two are so strongly correlated. I very much doubt that one could figure out how to robustly align general intelligence without also figuring out how to build it efficiently. Strong positive correlation between alignment and capabilities research problems mean that non-alignment researchers win the capabilities race mainly in worlds where the alignment researchers aren’t able to solve the alignment problem anyway.

Comment

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=A9tZwaRwHkZjXLH73

Getting the whole AI industry on board would buy time. It would not, in itself, be a win condition. Mm, I don’t think we’re disagreeing here, I just played fast and loose with definitions. Statement: "If we get the AI industry to take AI Safety seriously, it’s a sufficient condition for survival." If "we" = "humanity", then yes, there’ll still be the work of actually figuring out alignment left to do. I had "we" = "the extant AI Safety community", in the sense that if the AI industry is moved to that desirable state, we could (in theory) just sit on our hands and expect others to solve alignment "on their own". I expect that an alignment research group which was able to solve the hard parts of alignment would also be far ahead of the mainstream on capabilities, because the two are so strongly correlated But isn’t that a one-way relationship? Progressing alignment progresses capabilities, but progressing capabilities doesn’t necessarily strongly progress alignment (otherwise there’d be no problem to begin with). And I guess I still expect that alignment-orthogonal research would progress capabilities faster. (Or, at least, that it’d be faster up to some point. Past that point alignment research might become necessary for further progress… But that point is not necessarily above the level of capabilities that kills everyone.)

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=m2LgaWT3REbxCxGnp

"It is necessary that people working on alignment have a capabilities lead." Could you say a little more about this? Seems true but I’d be curious about your line of thought.

The theory of change could be as simple as "once we know how to build aligned AGI, we’ll tell everybody", or as radical as "once we have an aligned AGI, we can steer the course of human events to prevent future catastrophe". The more boring argument would be that any good ML research happens on the cutting edge of the field, so alignment needs big budgets and fancy labs just like any other researcher. Would you take a specific stance on which is most important?

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=ddERw6YfDwQoF4BPA

Small request: given that it’s plausible that a bunch of LW material on this topic will end up quoted out of context, would you mind changing the headline example in section 5 to something less bad-if-quoted-out-of-context?

Comment

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=oQgSKgnf5FhYA9AWk

Yeah, I’d been worrying about that as well. Though if anything, I’m more concerned about the examples I provide in the second paragraph of that section. The titular one is a bit absurd; the other ones are more plausible. Edited all of that. Better?

Comment

Though if anything, I’m more concerned about the examples I provide in the second paragraph of that section This is what I initially thought Richard’s suggestion was referring to. I was thinking that you’d keep that the same structure minus some of the more extreme examples in the second paragraph. It would have been a much better solution; the "do not start a thermonuclear war" line was pretty funny!

Comment

It was, it was. Unfortunately, we can’t have nice things. I was thinking replacing it with something like "But No Galaxy-Brain Stuff, Alright?" and "radioactive plans" with "galaxy-brain plans", but that seems a bit too tongue-in-cheek/​positive.

Yepp, thanks!

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=DR8jjvfurMkXBPfDv

most academic research work is done by grad students, and grad students need incremental, legible wins to put on their CV so they can prove they are capable of doing research. this has to happen pretty fast. an ML grad student who hasn’t contributed to any top conference papers by their second or third year in grad school might get pulled aside for a talk about their future. ideally you want a topic where you can go from zero to paper in less than a year, with multiple opportunities for followup work. get a few such projects going and you have a very strong chance of getting at least one through in time to not get managed out of your program—and of course, usually more will succeed and you’ll be doing great. I don’t think there’s anything like this in AI safety research. Section 3.4 seems to acknowledge this a little bit. If you want AI safety to become more popular, you’d hope that an incoming PhD student could say "I want to work on AI Safety" and be confident that in a year or two, they’ll have a finished research project that they can claim as a success and submit to a top venue. Otherwise, they are taking a pretty huge career risk, and most people won’t take it.

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=6aQJxmtdFRzJ5CthN

There are some good thoughts here, I like this enough that I am going to comment on the effective strategies angle. You state that

The wider AI research community is an almost-optimal engine of apocalypse.

and

AI capabilities are advancing rapidly, while our attempts to align it proceed at a frustratingly slow pace.

I have to observe that, even though certain people on this forum definitely do believe the above two statements, even on this forum this extreme level of pessimism is a minority opinion. Personally, I have been quite pleased with the pace of progress in alignment research.

This level of disagreement, which is almost inevitable as it involves estimates about about the future. has important implications for the problem of convincing people:

As per above, we’d be fighting an uphill battle here. Researchers and managers are knowledgeable on the subject, have undoubtedly heard about AI risk already, and weren’t convinced.

I’d say that you would indeed be facing an uphill battle, if you’d want to convince most researchers and managers that the recent late-stage Yudkowsky estimates about the inevitability of an AI apocalypse are correct.

The effective framing you are looking for, even if you believe yourself that Yudkowsky is fully correct, is that more work is needed on reducing long-term AI risks. Researchers and managers in the AI industry might agree with you on that, even if they disagree with you and Yudkowsky about other things.

Whether these researchers and managers will change their whole career just because they agree with you is a different matter. Most will not. This is a separate problem, and should be treated as such. Trying to solve both problems at once by making people deeply afraid about the AI apocalypse is a losing strategy.

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=XJju4K2f4ztQ7GdHP

Where’s the money? People who want more AI safety research should be willing to pay for it. Other industries seem to understand that you need to pay for the 90% crap to get the 10% good.

Comment

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=yb9yLAAkxHMNv3DGu

AI safety is a coordination problem, whereas other ML research like the kind OpenAI does can, to a point, be captured for profit by the actors that pioneer it.

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=mkhR9sJ7AuzC58Hhc

This is a wonderful piece and echoes many sentiments I have with the current state of AI safety. Lately, I have also thought more and more about the technical focus’ limitations in the necessary scope to handle the problems of AGI, i.e. the steam engine was an engineering/​tinkering feat loong before it was described technically/​scientifically and ML research seems much the same. When this is the case, focusing purely on hard technical solutions seems less important than focusing on AI governance or prosaic alignment and not doing this, as echoed in other comments, might indeed be a pitfall of specialists, some of which are also warned of here.

https://www.lesswrong.com/posts/mF8dkhZF9hAuLHXaD/reshaping-the-ai-industry?commentId=t2L6DS2HgvmArA9Tj

Thanks for this. There’s been an excess of panic and defeatism here lately, and it’s not good for our chances at success, or our mental health. This is actionable, and feels like it could help.