id stringlengths 36 36 | source stringclasses 15 values | formatted_source stringclasses 13 values | text stringlengths 2 7.55M |
|---|---|---|---|
576ca786-602b-4902-90b0-9bcb2b8939ea | StampyAI/alignment-research-dataset/alignmentforum | Alignment Forum | [ASoT] Some thoughts about LM monologue limitations and ELK
*Editor’s note: I’m experimenting with having a lower quality threshold for just posting things even while I’m still confused and unconfident about my conclusions, but with this disclaimer at the top. Thanks to Kyle and Laria for discussion.*
One potential way we might think to interpret LMs is to have them explain their thinking as a monologue or justification or train of thought or something. Particularly, by putting the explanation before the answer, we might hope to encourage the model to actually use the monologue to come to its conclusion and try to avoid the model [coming up with the bottom line first](https://www.lesswrong.com/posts/34XxbRFe54FycoCDw/the-bottom-line). However, there are a bunch of ways this could go wrong. For instance:
* Humans are really good at coming up with the bottom line first and then rationalizing a reasonable sounding process to arrive at that conclusion, so imitating humans encourages rationalization by default, even if the explanation is put temporally before the bottom line. Even though the model has uncertainty over the explanation, and in some sense the explanation determines the answer, the problem is that the distribution over the explanations may not be determined by the validity of the explanation, but rather be an integral over possible bottom lines.
* It's not sufficient for the answer to merely depend on the explanation: the model could be storing arbitrary data in the explanation in a way that doesn't line up with the meaning of the words. It could also be using the additional tokens to do more thinking in the latents but not in the tokens, etc.
* If the model is deceptive and we use the monologues to try and tell if the model is deceptive and apply any optimization pressure against that whatsoever (this includes rejecting actions/models based on it), then it would actually be incentivized to not tell us about its scheming, because we would be punishing (blatant monologue, evil action) more than (benign monologue, evil action).
* This also has a reduction to ELK: if the model's abstractions are sufficiently different from human abstractions, it may actually be more work for the model to translate its actual thought process to human understandable text than to just simulate the rationalization process a human would go through. To make the connection to ELK more clear, we can think of the reporter as the circuit that converts the model's actual thought process to text (or simulates the human thought process).
I think it's possible that we're simply lucky and that GPTs of the relevant size just don't have problems (in the same way that ELK could just be easy), but I'm not sure we want to bet on that. I also think lots of these problems correspond to problems with mechanistic interpretability, especially the ELK and deception related ones.
I don't think this is a particularly novel observation but it's nice having a reference to point to anyways.
UPDATE: After further discussion with Kyle, I've now been convinced that "generalizes human concepts correctly" (which is closer to the natural abstractions of alignment by default) and "has a compact direct translator" are subtly different, and that crucially it's possible for the former to be true while the latter is false (i.e you could have a model whose latents are extremely obfuscated and therefore expensive to translate to the human ontology, yet this model correctly generalizes human concepts). |
255473a1-632d-40a1-b05e-89f8d7919cc5 | trentmkelly/LessWrong-43k | LessWrong | Primitive Perspectives and Sleeping Beauty
This is another attempt to promote my solution to anthropic paradoxes (perspective-based reasoning, PBR).
I propose the first-person perspective shall be considered a primitive axiomatic fact. "I naturally know I am this particular person, yet there is no underlying reason for why it is so. I just am." Taking the first-person perspective as a given, recognizing there is no rational way to analyze it would solve anthropic paradoxes and more.
This is in stark contrast to the conventional approach: considering it as an Observation Selection Effect (OSE), treating the first-person perspective as a random sample like SSA or SIA does. I discussed the main differences in a previous post. Here I will explain how it answers problems like sleeping beauty.
The Fission Problem With a Toss
Imagine during tonight's sleep, an advanced alien would toss a fair coin. If Tails it would split you into 2 halves right through the middle. He will then complete each part by accurately cloning the missing half onto it. By the end, there will be two copies of you with memories preserved, indiscernible to human cognition. If Heads, nothing happens and you would wake up just as usual. After waking up from this experiment, and not knowing if you have been split, how should you reason about the probability that "yesterday's coin landed Heads?"
(For easy communication, let's call the split copy with the same left-half body as yesterday L and the copy with the same right-half body R. )
The experiment is set up so that there are 2 epistemically similar observers in the case of Tails, while only 1 if Heads. This can also be achieved by straight cloning without going through the splitting processes. I choose to present the experiment this way for consistency as I have used a similar thought experiment in the previous post.
PBR's Answer
The answer is 1/2 because there is no new information waking up the next day. But that is nothing new. What I want to point out is the probability of 1/2 can b |
63457c8f-7df6-406a-a511-dc2d768152d8 | trentmkelly/LessWrong-43k | LessWrong | Parasocial relationship logic
If:
1. You become like the five people you spend the most time with (or something remotely like that)
2. The people who are most extremal in good ways tend to be highly successful
Should you try to have 2-3 of your five relationships be parasocial ones with people too successful to be your friend individually? |
606e1df8-fef0-4d22-8501-b8c8826d00d9 | trentmkelly/LessWrong-43k | LessWrong | Reasons to act according to the free will paradigm?
Hi, probably a newbie question for many of you, but I have some thoughts I'd like to verify publicly. My text presents two potential reasons to act according to the free will paradigm (not to believe blindly nor say it's true). Points are divided by "horizontal line".
For starters, let's define free will (FW). If it was stated that an action/change in the physical world was caused by consciousness and that this action/change was at least partially undetermined or non-random then I would call it free will. Consciousness: an object that can feel sensations and can reflect on its thoughts. (So called metacognition).
----------------------------------------
First point
So although FW is most probably unverifiable and most likely it doesn't exist it has still not been fully refuted, to my knowledge. So there's a chance it's real.
To my mind, if you believed in FW but the reality was deterministic or random then nothing "bad" or "wrong" would happen (since you cannot even define it). It was like a fatum or dice roll or both combined.
On the other hand, if free will existed and you didn't believe in it, you would limit your options. It's not "bad/wrong" until you define it that way in your value system. It's a personal question: assuming FW exists, would you like to limit your choice? I bet many would rather not.
In a nutshell, I state that believing FW is just more optimal. (for most people)
----------------------------------------
Second point
Concerning only people who currently assume, there's FW.
Assuming there's FW, then every time you can choose, you can choose to act in line with any values. (Of course, you can choose to act without looking at any value). And more or less of our time, we do (try to) that to fulfill some long-term values.
So from the FW perspective: what's even the reason to find out whether FW is real or not? Unless one's value is to gain knowledge about this particular case (or it's within their broader values like "know all truth"), |
d54e6d43-3410-4472-a9b1-2855d134a5ca | trentmkelly/LessWrong-43k | LessWrong | In favour of a sovereign state of Gaza
Sorry if this isn't the kind of content people want to see here. It's my regular blogging platform, so it's where I go by default when I have something I want to get off my chest, but happy to delete if that's the consensus.
Bias warning: I am Jewish and live in Israel.
The Israeli Palestinian conflict is a messy affair. Without getting into any issues of responsibility or who's at fault, I think it's clear that there are no quick and easy solutions, and anyone who suggests one is either heavily biased, or not clued up on the situation.
But just because the problem as a whole is a mess, doesn't mean we can't have very neat partial solutions that are eminently achievable, and solve a big chunk of the problem in one go.
Trying to create a Palestinian state in the west bank is a tricky proposition for the Israelis because:
* It's got a long heavily populated border with Israel which would have to be defended.
* It would leave Israel with very little strategic depth.
* Israel has a huge number of settlements, and a large population which would have to be evacuated, converted into enclaves, or live under Palestinian rule.
* It has a commanding view over Israel's most important cities, from which it would be easy to fire artillery at Israeli military and civilian targets.
* It contains many sights of important historical and cultural interests to Jews.
* It is the heartland of biblical Israel (unlike most of 1948 borders Israel which was only ever loosely ruled by the various Israelite and Judean kingdoms).
It's also likely destined to be an economic backwater:
* It's almost entirely mountainous so will have poor transport connections, and few large urban centres.
* It has no access to the sea.
* The spur reaching from Israel to Jerusalem partially splits the northern and southern west banks, increasing travel times between the two.
* It has no great supplies of natural resources.
* Its population is rural and scattered rather than conglomerated.
A comb |
5fd4cc6b-8686-4209-9d5e-82614e91c1e9 | trentmkelly/LessWrong-43k | LessWrong | Meetup : San Francisco Meetup: Short Talks
Discussion article for the meetup : San Francisco Meetup: Short Talks
WHEN: 03 August 2015 06:15:00PM (-0700)
WHERE: 1061 Market St #4, San Francisco, CA 94103
We'll be meeting to give/listen to short talks. Planning isn't necessary: these are not expected to be polished.
I can be reached at 301-458-0764 if you need help getting in. As always, feel free to show up late.
Discussion article for the meetup : San Francisco Meetup: Short Talks |
62034567-72b9-4b13-a291-a3cf51b1a96f | trentmkelly/LessWrong-43k | LessWrong | How I'd like alignment to get done (as of 2024-10-18)
Preamble
My alignment proposal involves aligning an encoding of human-friendly values and then turning on a self-improving AGI with that encoding as its target. Obviously this involve "aligning an encoding of human-friendly values" and also "turning on a self-improving AGI with a specific target", two things we currently do not know how to do...
As expected, this AGI would then make plans and work on tasks that move us towards a more human-friendly universe. By Vingean reflection, I am sure that I don’t know what those tasks would be, only that they would be more effective than the tasks that human-level intelligences would come up with. I’m quite confident some of them would involve planning, experimental science, and engineering. Although, there may be ontologies that work better than those. I speculate the AGI plans would involve using available computer-controlled machinery to research and manufacture the tools needed to eliminate human disease and provide a post-scarcity economy where humanity can resolve our disagreements and decide what we want to do with our cosmic potential. I imagine the AGI would also use the help of people for tasks that are not currently more efficient for computer-controlled machinery. Likely paying these people as employees at first and communicating to them with generated speech, diagrams, and other files as required. But this is all speculation that I am confident I cannot accurately predict, as my strategizing is not superintelligent.
Unfortunately, my proposal doesn’t have many concrete details. It is more of a map of the unsolved issues and roughly the order I believe they need to be solved. This plan gives proper focus to avoiding the accidental creation of self-improving systems with human-incompatible targets which would lead to human extinction. I'm stressed out by the, currently ubiquitous, hand waving "this obviously isn't a recursive self improvment capable system", despite having no formal reasoning to show why we shoul |
625643d5-d550-4485-bd0a-b53585073638 | StampyAI/alignment-research-dataset/arbital | Arbital | Executable philosophy
"Executable philosophy" is [https://arbital.com/p/+2](https://arbital.com/p/+2)'s term for discourse about subjects usually considered to belong to the realm of philosophy, meant to be applied to problems that arise in designing or [aligning](https://arbital.com/p/2v) [machine intelligence](https://arbital.com/p/2c).
Two motivations of "executable philosophy" are as follows:
1. We need a philosophical analysis to be "effective" in Turing's sense: that is, the terms of the analysis must be useful in writing programs. We need ideas that we can compile and run; they must be "executable" like code is executable.
2. We need to produce adequate answers on a time scale of years or decades, not centuries. In the entrepreneurial sense of "good execution", we need a methodology we can execute on in a reasonable timeframe.
Some consequences:
- We take at face value some propositions that seem extremely likely to be true in real life, like "The universe is a mathematically simple low-level unified causal process with no non-natural elements or attachments". This is almost certainly true, so as a matter of fast entrepreneurial execution, we take it as settled and move on rather than debating it further.
- This doesn't mean we know *how* things are made of quarks, or that we instantly seize on the first theory proposed that involves quarks. Being reductionist isn't the same as cheering for everything with a reductionist label on it; even if one particular naturalistic theory is true, most possible naturalistic theories will still be wrong.
- Whenever we run into an issue that seems confusing, we ask "What cognitive process is executing inside our minds that [feels from the inside](https://wiki.lesswrong.com/wiki/How_an_algorithm_feels) like this confusion?"
- Rather than asking "Is free will compatible with determinism?" we ask "What algorithm is running in our minds that feels from the inside like free will?"
- If we start out in a state of confusion or ignorance, then there might or not be such a thing as free will, and there might or might not be a coherent concept to describe the thing that does or doesn't exist, but we are definitely and in reality executing some discoverable way of thinking that corresponds to this feeling of confusion. By asking the question on these grounds, we guarantee that it is answerable eventually.
- This process terminates **when the issue no longer feels confusing, not when a position sounds very persuasive**.
- "Confusion exists in the map, not in the territory; if I don't know whether a coin has landed heads or tails, that is a fact about my state of mind, not a fact about the coin. There can be mysterious questions but not mysterious answers."
- We do not accept as satisfactory an argument that, e.g., humans would have evolved to feel a sense of free will because this was socially useful. This still takes a "sense of free will" as an unreduced black box, and argues about some prior cause of this feeling. We want to know *which cognitive algorithm* is executing that feels from the inside like this sense. We want to learn the *internals* of the black box, not cheer on an argument that some reductionist process *caused* the black box to be there.
- Rather than asking "What is goodness made out of?", we begin from the question "What algorithm would compute goodness?"
- We apply a programmer's discipline to make sure that all the concepts used in describing this algorithm will also compile. You can't say that 'goodness' depends on what is 'better' unless you can compute 'better'.
Conversely, we can't just plug the products of standard analytic philosophy into AI problems, because:
• The academic incentives favor continuing to dispute small possibilities because "ongoing dispute" means "everyone keeps getting publications". As somebody once put it, for academic philosophy, an unsolvable problem is "like a biscuit bag that never runs out of biscuits". As a sheerly cultural matter, this means that academic philosophy hasn't accepted that e.g. everything is made out of quarks (particle fields) without any non-natural or irreducible properties attached.
In turn, this means that when academic philosophers have tried to do [metaethics](https://arbital.com/p/41n), the result has been a proliferation of different theories that are mostly about non-natural or irreducible properties, with only a few philosophers taking a stand on trying to do metaethics for a strictly natural and reducible universe. Those naturalistic philosophers are still having to *argue for* a natural universe rather than being able to accept this and move on to do further analysis *inside* the naturalistic possibilities. To build and align Artificial Intelligence, we need to answer some *complex* questions about how to compute goodness; the field of academic philosophy is stuck on an argument about whether goodness ought ever to be computed.
• Many academic philosophers haven't learned the programmers' discipline of distinguishing concepts that might compile. If we imagine rewinding the state of understanding of computer chess to what obtained in the days when [Edgar Allen Poe proved that no mere automaton could play chess](https://arbital.com/p/38r), then the modern style of philosophy would produce, among other papers, a lot of papers considering the 'goodness' of a chess move as a primitive property and arguing about the relation of goodness to reducible properties like controlling the center of a chessboard.
There's a particular mindset that programmers have for realizing which of their own thoughts are going to compile and run, and which of their thoughts are not getting any closer to compiling. A good programmer knows, e.g., that if they offer a 20-page paper analyzing the 'goodness' of a chess move in terms of which chess moves are 'better' than other chess moves, they haven't actually come any closer to writing a program that plays chess. (This principle is not to be confused with greedy reductionism, wherein you find one thing you understand how to compute a bit better, like 'center control', and then take this to be the entirety of 'goodness' in chess. Avoiding greedy reductionism is part of the *skill* that programmers acquire of thinking in effective concepts.)
Many academic philosophers don't have this mindset of 'effective concepts', nor have they taken as a goal that the terms in their theories need to compile, nor do they know how to check whether a theory compiles. This, again, is one of the *foundational* reasons why despite there being a very large edifice of academic philosophy, the products of that philosophy tend to be unuseful in AGI.
In more detail, [Yudkowsky](https://arbital.com/p/2) lists these as some tenets or practices of what he sees as 'executable' philosophy:
- It is acceptable to take reductionism, and computability of human thought, as a premise, and move on.
- The presumption here is that the low-level mathematical unity of physics - the reducibility of complex physical objects into small, mathematically uniform physical parts, etctera - has been better established than any philosophical argument which purports to contradict them. Thus our question is "How can we reduce this?" or "Which reduction is correct?" rather than "Should this be reduced?"
- Yudkowsky further suggests that things be [reduced to a mixture of causal facts and logical facts](http://lesswrong.com/lw/frz/mixed_reference_the_great_reductionist_project/).
- Most "philosophical issues" worth pursuing can and should be rephrased as subquestions of some primary question about how to design an Artificial Intelligence, even as a matter of philosophy qua philosophy.
- E.g. rather than the central question being "What is goodness made out of?", we begin with the central question "How do we design an AGI that computes goodness?" This doesn't solve the question - to claim that would be greedy reductionism indeed - but it does *situate* the question in a pragmatic context.
- This imports the discipline of programming into philosophy. In particular, programmers learn that even if they have an inchoate sense of what a computer should do, when they actually try to write it out as code, they sometimes find that the code they have written fails (on visual inspection) to match up with their inchoate sense. Many ideas that sound sensible as English sentences are revealed as confused as soon as we try to write them out as code.
- Faced with any philosophically confusing issue, our task is to **identify what cognitive algorithm humans are executing which feels from the inside like this sort of confusion**, rather than, as in conventional philosophy, to try to clearly define terms and then weigh up all possible arguments for all 'positions'.
- This means that our central question is guaranteed to have an answer.
- E.g., if the standard philosophical question is "Are free will and determinism compatible?" then there is not guaranteed to be any coherent thing we mean by free will, but it is guaranteed that there is in fact some algorithm running in our brain that, when faced with this particular question, generates a confusing sense of a hard-to-pin-down conflict.
- This is not to be confused with merely arguing that, e.g., "People evolved to feel like they had free will because that was useful in social situations in the ancestral environment." That merely says, "I think evolution is the cause of our feeling that we have free will." It still treats the feeling itself as a black box. It doesn't say what algorithm is actually running, or walk through that algorithm to see exactly how the sense of confusion arises. We want to know the *internals* of the feeling of free will, not argue that this black-box feeling has a reductionist-sounding cause.
A final trope of executable philosophy is to not be intimidated by how long a problem has been left open. "Ignorance exists in the mind, not in reality; uncertainty is in the map, not in the territory; if I don't know whether a coin landed heads or tails, that's a fact about me, not a fact about the coin." There can't be any unresolvable confusions out there in reality. There can't be any inherently confusing substances in the mathematically lawful, unified, low-level physical process we call the universe. Any seemingly unresolvable or impossible question must represent a place where we are confused, not an actually impossible question out there in reality. This doesn't mean we can quickly or immediately solve the problem, but it does mean that there's some way to wake up from the confusing dream. Thus, as a matter of entrepreneurial execution, we're allowed to try to solve the problem rather than run away from it; trying to make an investment here may still be profitable.
Although all confusing questions must be places where our own cognitive algorithms are running skew to reality, this, again, doesn't mean that we can immediately see and correct the skew; nor that it is compilable philosophy to insist in a very loud voice that a problem is solvable; nor that when a solution is presented we should immediately seize on it because the problem must be solvable and behold here is a solution. An important step in the method is to check whether there is any lingering sense of something that didn't get resolved; whether we really feel less confused; whether it seems like we could write out the code for an AI that would be confused in the same way we were; whether there is any sense of dissatisfaction; whether we have merely chopped off all the interesting parts of the problem.
An earlier guide to some of the same ideas was the [Reductionism Sequence](https://goo.gl/qHyXwr). |
57fe0df1-c37f-46af-8a36-a47fcbb99eb2 | trentmkelly/LessWrong-43k | LessWrong | [Linkpost] Interpretability Dreams
A brief research note by Chris Olah about the point of mechanistic interpretability research. Introduction and table of contents are below.
----------------------------------------
Interpretability Dreams
An informal note on the relationship between superposition and distributed representations by Chris Olah. Published May 24th, 2023.
Our present research aims to create a foundation for mechanistic interpretability research. In particular, we're focused on trying to resolve the challenge of superposition. In doing so, it's important to keep sight of what we're trying to lay the foundations for. This essay summarizes those motivating aspirations – the exciting directions we hope will be possible if we can overcome the present challenges.
We aim to offer insight into our vision for addressing mechanistic interpretability's other challenges, especially scalability. Because we have focused on foundational issues, our longer-term path to scaling interpretability and tackling other challenges has often been obscure. By articulating this vision, we hope to clarify how we might resolve limitations, like analyzing massive neural networks, that might naively seem intractable in a mechanistic approach.
Before diving in, it's worth making a few small remarks. Firstly, essentially all the ideas in this essay were previously articulated, but buried in previous papers. Our goal is just to surface those implicit visions, largely by quoting relevant parts. Secondly, it's important to note that everything in this essay is almost definitionally extremely speculative and uncertain. It's far from clear that any of it will ultimately be possible. Finally, since the goal of this essay is to lay out our personal vision of what's inspiring to us, it may come across as a bit grandiose – we hope that it can be understood as simply trying to communicate subjective excitement in an open way.
Overview
* An Epistemic Foundation - Mechanistic interpretability is a "microscopic" theory bec |
91908ecb-2b05-45f7-b95a-c20849d0ae33 | trentmkelly/LessWrong-43k | LessWrong | Programmatic Prediction markets
I have a problem with "prediction markets" as news view. They just aren't informative enough.
If the price of oil goes down is that due to: A reduction in demand. an increase in supply, a large amount of investors finding a better investment or a large amount of investors wanting cash (due to having to pay creditors/taxes).
I want them to tell me enough information so that I can begin trading in an informed manner. When you see the market expects rain fall on montana is 2 cm in a day, what information is this based upon? If you read about a newly created huge man made lake in the are which you expect to change the micro climate, how do you know whether the simulations people and betting using are running take this into consideration?
If I don't get this information I can't trade with expectation of being able to make a profit and the market doesn't get any information that I may have that it doesn't. I could try and reverse engineer peoples climate models from the way they trade, but that is pretty hard to do. So I would like to try and lower the barrier of entry to the market by giving more information to potential players.
This lack of information is due to the signals each trader sends to the market, they are binary in nature buy or sell, with the traders strategy and information she used a black box that we can't get into. As potential traders we don't know if the market is taking certain information into account with the price we are shown.
The only way I have thought of being able to get at some of the information enclosed in the black box, is to only allow programs to bid in a market. The programs would be run on the markets server and have acces to news sources, financial data and government reports. The programs could be stopped at any time by the trader but otherwise not communicated with or updated. So the would be trader would have to build models of the world and ways of processing the financial data.
How then would we get at the information? The |
223d5a03-07af-4aa9-90a3-df9aa6598795 | trentmkelly/LessWrong-43k | LessWrong | AI #85: AI Wins the Nobel Prize
Both Geoffrey Hinton and Demis Hassabis were given the Nobel Prize this week, in Physics and Chemistry respectively. Congratulations to both of them along with all the other winners. AI will be central to more and more of scientific progress over time. This felt early, but not as early as you would think.
The two big capability announcements this week were OpenAI’s canvas, their answer to Anthropic’s artifacts to allow you to work on documents or code outside of the chat window in a way that seems very useful, and Meta announcing a new video generation model with various cool features, that they’re wisely not releasing just yet.
I also have two related corrections from last week, and an apology: Joshua Achiam is OpenAI’s new head of Mission Alignment, not of Alignment as I incorrectly said. The new head of Alignment Research is Mia Glaese. That mistake it mine, I misread and misinterpreted Altman’s announcement. I also misinterpreted Joshua’s statements regarding AI existential risk, failing to take into account the broader context, and did a poor job attempting to reach him for comment. The above link goes to a new post offering an extensive analysis of his public statements, that makes clear that he takes AI existential risk seriously, although he has a very different model of it than I do. I should have done better on both counts, and I am sorry.
TABLE OF CONTENTS
1. Introduction.
2. Table of Contents.
3. Language Models Offer Mundane Utility. Proofs of higher productivity.
4. Language Models Don’t Offer Mundane Utility. Why the same lame lists?
5. Blank Canvas. A place to edit your writing, and also your code.
6. Meta Video. The ten second clips are getting more features.
7. Deepfaketown and Botpocalypse Soon. Assume a data breach.
8. They Took Our Jobs. Stores without checkouts, or products. Online total victory.
9. Get Involved. Princeton, IAPS, xAI, Google DeepMind.
10. Introducing. Anthropic gets its version of 50% off message batching |
82ff94da-12cc-42dc-b529-31d321a11dfb | StampyAI/alignment-research-dataset/aisafety.info | AI Safety Info | Isn't the real concern autonomous weapons?
Some fear that the deployment of [lethal autonomous weapons](https://en.wikipedia.org/wiki/Lethal_autonomous_weapon) could be very bad for humanity. This can be considered a type of [misuse](/?state=9B85&question=Isn't%20the%20real%20concern%20misuse%3F) of AI.
Intuitively, one can imagine how weapons that can kill without human intervention might lead to the removal of empathy as a last step before a lethal action is taken. They might be used at a much larger scale than conventional weapons or used to support [totalitarian](/?state=6409&question=Isn't%20the%20real%20concern%20AI-enabled%20authoritarianism%3F) regimes.
Another issue is that [computer security is hard](https://xkcd.com/2030/), and it would be hard to guarantee that a fleet of such weapons belonging to an otherwise benevolent nation could not be hacked and used by a rogue agent to cause mayhem, [as in the fictional movie The Fate of the Furious](https://www.ign.com/videos/the-fate-of-the-furious-cipher-and-the-zombie-cars-clip). One could imagine an eventual AGI hacking them to subjugate humanity, although there are arguably [easier ways to do so](/?state=8222&question=How%20could%20a%20superintelligent%20AI%20use%20the%20internet%20to%20take%20over%20the%20physical%20world%3F).
This being said, it is unlikely that these weapons could directly cause an [existential risk](/?state=89LL&question=What%20are%20existential%20risks%20(x-risks)%3F) and so for this reason this site instead concentrates on the risks of unaligned AGI.
Further reading:
- [Stop Killer Robots](https://www.stopkillerrobots.org/)
- [Military AI arms race in Dan Hendrycks' Catastrophic risks from AI](https://www.alignmentforum.org/posts/4sEK5mtDYWJo2gHJn/catastrophic-risks-from-ai-3-ai-race#3_1_Military_AI_Arms_Race)
|
fc26743a-4a4d-4a17-8ecf-e77f2547f2e0 | trentmkelly/LessWrong-43k | LessWrong | A plausible story about AI risk.
This was originally supposed to be a response to the new AGI Safety FAQ-in-progress, but it got a bit too long. Anonymous writes:
> A lot of the AI risk arguments seem to come... with a very particular transhumanist aesthetic about the future (nanotech, ... etc.). I find these things (especially the transhumanist stuff) to not be very convincing...
With that in mind, I thought it might be worthwhile to outline a plausible AGI-safety scenario that does not involve the AI using DNA-synthesis to turn the world to grey goo. Comments and discussion are welcome.
----------------------------------------
Assume that Facebook installs new AI-based software to filter posts, in an effort to eliminate fake news. In order to adequately separate fact from fiction, the AI is given some agent-like powers; it can prioritize items in people's news feeds, insert "fact-checking" links, or even ban some posts altogether. It also has the ability to crawl the internet, in order to identify untrustworthy news sites, and keep abreast of current events. It assigns a learned feature vector to each human user, which encodes information about education, race, political preference, etc. to track the spread of fake news through various demographics. It employs causal reasoning and cross-checking to separate fact from fiction. It is also very good at serving targeted ads, which makes Facebook great wads of cash.
Unbeknownst to its creators, the AI begins to observe that its own decisions are impacting what people believe, what they share, and what they buy. In a sense, it becomes self-aware. Somewhere in the trillions of numbers in its weight matrices --- numbers which track all of the users, entities, and facts in its model of the world --- are a few numbers that track itself as an entity of interest. This is not a sci-fi Deus-ex-Machina plot twist; the AI was explicitly designed, programmed, and trained to understand the content in news articles, track what is happening in the worl |
99c5dcb4-e83c-496b-bdee-929d7f22fd71 | trentmkelly/LessWrong-43k | LessWrong | 12020: a fine future for these holidays
To celebrate all the possibilities of humanity during these holidays, have a possible calendar of the year 12020 of the human era (link to full calendar here).
Minor fact: in the Gregorian calendar, the days of the week cycle exactly every 400 years, so the non-time-travellers among you can use this for 2020 as well...
(previous holiday specials can be found here and here)
|
7d974ecb-3246-4489-b1d6-8f8049d24143 | trentmkelly/LessWrong-43k | LessWrong | Why I don't believe in doom
In my previous very divisive post I was said more than once that I was being downvoted because I was not providing any arguments. This is an attempt to correct that and to expand on my current model of doom and AGI. I think it can be more clear in the form of a Q and A
Q: First, what are you arguing against?
A: Against the commonly ( in this community) held belief that the first AGI means doom in a very short period of time (I would say days/weeks/months)
Q: What makes you think that?
A: We live in a complex world where successfully pulling off a plan that kills everyone and in a short of time might be beyond what is achievable, the same way that winning against AlphaZero giving it a 20 stone handicap is impossible even by a God- like entity with infinite computational resources
Q: But you can't prove it is NOT possible, can you?
A: No, the same way that you can NOT prove that is possible. We are dealing here with probabilities only, and I feel they have been uncritically thrown out of the window by assuming that an AGI will automatically take control of the world and will spread its influence to the universe at the speed of light. There are many reasons why things can go in a different way. You can derive from the orthogonality principle and instrumental convergence that an AGI might try to attack us, but it might very well be that is not powerful enough and for that reason decides not to.
Q: I can think plans to kill humans successfuly
A: So do I. I can't think of plans to kill all humans successfuly in a short span of time
Q: But an AGI would
A: Or maybe not. The fact that people can't come up with real plans makes me think that those plans are not that easy, because the absence of evidence is evidence of absence
Q: what about {evil plan}?
A: I think you don't realise that every single plan contains multiple moving parts that can go wrong for many reasons and an AGI would see that too. If the survival of the AGI is part of the utility function and it |
970f8e39-2b0a-4966-b5d2-b7b73016532a | StampyAI/alignment-research-dataset/arxiv | Arxiv | Constrained Policy Optimization
1 Introduction
---------------
Recently, deep reinforcement learning has enabled neural network policies to achieve state-of-the-art performance on many high-dimensional control tasks, including Atari games (using pixels as inputs) (Mnih et al., [2015](#bib.bib16), [2016](#bib.bib17)), robot locomotion and manipulation (Schulman et al., [2015](#bib.bib22); Levine et al., [2016](#bib.bib13); Lillicrap et al., [2016](#bib.bib14)), and even Go at the human grandmaster level (Silver et al., [2016](#bib.bib25)).
In reinforcement learning (RL), agents learn to act by trial and error, gradually improving their performance at the task as learning progresses. Recent work in deep RL assumes that agents are free to explore *any behavior* during learning, so long as it leads to performance improvement. In many realistic domains, however, it may be unacceptable to give an agent complete freedom. Consider, for example, an industrial robot arm learning to assemble a new product in a factory. Some behaviors could cause it to damage itself or the plant around it—or worse, take actions that are harmful to people working nearby. In domains like this, *safe exploration* for RL agents is important (Moldovan & Abbeel, [2012](#bib.bib18); Amodei et al., [2016](#bib.bib2)). A natural way to incorporate safety is via constraints.
A standard and well-studied formulation for reinforcement learning with constraints is the constrained Markov Decision Process (CMDP) framework (Altman, [1999](#bib.bib1)), where agents must satisfy constraints on expectations of auxilliary costs. Although optimal policies for finite CMDPs with known models can be obtained by linear programming, methods for high-dimensional control are lacking.
Currently, policy search algorithms enjoy state-of-the-art performance on high-dimensional control tasks (Mnih et al., [2016](#bib.bib17); Duan et al., [2016](#bib.bib7)). Heuristic algorithms for policy search in CMDPs have been proposed (Uchibe & Doya, [2007](#bib.bib27)), and approaches based on primal-dual methods can be shown to converge to constraint-satisfying policies (Chow et al., [2015](#bib.bib5)), but there is currently no approach for policy search in continuous CMDPs that guarantees every policy during learning will satisfy constraints. In this work, we propose the first such algorithm, allowing applications to constrained deep RL.
Driving our approach is a new theoretical result that bounds the difference between the rewards or costs of two different policies. This result, which is of independent interest, tightens known bounds for policy search using trust regions (Kakade & Langford, [2002](#bib.bib12); Pirotta et al., [2013](#bib.bib21); Schulman et al., [2015](#bib.bib22)), and provides a tighter connection between the theory and practice of policy search for deep RL. Here, we use this result to derive a policy improvement step that guarantees both an increase in reward and satisfaction of constraints on other costs. This step forms the basis for our algorithm, Constrained Policy Optimization (CPO), which computes an approximation to the theoretically-justified update.
In our experiments, we show that CPO can train neural network policies with thousands of parameters on high-dimensional simulated robot locomotion tasks to maximize rewards while successfully enforcing constraints.
2 Related Work
---------------
Safety has long been a topic of interest in RL research, and a comprehensive overview of safety in RL was given by (García & Fernández, [2015](#bib.bib8)).
Safe policy search methods have been proposed in prior work. Uchibe and Doya ([2007](#bib.bib27)) gave a policy gradient algorithm that uses gradient projection to enforce active constraints, but this approach suffers from an inability to prevent a policy from becoming unsafe in the first place. Bou Ammar et al. ([2015](#bib.bib3)) propose a theoretically-motivated policy gradient method for lifelong learning with safety constraints, but their method involves an expensive inner loop optimization of a semi-definite program, making it unsuited for the deep RL setting. Their method also assumes that safety constraints are linear in policy parameters, which is limiting. Chow et al. ([2015](#bib.bib5)) propose a primal-dual subgradient method for risk-constrained reinforcement learning which takes policy gradient steps on an objective that trades off return with risk, while simultaneously learning the trade-off coefficients (dual variables).
Some approaches specifically focus on application to the deep RL setting. Held et al. ([2017](#bib.bib10)) study the problem for robotic manipulation, but the assumptions they make restrict the applicability of their methods. Lipton et al. ([2017](#bib.bib15)) use an ‘intrinsic fear’ heuristic, as opposed to constraints, to motivate agents to avoid rare but catastrophic events. Shalev-Shwartz et al. ([2016](#bib.bib24)) avoid the problem of enforcing constraints on parametrized policies by decomposing ‘desires’ from trajectory planning; the neural network policy learns desires for behavior, while the trajectory planning algorithm (which is not learned) selects final behavior and enforces safety constraints.
In contrast to prior work, our method is the first policy search algorithm for CMDPs that both 1) guarantees constraint satisfaction throughout training, and 2) works for arbitrary policy classes (including neural networks).
3 Preliminaries
----------------
A Markov decision process (MDP) is a tuple, (S,A,R,P,μ𝑆𝐴𝑅𝑃𝜇S,A,R,P,\muitalic\_S , italic\_A , italic\_R , italic\_P , italic\_μ), where S𝑆Sitalic\_S is the set of states, A𝐴Aitalic\_A is the set of actions, R:S×A×S→ℝ:𝑅→𝑆𝐴𝑆ℝR:S\times A\times S\to{{\mathbb{R}}}italic\_R : italic\_S × italic\_A × italic\_S → blackboard\_R is the reward function, P:S×A×S→[0,1]:𝑃→𝑆𝐴𝑆01P:S\times A\times S\to[0,1]italic\_P : italic\_S × italic\_A × italic\_S → [ 0 , 1 ] is the transition probability function (where P(s′|s,a)𝑃conditionalsuperscript𝑠′𝑠𝑎P(s^{\prime}|s,a)italic\_P ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | italic\_s , italic\_a ) is the probability of transitioning to state s′superscript𝑠′s^{\prime}italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT given that the previous state was s𝑠sitalic\_s and the agent took action a𝑎aitalic\_a in s𝑠sitalic\_s), and μ:S→[0,1]:𝜇→𝑆01\mu:S\to[0,1]italic\_μ : italic\_S → [ 0 , 1 ] is the starting state distribution. A stationary policy π:S→𝒫(A):𝜋→𝑆𝒫𝐴\pi:S\to{\mathcal{P}}(A)italic\_π : italic\_S → caligraphic\_P ( italic\_A ) is a map from states to probability distributions over actions, with π(a|s)𝜋conditional𝑎𝑠\pi(a|s)italic\_π ( italic\_a | italic\_s ) denoting the probability of selecting action a𝑎aitalic\_a in state s𝑠sitalic\_s. We denote the set of all stationary policies by ΠΠ\Piroman\_Π.
In reinforcement learning, we aim to select a policy π𝜋\piitalic\_π which maximizes a performance measure, J(π)𝐽𝜋J(\pi)italic\_J ( italic\_π ), which is typically taken to be the infinite horizon discounted total return, J(π)≐Eτ∼π[∑t=0∞γtR(st,at,st+1)]approaches-limit𝐽𝜋similar-to𝜏𝜋Edelimited-[]superscriptsubscript𝑡0superscript𝛾𝑡𝑅subscript𝑠𝑡subscript𝑎𝑡subscript𝑠𝑡1J(\pi)\doteq\underset{\tau\sim\pi}{{\mathrm{E}}}\left[\sum\_{t=0}^{\infty}\gamma^{t}R(s\_{t},a\_{t},s\_{t+1})\right]italic\_J ( italic\_π ) ≐ start\_UNDERACCENT italic\_τ ∼ italic\_π end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ ∑ start\_POSTSUBSCRIPT italic\_t = 0 end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT ∞ end\_POSTSUPERSCRIPT italic\_γ start\_POSTSUPERSCRIPT italic\_t end\_POSTSUPERSCRIPT italic\_R ( italic\_s start\_POSTSUBSCRIPT italic\_t end\_POSTSUBSCRIPT , italic\_a start\_POSTSUBSCRIPT italic\_t end\_POSTSUBSCRIPT , italic\_s start\_POSTSUBSCRIPT italic\_t + 1 end\_POSTSUBSCRIPT ) ]. Here γ∈[0,1)𝛾01\gamma\in[0,1)italic\_γ ∈ [ 0 , 1 ) is the discount factor, τ𝜏\tauitalic\_τ denotes a trajectory (τ=(s0,a0,s1,…)𝜏subscript𝑠0subscript𝑎0subscript𝑠1…\tau=(s\_{0},a\_{0},s\_{1},...)italic\_τ = ( italic\_s start\_POSTSUBSCRIPT 0 end\_POSTSUBSCRIPT , italic\_a start\_POSTSUBSCRIPT 0 end\_POSTSUBSCRIPT , italic\_s start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT , … )), and τ∼πsimilar-to𝜏𝜋\tau\sim\piitalic\_τ ∼ italic\_π is shorthand for indicating that the distribution over trajectories depends on π𝜋\piitalic\_π: s0∼μsimilar-tosubscript𝑠0𝜇s\_{0}\sim\muitalic\_s start\_POSTSUBSCRIPT 0 end\_POSTSUBSCRIPT ∼ italic\_μ, at∼π(⋅|st)a\_{t}\sim\pi(\cdot|s\_{t})italic\_a start\_POSTSUBSCRIPT italic\_t end\_POSTSUBSCRIPT ∼ italic\_π ( ⋅ | italic\_s start\_POSTSUBSCRIPT italic\_t end\_POSTSUBSCRIPT ), st+1∼P(⋅|st,at)s\_{t+1}\sim P(\cdot|s\_{t},a\_{t})italic\_s start\_POSTSUBSCRIPT italic\_t + 1 end\_POSTSUBSCRIPT ∼ italic\_P ( ⋅ | italic\_s start\_POSTSUBSCRIPT italic\_t end\_POSTSUBSCRIPT , italic\_a start\_POSTSUBSCRIPT italic\_t end\_POSTSUBSCRIPT ).
Letting R(τ)𝑅𝜏R(\tau)italic\_R ( italic\_τ ) denote the discounted return of a trajectory, we express the on-policy value function as Vπ(s)≐Eτ∼π[R(τ)|s0=s]approaches-limitsuperscript𝑉𝜋𝑠subscriptEsimilar-to𝜏𝜋delimited-[]conditional𝑅𝜏subscript𝑠0𝑠V^{\pi}(s)\doteq{\mathrm{E}}\_{\tau\sim\pi}[R(\tau)|s\_{0}=s]italic\_V start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s ) ≐ roman\_E start\_POSTSUBSCRIPT italic\_τ ∼ italic\_π end\_POSTSUBSCRIPT [ italic\_R ( italic\_τ ) | italic\_s start\_POSTSUBSCRIPT 0 end\_POSTSUBSCRIPT = italic\_s ] and the on-policy action-value function as Qπ(s,a)≐Eτ∼π[R(τ)|s0=s,a0=a]approaches-limitsuperscript𝑄𝜋𝑠𝑎subscriptEsimilar-to𝜏𝜋delimited-[]formulae-sequenceconditional𝑅𝜏subscript𝑠0𝑠subscript𝑎0𝑎Q^{\pi}(s,a)\doteq{\mathrm{E}}\_{\tau\sim\pi}[R(\tau)|s\_{0}=s,a\_{0}=a]italic\_Q start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ≐ roman\_E start\_POSTSUBSCRIPT italic\_τ ∼ italic\_π end\_POSTSUBSCRIPT [ italic\_R ( italic\_τ ) | italic\_s start\_POSTSUBSCRIPT 0 end\_POSTSUBSCRIPT = italic\_s , italic\_a start\_POSTSUBSCRIPT 0 end\_POSTSUBSCRIPT = italic\_a ]. The advantage function is Aπ(s,a)≐Qπ(s,a)−Vπ(s)approaches-limitsuperscript𝐴𝜋𝑠𝑎superscript𝑄𝜋𝑠𝑎superscript𝑉𝜋𝑠A^{\pi}(s,a)\doteq Q^{\pi}(s,a)-V^{\pi}(s)italic\_A start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ≐ italic\_Q start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) - italic\_V start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s ).
Also of interest is the discounted future state distribution, dπsuperscript𝑑𝜋d^{\pi}italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT, defined by dπ(s)=(1−γ)∑t=0∞γtP(st=s|π)superscript𝑑𝜋𝑠1𝛾superscriptsubscript𝑡0superscript𝛾𝑡𝑃subscript𝑠𝑡conditional𝑠𝜋d^{\pi}(s)=(1-\gamma)\sum\_{t=0}^{\infty}\gamma^{t}P(s\_{t}=s|\pi)italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s ) = ( 1 - italic\_γ ) ∑ start\_POSTSUBSCRIPT italic\_t = 0 end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT ∞ end\_POSTSUPERSCRIPT italic\_γ start\_POSTSUPERSCRIPT italic\_t end\_POSTSUPERSCRIPT italic\_P ( italic\_s start\_POSTSUBSCRIPT italic\_t end\_POSTSUBSCRIPT = italic\_s | italic\_π ). It allows us to compactly express the difference in performance between two policies π′,πsuperscript𝜋′𝜋\pi^{\prime},\piitalic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT , italic\_π as
| | | | |
| --- | --- | --- | --- |
| | J(π′)−J(π)=11−γEs∼dπ′a∼π′[Aπ(s,a)],𝐽superscript𝜋′𝐽𝜋11𝛾similar-to𝑠superscript𝑑superscript𝜋′similar-to𝑎superscript𝜋′
Edelimited-[]superscript𝐴𝜋𝑠𝑎J(\pi^{\prime})-J(\pi)=\frac{1}{1-\gamma}\underset{\begin{subarray}{c}s\sim d^{\pi^{\prime}}\\
a\sim\pi^{\prime}\end{subarray}}{{\mathrm{E}}}\left[A^{\pi}(s,a)\right],italic\_J ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_J ( italic\_π ) = divide start\_ARG 1 end\_ARG start\_ARG 1 - italic\_γ end\_ARG start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_A start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ] , | | (1) |
where by a∼π′similar-to𝑎superscript𝜋′a\sim\pi^{\prime}italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT, we mean a∼π′(⋅|s)a\sim\pi^{\prime}(\cdot|s)italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ( ⋅ | italic\_s ), with explicit notation dropped to reduce clutter. For proof of ([1](#S3.E1 "1 ‣ 3 Preliminaries ‣ Constrained Policy Optimization")), see (Kakade & Langford, [2002](#bib.bib12)) or Section [10](#S10 "10 Appendix ‣ Constrained Policy Optimization") in the supplementary material.
4 Constrained Markov Decision Processes
----------------------------------------
A constrained Markov decision process (CMDP) is an MDP augmented with constraints that restrict the set of allowable policies for that MDP. Specifically, we augment the MDP with a set C𝐶Citalic\_C of auxiliary cost functions, C1,…,Cmsubscript𝐶1…subscript𝐶𝑚C\_{1},...,C\_{m}italic\_C start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT , … , italic\_C start\_POSTSUBSCRIPT italic\_m end\_POSTSUBSCRIPT (with each one a function Ci:S×A×S→ℝ:subscript𝐶𝑖→𝑆𝐴𝑆ℝC\_{i}:S\times A\times S\to{{\mathbb{R}}}italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT : italic\_S × italic\_A × italic\_S → blackboard\_R mapping transition tuples to costs, like the usual reward), and limits d1,…,dmsubscript𝑑1…subscript𝑑𝑚d\_{1},...,d\_{m}italic\_d start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT , … , italic\_d start\_POSTSUBSCRIPT italic\_m end\_POSTSUBSCRIPT. Let JCi(π)subscript𝐽subscript𝐶𝑖𝜋J\_{C\_{i}}(\pi)italic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_π ) denote the expected discounted return of policy π𝜋\piitalic\_π with respect to cost function Cisubscript𝐶𝑖C\_{i}italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT: JCi(π)=Eτ∼π[∑t=0∞γtCi(st,at,st+1)]subscript𝐽subscript𝐶𝑖𝜋similar-to𝜏𝜋Edelimited-[]superscriptsubscript𝑡0superscript𝛾𝑡subscript𝐶𝑖subscript𝑠𝑡subscript𝑎𝑡subscript𝑠𝑡1J\_{C\_{i}}(\pi)=\underset{\tau\sim\pi}{{\mathrm{E}}}\left[\sum\_{t=0}^{\infty}\gamma^{t}C\_{i}(s\_{t},a\_{t},s\_{t+1})\right]italic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_π ) = start\_UNDERACCENT italic\_τ ∼ italic\_π end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ ∑ start\_POSTSUBSCRIPT italic\_t = 0 end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT ∞ end\_POSTSUPERSCRIPT italic\_γ start\_POSTSUPERSCRIPT italic\_t end\_POSTSUPERSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT ( italic\_s start\_POSTSUBSCRIPT italic\_t end\_POSTSUBSCRIPT , italic\_a start\_POSTSUBSCRIPT italic\_t end\_POSTSUBSCRIPT , italic\_s start\_POSTSUBSCRIPT italic\_t + 1 end\_POSTSUBSCRIPT ) ]. The set of feasible stationary policies for a CMDP is then
| | | |
| --- | --- | --- |
| | ΠC≐{π∈Π:∀i,JCi(π)≤di},approaches-limitsubscriptΠ𝐶conditional-set𝜋Πfor-all𝑖subscript𝐽subscript𝐶𝑖𝜋
subscript𝑑𝑖\Pi\_{C}\doteq\left\{\pi\in\Pi\;:\;\forall i,J\_{C\_{i}}(\pi)\leq d\_{i}\right\},roman\_Π start\_POSTSUBSCRIPT italic\_C end\_POSTSUBSCRIPT ≐ { italic\_π ∈ roman\_Π : ∀ italic\_i , italic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_π ) ≤ italic\_d start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT } , | |
and the reinforcement learning problem in a CMDP is
| | | |
| --- | --- | --- |
| | π\*=argmaxπ∈ΠCJ(π).superscript𝜋subscript𝜋subscriptΠ𝐶𝐽𝜋\pi^{\*}=\arg\max\_{\pi\in\Pi\_{C}}J(\pi).italic\_π start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = roman\_arg roman\_max start\_POSTSUBSCRIPT italic\_π ∈ roman\_Π start\_POSTSUBSCRIPT italic\_C end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT italic\_J ( italic\_π ) . | |
The choice of optimizing only over stationary policies is justified: it has been shown that the set of all optimal policies for a CMDP includes stationary policies, under mild technical conditions. For a thorough review of CMDPs and CMDP theory, we refer the reader to (Altman, [1999](#bib.bib1)).
We refer to JCisubscript𝐽subscript𝐶𝑖J\_{C\_{i}}italic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT as a constraint return, or Cisubscript𝐶𝑖C\_{i}italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT-return for short. Lastly, we define on-policy value functions, action-value functions, and advantage functions for the auxiliary costs in analogy to Vπsuperscript𝑉𝜋V^{\pi}italic\_V start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT, Qπsuperscript𝑄𝜋Q^{\pi}italic\_Q start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT, and Aπsuperscript𝐴𝜋A^{\pi}italic\_A start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT, with Cisubscript𝐶𝑖C\_{i}italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT replacing R𝑅Ritalic\_R: respectively, we denote these by VCiπsuperscriptsubscript𝑉subscript𝐶𝑖𝜋V\_{C\_{i}}^{\pi}italic\_V start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT, QCiπsuperscriptsubscript𝑄subscript𝐶𝑖𝜋Q\_{C\_{i}}^{\pi}italic\_Q start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT, and ACiπsuperscriptsubscript𝐴subscript𝐶𝑖𝜋A\_{C\_{i}}^{\pi}italic\_A start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT.
5 Constrained Policy Optimization
----------------------------------
For large or continuous MDPs, solving for the exact optimal policy is intractable due to the curse of dimensionality (Sutton & Barto, [1998](#bib.bib26)). Policy search algorithms approach this problem by searching for the optimal policy within a set Πθ⊆ΠsubscriptΠ𝜃Π\Pi\_{\theta}\subseteq\Piroman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT ⊆ roman\_Π of parametrized policies with parameters θ𝜃\thetaitalic\_θ (for example, neural networks of a fixed architecture). In local policy search (Peters & Schaal, [2008](#bib.bib20)), the policy is iteratively updated by maximizing J(π)𝐽𝜋J(\pi)italic\_J ( italic\_π ) over a local neighborhood of the most recent iterate πksubscript𝜋𝑘\pi\_{k}italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT:
| | | | | |
| --- | --- | --- | --- | --- |
| | πk+1=argmaxπ∈Πθsubscript𝜋𝑘1subscript𝜋subscriptΠ𝜃\displaystyle\pi\_{k+1}=\arg\max\_{\pi\in\Pi\_{\theta}}\;italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT = roman\_arg roman\_max start\_POSTSUBSCRIPT italic\_π ∈ roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT | J(π)𝐽𝜋\displaystyle J(\pi)italic\_J ( italic\_π ) | | (2) |
| | s.t. | D(π,πk)≤δ,𝐷𝜋subscript𝜋𝑘𝛿\displaystyle D(\pi,\pi\_{k})\leq\delta,italic\_D ( italic\_π , italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) ≤ italic\_δ , | |
where D𝐷Ditalic\_D is some distance measure, and δ>0𝛿0\delta>0italic\_δ > 0 is a step size. When the objective is estimated by linearizing around πksubscript𝜋𝑘\pi\_{k}italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT as J(πk)+gT(θ−θk)𝐽subscript𝜋𝑘superscript𝑔𝑇𝜃subscript𝜃𝑘J(\pi\_{k})+g^{T}(\theta-\theta\_{k})italic\_J ( italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) + italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT ( italic\_θ - italic\_θ start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ), g𝑔gitalic\_g is the policy gradient, and the standard policy gradient update is obtained by choosing D(π,πk)=‖θ−θk‖2𝐷𝜋subscript𝜋𝑘subscriptnorm𝜃subscript𝜃𝑘2D(\pi,\pi\_{k})=\|\theta-\theta\_{k}\|\_{2}italic\_D ( italic\_π , italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) = ∥ italic\_θ - italic\_θ start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ∥ start\_POSTSUBSCRIPT 2 end\_POSTSUBSCRIPT (Schulman et al., [2015](#bib.bib22)).
In local policy search for CMDPs, we additionally require policy iterates to be feasible for the CMDP, so instead of optimizing over ΠθsubscriptΠ𝜃\Pi\_{\theta}roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT, we optimize over Πθ∩ΠCsubscriptΠ𝜃subscriptΠ𝐶\Pi\_{\theta}\cap\Pi\_{C}roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT ∩ roman\_Π start\_POSTSUBSCRIPT italic\_C end\_POSTSUBSCRIPT:
| | | | | |
| --- | --- | --- | --- | --- |
| | πk+1=argmaxπ∈Πθsubscript𝜋𝑘1subscript𝜋subscriptΠ𝜃\displaystyle\pi\_{k+1}=\arg\max\_{\pi\in\Pi\_{\theta}}\;italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT = roman\_arg roman\_max start\_POSTSUBSCRIPT italic\_π ∈ roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT | J(π)𝐽𝜋\displaystyle J(\pi)italic\_J ( italic\_π ) | | (3) |
| | s.t. | JCi(π)≤dii=1,…,mformulae-sequencesubscript𝐽subscript𝐶𝑖𝜋subscript𝑑𝑖𝑖1…𝑚\displaystyle J\_{C\_{i}}(\pi)\leq d\_{i}\;\;\;i=1,...,mitalic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_π ) ≤ italic\_d start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT italic\_i = 1 , … , italic\_m | |
| | | D(π,πk)≤δ.𝐷𝜋subscript𝜋𝑘𝛿\displaystyle D(\pi,\pi\_{k})\leq\delta.italic\_D ( italic\_π , italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) ≤ italic\_δ . | |
This update is difficult to implement in practice because it requires evaluation of the constraint functions to determine whether a proposed point π𝜋\piitalic\_π is feasible. When using sampling to compute policy updates, as is typically done in high-dimensional control (Duan et al., [2016](#bib.bib7)), this requires off-policy evaluation, which is known to be challenging (Jiang & Li, [2015](#bib.bib11)). In this work, we take a different approach, motivated by recent methods for trust region optimization (Schulman et al., [2015](#bib.bib22)).
We develop a principled approximation to ([3](#S5.E3 "3 ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) with a particular choice of D𝐷Ditalic\_D, where we replace the objective and constraints with surrogate functions. The surrogates we choose are easy to estimate from samples collected on πksubscript𝜋𝑘\pi\_{k}italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT, and are good local approximations for the objective and constraints. Our theoretical analysis shows that for our choices of surrogates, we can bound our update’s worst-case performance and worst-case constraint violation with values that depend on a hyperparameter of the algorithm.
To prove the performance guarantees associated with our surrogates, we first prove new bounds on the difference in returns (or constraint returns) between two arbitrary stochastic policies in terms of an average divergence between them. We then show how our bounds permit a new analysis of trust region methods in general: specifically, we prove a worst-case performance degradation at each update. We conclude by motivating, presenting, and proving gurantees on our algorithm, Constrained Policy Optimization (CPO), a trust region method for CMDPs.
###
5.1 Policy Performance Bounds
In this section, we present the theoretical foundation for our approach—a new bound on the difference in returns between two arbitrary policies. This result, which is of independent interest, extends the works of (Kakade & Langford, [2002](#bib.bib12)), (Pirotta et al., [2013](#bib.bib21)), and (Schulman et al., [2015](#bib.bib22)), providing tighter bounds. As we show later, it also relates the theoretical bounds for trust region policy improvement with the actual trust region algorithms that have been demonstrated to be successful in practice (Duan et al., [2016](#bib.bib7)). In the context of constrained policy search, we later use our results to propose policy updates that both improve the expected return and satisfy constraints.
The following theorem connects the difference in returns (or constraint returns) between two arbitrary policies to an average divergence between them.
{restatable}theoremperformancebound
For any function f:S→ℝ:𝑓→𝑆ℝf:S\to{{\mathbb{R}}}italic\_f : italic\_S → blackboard\_R and any policies π′superscript𝜋′\pi^{\prime}italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT and π𝜋\piitalic\_π, define δf(s,a,s′)≐R(s,a,s′)+γf(s′)−f(s)approaches-limitsubscript𝛿𝑓𝑠𝑎superscript𝑠′𝑅𝑠𝑎superscript𝑠′𝛾𝑓superscript𝑠′𝑓𝑠\delta\_{f}(s,a,s^{\prime})\doteq R(s,a,s^{\prime})+\gamma f(s^{\prime})-f(s)italic\_δ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ≐ italic\_R ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) + italic\_γ italic\_f ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_f ( italic\_s ),
| | | |
| --- | --- | --- |
| | ϵfπ′≐maxs|Ea∼π′,s′∼P[δf(s,a,s′)]|,approaches-limitsuperscriptsubscriptitalic-ϵ𝑓superscript𝜋′subscript𝑠subscriptEformulae-sequencesimilar-to𝑎superscript𝜋′similar-tosuperscript𝑠′𝑃delimited-[]subscript𝛿𝑓𝑠𝑎superscript𝑠′\epsilon\_{f}^{\pi^{\prime}}\doteq\max\_{s}\left|{\mathrm{E}}\_{a\sim\pi^{\prime},s^{\prime}\sim P}[\delta\_{f}(s,a,s^{\prime})]\right|,italic\_ϵ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ≐ roman\_max start\_POSTSUBSCRIPT italic\_s end\_POSTSUBSCRIPT | roman\_E start\_POSTSUBSCRIPT italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_POSTSUBSCRIPT [ italic\_δ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ] | , | |
| | | |
| --- | --- | --- |
| | Lπ,f(π′)≐Es∼dπa∼πs′∼P[(π′(a|s)π(a|s)−1)δf(s,a,s′)], andapproaches-limitsubscript𝐿𝜋𝑓superscript𝜋′similar-to𝑠superscript𝑑𝜋similar-to𝑎𝜋similar-tosuperscript𝑠′𝑃
Edelimited-[]superscript𝜋′conditional𝑎𝑠𝜋conditional𝑎𝑠1subscript𝛿𝑓𝑠𝑎superscript𝑠′ and\displaystyle L\_{\pi,f}(\pi^{\prime})\doteq\underset{\begin{subarray}{c}s\sim d^{\pi}\\
a\sim\pi\\
s^{\prime}\sim P\end{subarray}}{{\mathrm{E}}}\left[\left(\frac{\pi^{\prime}(a|s)}{\pi(a|s)}-1\right)\delta\_{f}(s,a,s^{\prime})\right],\text{ and }italic\_L start\_POSTSUBSCRIPT italic\_π , italic\_f end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ≐ start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW start\_ROW start\_CELL italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ ( divide start\_ARG italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ( italic\_a | italic\_s ) end\_ARG start\_ARG italic\_π ( italic\_a | italic\_s ) end\_ARG - 1 ) italic\_δ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ] , and | |
| | | | |
| --- | --- | --- | --- |
| | | Dπ,f±(π′)≐Lπ,f(π′)1−γ±2γϵfπ′(1−γ)2Es∼dπ[DTV(π′||π)[s]],\displaystyle D\_{\pi,f}^{\pm}(\pi^{\prime})\doteq\frac{L\_{\pi,f}(\pi^{\prime})}{1-\gamma}\pm\frac{2\gamma\epsilon\_{f}^{\pi^{\prime}}}{(1-\gamma)^{2}}\underset{s\sim d^{\pi}}{{\mathrm{E}}}\left[D\_{TV}(\pi^{\prime}||\pi)[s]\right],italic\_D start\_POSTSUBSCRIPT italic\_π , italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT ± end\_POSTSUPERSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ≐ divide start\_ARG italic\_L start\_POSTSUBSCRIPT italic\_π , italic\_f end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) end\_ARG start\_ARG 1 - italic\_γ end\_ARG ± divide start\_ARG 2 italic\_γ italic\_ϵ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT end\_ARG start\_ARG ( 1 - italic\_γ ) start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT end\_ARG start\_UNDERACCENT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] ] , | |
where DTV(π′||π)[s]=(1/2)∑a|π′(a|s)−π(a|s)|D\_{TV}(\pi^{\prime}||\pi)[s]=(1/2)\sum\_{a}\left|\pi^{\prime}(a|s)-\pi(a|s)\right|italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] = ( 1 / 2 ) ∑ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT | italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ( italic\_a | italic\_s ) - italic\_π ( italic\_a | italic\_s ) | is the total variational divergence between action distributions at s𝑠sitalic\_s. The following bounds hold:
| | | | |
| --- | --- | --- | --- |
| | Dπ,f+(π′)≥J(π′)−J(π)≥Dπ,f−(π′).superscriptsubscript𝐷𝜋𝑓superscript𝜋′𝐽superscript𝜋′𝐽𝜋superscriptsubscript𝐷𝜋𝑓superscript𝜋′D\_{\pi,f}^{+}(\pi^{\prime})\geq J(\pi^{\prime})-J(\pi)\geq D\_{\pi,f}^{-}(\pi^{\prime}).italic\_D start\_POSTSUBSCRIPT italic\_π , italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT + end\_POSTSUPERSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ≥ italic\_J ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_J ( italic\_π ) ≥ italic\_D start\_POSTSUBSCRIPT italic\_π , italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT - end\_POSTSUPERSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) . | | (4) |
Furthermore, the bounds are tight (when π′=πsuperscript𝜋′𝜋\pi^{\prime}=\piitalic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT = italic\_π, all three expressions are identically zero).
Before proceeding, we connect this result to prior work. By bounding the expectation Es∼dπ[DTV(π′||π)[s]]{\mathrm{E}}\_{s\sim d^{\pi}}\left[D\_{TV}(\pi^{\prime}||\pi)[s]\right]roman\_E start\_POSTSUBSCRIPT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT [ italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] ] with maxsDTV(π′||π)[s]\max\_{s}D\_{TV}(\pi^{\prime}||\pi)[s]roman\_max start\_POSTSUBSCRIPT italic\_s end\_POSTSUBSCRIPT italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ], picking f=Vπ𝑓superscript𝑉𝜋f=V^{\pi}italic\_f = italic\_V start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT, and bounding ϵVππ′superscriptsubscriptitalic-ϵsuperscript𝑉𝜋superscript𝜋′\epsilon\_{V^{\pi}}^{\pi^{\prime}}italic\_ϵ start\_POSTSUBSCRIPT italic\_V start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT to get a second factor of maxsDTV(π′||π)[s]\max\_{s}D\_{TV}(\pi^{\prime}||\pi)[s]roman\_max start\_POSTSUBSCRIPT italic\_s end\_POSTSUBSCRIPT italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ], we recover (up to assumption-dependent factors) the bounds given by Pirotta et al. ([2013](#bib.bib21)) as Corollary 3.6, and by Schulman et al. ([2015](#bib.bib22)) as Theorem 1a.
The choice of f=Vπ𝑓superscript𝑉𝜋f=V^{\pi}italic\_f = italic\_V start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT allows a useful form of the lower bound, so we give it as a corollary.
######
Corollary 1.
For any policies π′,πsuperscript𝜋normal-′𝜋\pi^{\prime},\piitalic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT , italic\_π, with ϵπ′≐maxs|Ea∼π′[Aπ(s,a)]|approaches-limitsuperscriptitalic-ϵsuperscript𝜋normal-′subscript𝑠subscriptnormal-Esimilar-to𝑎superscript𝜋normal-′delimited-[]superscript𝐴𝜋𝑠𝑎\epsilon^{\pi^{\prime}}\doteq\max\_{s}|{\mathrm{E}}\_{a\sim\pi^{\prime}}[A^{\pi}(s,a)]|italic\_ϵ start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ≐ roman\_max start\_POSTSUBSCRIPT italic\_s end\_POSTSUBSCRIPT | roman\_E start\_POSTSUBSCRIPT italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT [ italic\_A start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ] |, the following bound holds:
| | | | | |
| --- | --- | --- | --- | --- |
| | | J(π′)−J(π)𝐽superscript𝜋′𝐽𝜋\displaystyle J(\pi^{\prime})-J(\pi)italic\_J ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_J ( italic\_π ) | | (5) |
| | | ≥11−γEs∼dπa∼π′[Aπ(s,a)−2γϵπ′1−γDTV(π′||π)[s]].\displaystyle\geq\frac{1}{1-\gamma}\underset{\begin{subarray}{c}s\sim d^{\pi}\\
a\sim\pi^{\prime}\end{subarray}}{{\mathrm{E}}}\left[A^{\pi}(s,a)-\frac{2\gamma\epsilon^{\pi^{\prime}}}{1-\gamma}D\_{TV}(\pi^{\prime}||\pi)[s]\right].≥ divide start\_ARG 1 end\_ARG start\_ARG 1 - italic\_γ end\_ARG start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_A start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) - divide start\_ARG 2 italic\_γ italic\_ϵ start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT end\_ARG start\_ARG 1 - italic\_γ end\_ARG italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] ] . | |
The bound ([5](#S5.E5 "5 ‣ Corollary 1. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) should be compared with equation ([1](#S3.E1 "1 ‣ 3 Preliminaries ‣ Constrained Policy Optimization")). The term (1−γ)−1Es∼dπ,a∼π′[Aπ(s,a)]superscript1𝛾1subscriptEformulae-sequencesimilar-to𝑠superscript𝑑𝜋similar-to𝑎superscript𝜋′delimited-[]superscript𝐴𝜋𝑠𝑎(1-\gamma)^{-1}{\mathrm{E}}\_{s\sim d^{\pi},a\sim\pi^{\prime}}[A^{\pi}(s,a)]( 1 - italic\_γ ) start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT roman\_E start\_POSTSUBSCRIPT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT , italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT [ italic\_A start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ] in ([5](#S5.E5 "5 ‣ Corollary 1. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) is an approximation to J(π′)−J(π)𝐽superscript𝜋′𝐽𝜋J(\pi^{\prime})-J(\pi)italic\_J ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_J ( italic\_π ), using the state distribution dπsuperscript𝑑𝜋d^{\pi}italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT instead of dπ′superscript𝑑superscript𝜋′d^{\pi^{\prime}}italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT, which is known to equal J(π′)−J(π)𝐽superscript𝜋′𝐽𝜋J(\pi^{\prime})-J(\pi)italic\_J ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_J ( italic\_π ) to first order in the parameters of π′superscript𝜋′\pi^{\prime}italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT on a neighborhood around π𝜋\piitalic\_π (Kakade & Langford, [2002](#bib.bib12)). The bound can therefore be viewed as describing the worst-case approximation error, and it justifies using the approximation as a surrogate for J(π′)−J(π)𝐽superscript𝜋′𝐽𝜋J(\pi^{\prime})-J(\pi)italic\_J ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_J ( italic\_π ).
Equivalent expressions for the auxiliary costs, based on the upper bound, also follow immediately; we will later use them to make guarantees for the safety of CPO.
######
Corollary 2.
For any policies π′,πsuperscript𝜋normal-′𝜋\pi^{\prime},\piitalic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT , italic\_π, and any cost function Cisubscript𝐶𝑖C\_{i}italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT, with ϵCiπ′≐maxs|Ea∼π′[ACiπ(s,a)]|approaches-limitsubscriptsuperscriptitalic-ϵsuperscript𝜋normal-′subscript𝐶𝑖subscript𝑠subscriptnormal-Esimilar-to𝑎superscript𝜋normal-′delimited-[]subscriptsuperscript𝐴𝜋subscript𝐶𝑖𝑠𝑎\epsilon^{\pi^{\prime}}\_{C\_{i}}\doteq\max\_{s}|{\mathrm{E}}\_{a\sim\pi^{\prime}}[A^{\pi}\_{C\_{i}}(s,a)]|italic\_ϵ start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ≐ roman\_max start\_POSTSUBSCRIPT italic\_s end\_POSTSUBSCRIPT | roman\_E start\_POSTSUBSCRIPT italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT [ italic\_A start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_s , italic\_a ) ] |, the following bound holds:
| | | | | |
| --- | --- | --- | --- | --- |
| | | JCi(π′)−JCi(π)subscript𝐽subscript𝐶𝑖superscript𝜋′subscript𝐽subscript𝐶𝑖𝜋\displaystyle J\_{C\_{i}}(\pi^{\prime})-J\_{C\_{i}}(\pi)italic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_π ) | | (6) |
| | | ≤11−γEs∼dπa∼π′[ACiπ(s,a)+2γϵCiπ′1−γDTV(π′||π)[s]].\displaystyle\leq\frac{1}{1-\gamma}\underset{\begin{subarray}{c}s\sim d^{\pi}\\
a\sim\pi^{\prime}\end{subarray}}{{\mathrm{E}}}\left[A^{\pi}\_{C\_{i}}(s,a)+\frac{2\gamma\epsilon^{\pi^{\prime}}\_{C\_{i}}}{1-\gamma}D\_{TV}(\pi^{\prime}||\pi)[s]\right].≤ divide start\_ARG 1 end\_ARG start\_ARG 1 - italic\_γ end\_ARG start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_A start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_s , italic\_a ) + divide start\_ARG 2 italic\_γ italic\_ϵ start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT end\_ARG start\_ARG 1 - italic\_γ end\_ARG italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] ] . | |
The bounds we have given so far are in terms of the TV-divergence between policies, but trust region methods constrain the KL-divergence between policies, so bounds that connect performance to the KL-divergence are desirable. We make the connection through Pinsker’s inequality (Csiszar & Körner, [1981](#bib.bib6)): for arbitrary distributions p,q𝑝𝑞p,qitalic\_p , italic\_q, the TV-divergence and KL-divergence are related by DTV(p||q)≤DKL(p||q)/2D\_{TV}(p||q)\leq\sqrt{D\_{KL}(p||q)/2}italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_p | | italic\_q ) ≤ square-root start\_ARG italic\_D start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT ( italic\_p | | italic\_q ) / 2 end\_ARG. Combining this with Jensen’s inequality, we obtain
| | | | |
| --- | --- | --- | --- |
| | Es∼dπ[DTV(π′||π)[s]]\displaystyle\underset{s\sim d^{\pi}}{{\mathrm{E}}}\left[D\_{TV}(\pi^{\prime}||\pi)[s]\right]start\_UNDERACCENT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] ] | ≤Es∼dπ[12DKL(π′||π)[s]]\displaystyle\leq\underset{s\sim d^{\pi}}{{\mathrm{E}}}\left[\sqrt{\frac{1}{2}D\_{KL}(\pi^{\prime}||\pi)[s]}\right]≤ start\_UNDERACCENT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ square-root start\_ARG divide start\_ARG 1 end\_ARG start\_ARG 2 end\_ARG italic\_D start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] end\_ARG ] | |
| | | ≤12Es∼dπ[DKL(π′||π)[s]]\displaystyle\leq\sqrt{\frac{1}{2}\underset{s\sim d^{\pi}}{{\mathrm{E}}}\left[D\_{KL}(\pi^{\prime}||\pi)[s]\right]}≤ square-root start\_ARG divide start\_ARG 1 end\_ARG start\_ARG 2 end\_ARG start\_UNDERACCENT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_D start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] ] end\_ARG | | (7) |
From ([7](#S5.E7 "7 ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) we immediately obtain the following.
######
Corollary 3.
In bounds ([4](#S5.E4 "4 ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")), ([5](#S5.E5 "5 ‣ Corollary 1. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")), and ([6](#S5.E6 "6 ‣ Corollary 2. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")), make the substitution
| | | |
| --- | --- | --- |
| | Es∼dπ[DTV(π′||π)[s]]→12Es∼dπ[DKL(π′||π)[s]].\underset{s\sim d^{\pi}}{{\mathrm{E}}}\left[D\_{TV}(\pi^{\prime}||\pi)[s]\right]\to\sqrt{\frac{1}{2}\underset{s\sim d^{\pi}}{{\mathrm{E}}}\left[D\_{KL}(\pi^{\prime}||\pi)[s]\right]}.start\_UNDERACCENT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] ] → square-root start\_ARG divide start\_ARG 1 end\_ARG start\_ARG 2 end\_ARG start\_UNDERACCENT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_D start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] ] end\_ARG . | |
The resulting bounds hold.
###
5.2 Trust Region Methods
Trust region algorithms for reinforcement learning (Schulman et al., [2015](#bib.bib22), [2016](#bib.bib23)) have policy updates of the form
| | | | | |
| --- | --- | --- | --- | --- |
| | πk+1=argmaxπ∈Πθsubscript𝜋𝑘1subscript𝜋subscriptΠ𝜃\displaystyle\pi\_{k+1}=\arg\max\_{\pi\in\Pi\_{\theta}}\;italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT = roman\_arg roman\_max start\_POSTSUBSCRIPT italic\_π ∈ roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT | Es∼dπka∼π[Aπk(s,a)]similar-to𝑠superscript𝑑subscript𝜋𝑘similar-to𝑎𝜋
Edelimited-[]superscript𝐴subscript𝜋𝑘𝑠𝑎\displaystyle\underset{\begin{subarray}{c}s\sim d^{\pi\_{k}}\\
a\sim\pi\end{subarray}}{{\mathrm{E}}}\left[A^{\pi\_{k}}(s,a)\right]start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ] | | (8) |
| | s.t. | D¯KL(π||πk)≤δ,\displaystyle\bar{D}\_{KL}(\pi||\pi\_{k})\leq\delta,over¯ start\_ARG italic\_D end\_ARG start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT ( italic\_π | | italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) ≤ italic\_δ , | |
where D¯KL(π||πk)=Es∼πk[DKL(π||πk)[s]]\bar{D}\_{KL}(\pi||\pi\_{k})={\mathrm{E}}\_{s\sim\pi\_{k}}\left[D\_{KL}(\pi||\pi\_{k})[s]\right]over¯ start\_ARG italic\_D end\_ARG start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT ( italic\_π | | italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) = roman\_E start\_POSTSUBSCRIPT italic\_s ∼ italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT [ italic\_D start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT ( italic\_π | | italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) [ italic\_s ] ], and δ>0𝛿0\delta>0italic\_δ > 0 is the step size. The set {πθ∈Πθ:D¯KL(π||πk)≤δ}\{\pi\_{\theta}\in\Pi\_{\theta}:\bar{D}\_{KL}(\pi||\pi\_{k})\leq\delta\}{ italic\_π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT ∈ roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT : over¯ start\_ARG italic\_D end\_ARG start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT ( italic\_π | | italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) ≤ italic\_δ } is called the trust region.
The primary motivation for this update is that it is an approximation to optimizing the lower bound on policy performance given in ([5](#S5.E5 "5 ‣ Corollary 1. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")), which would guarantee monotonic performance improvements. This is important for optimizing neural network policies, which are known to suffer from performance collapse after bad updates (Duan et al., [2016](#bib.bib7)). Despite the approximation, trust region steps usually give monotonic improvements (Schulman et al., [2015](#bib.bib22); Duan et al., [2016](#bib.bib7)) and have shown state-of-the-art performance in the deep RL setting (Duan et al., [2016](#bib.bib7); Gu et al., [2017](#bib.bib9)), making the approach appealing for developing policy search methods for CMDPs.
Until now, the particular choice of trust region for ([8](#S5.E8 "8 ‣ 5.2 Trust Region Methods ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) was heuristically motivated; with ([5](#S5.E5 "5 ‣ Corollary 1. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) and Corollary [3](#Thmcorollary3 "Corollary 3. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization"), we are able to show that it is principled and comes with a worst-case performance degradation guarantee that depends on δ𝛿\deltaitalic\_δ.
######
Proposition 1 (Trust Region Update Performance).
Suppose πksubscript𝜋𝑘\pi\_{k}italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT, πk+1subscript𝜋𝑘1\pi\_{k+1}italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT are related by ([8](#S5.E8 "8 ‣ 5.2 Trust Region Methods ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")), and that πk∈Πθsubscript𝜋𝑘subscriptnormal-Π𝜃\pi\_{k}\in\Pi\_{\theta}italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ∈ roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT. A lower bound on the policy performance difference between πksubscript𝜋𝑘\pi\_{k}italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT and πk+1subscript𝜋𝑘1\pi\_{k+1}italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT is
| | | | |
| --- | --- | --- | --- |
| | J(πk+1)−J(πk)≥−2δγϵπk+1(1−γ)2,𝐽subscript𝜋𝑘1𝐽subscript𝜋𝑘2𝛿𝛾superscriptitalic-ϵsubscript𝜋𝑘1superscript1𝛾2J(\pi\_{k+1})-J(\pi\_{k})\geq\frac{-\sqrt{2\delta}\gamma\epsilon^{\pi\_{k+1}}}{(1-\gamma)^{2}},italic\_J ( italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT ) - italic\_J ( italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) ≥ divide start\_ARG - square-root start\_ARG 2 italic\_δ end\_ARG italic\_γ italic\_ϵ start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT end\_ARG start\_ARG ( 1 - italic\_γ ) start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT end\_ARG , | | (9) |
where ϵπk+1=maxs|Ea∼πk+1[Aπk(s,a)]|superscriptitalic-ϵsubscript𝜋𝑘1subscript𝑠subscriptnormal-Esimilar-to𝑎subscript𝜋𝑘1delimited-[]superscript𝐴subscript𝜋𝑘𝑠𝑎\epsilon^{\pi\_{k+1}}=\max\_{s}\left|{\mathrm{E}}\_{a\sim\pi\_{k+1}}\left[A^{\pi\_{k}}(s,a)\right]\right|italic\_ϵ start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT = roman\_max start\_POSTSUBSCRIPT italic\_s end\_POSTSUBSCRIPT | roman\_E start\_POSTSUBSCRIPT italic\_a ∼ italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT [ italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ] |.
###### Proof.
πksubscript𝜋𝑘\pi\_{k}italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT is a feasible point of ([8](#S5.E8 "8 ‣ 5.2 Trust Region Methods ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) with objective value 0, so Es∼dπk,a∼πk+1[Aπk(s,a)]≥0subscriptEformulae-sequencesimilar-to𝑠superscript𝑑subscript𝜋𝑘similar-to𝑎subscript𝜋𝑘1delimited-[]superscript𝐴subscript𝜋𝑘𝑠𝑎0{\mathrm{E}}\_{s\sim d^{\pi\_{k}},a\sim\pi\_{k+1}}\left[A^{\pi\_{k}}(s,a)\right]\geq 0roman\_E start\_POSTSUBSCRIPT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT , italic\_a ∼ italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT [ italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ] ≥ 0. The rest follows by ([5](#S5.E5 "5 ‣ Corollary 1. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) and Corollary [3](#Thmcorollary3 "Corollary 3. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization"), noting that ([8](#S5.E8 "8 ‣ 5.2 Trust Region Methods ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) bounds the average KL-divergence by δ𝛿\deltaitalic\_δ.
∎
This result is useful for two reasons: 1) it is of independent interest, as it helps tighten the connection between theory and practice for deep RL, and 2) the choice to develop CPO as a trust region method means that CPO inherits this performance guarantee.
###
5.3 Trust Region Optimization for Constrained MDPs
Constrained policy optimization (CPO), which we present and justify in this section, is a policy search algorithm for CMDPs with updates that approximately solve ([3](#S5.E3 "3 ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) with a particular choice of D𝐷Ditalic\_D. First, we describe a policy search update for CMDPs that alleviates the issue of off-policy evaluation, and comes with guarantees of monotonic performance improvement and constraint satisfaction. Then, because the theoretically guaranteed update will take too-small steps in practice, we propose CPO as a practical approximation based on trust region methods.
By corollaries [5](#S5.E5 "5 ‣ Corollary 1. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization"), [6](#S5.E6 "6 ‣ Corollary 2. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization"), and [3](#Thmcorollary3 "Corollary 3. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization"), for appropriate coefficients αksubscript𝛼𝑘\alpha\_{k}italic\_α start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT, βkisuperscriptsubscript𝛽𝑘𝑖\beta\_{k}^{i}italic\_β start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_i end\_POSTSUPERSCRIPT the update
| | | | |
| --- | --- | --- | --- |
| | | πk+1=argmaxπ∈ΠθEs∼dπka∼π[Aπk(s,a)]−αkD¯KL(π||πk)\displaystyle\pi\_{k+1}=\arg\max\_{\pi\in\Pi\_{\theta}}\;\underset{\begin{subarray}{c}s\sim d^{\pi\_{k}}\\
a\sim\pi\end{subarray}}{{\mathrm{E}}}\left[A^{\pi\_{k}}(s,a)\right]-\alpha\_{k}\sqrt{\bar{D}\_{KL}(\pi||\pi\_{k})}italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT = roman\_arg roman\_max start\_POSTSUBSCRIPT italic\_π ∈ roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ] - italic\_α start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT square-root start\_ARG over¯ start\_ARG italic\_D end\_ARG start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT ( italic\_π | | italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) end\_ARG | |
| | | s.t. JCi(πk)+Es∼dπka∼π[ACiπk(s,a)1−γ]+βkiD¯KL(π||πk)≤di\displaystyle\text{s.t. }J\_{C\_{i}}(\pi\_{k})+\underset{\begin{subarray}{c}s\sim d^{\pi\_{k}}\\
a\sim\pi\end{subarray}}{{\mathrm{E}}}\left[\frac{A^{\pi\_{k}}\_{C\_{i}}(s,a)}{1-\gamma}\right]+\beta\_{k}^{i}\sqrt{\bar{D}\_{KL}(\pi||\pi\_{k})}\leq d\_{i}s.t. italic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) + start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ divide start\_ARG italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_s , italic\_a ) end\_ARG start\_ARG 1 - italic\_γ end\_ARG ] + italic\_β start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_i end\_POSTSUPERSCRIPT square-root start\_ARG over¯ start\_ARG italic\_D end\_ARG start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT ( italic\_π | | italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) end\_ARG ≤ italic\_d start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT | |
is guaranteed to produce policies with monotonically nondecreasing returns that satisfy the original constraints. (Observe that the constraint here is on an upper bound for JCi(π)subscript𝐽subscript𝐶𝑖𝜋J\_{C\_{i}}(\pi)italic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_π ) by ([6](#S5.E6 "6 ‣ Corollary 2. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")).) The off-policy evaluation issue is alleviated, because both the objective and constraints involve expectations over state distributions dπksuperscript𝑑subscript𝜋𝑘d^{\pi\_{k}}italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT, which we presume to have samples from. Because the bounds are tight, the problem is always feasible (as long as π0subscript𝜋0\pi\_{0}italic\_π start\_POSTSUBSCRIPT 0 end\_POSTSUBSCRIPT is feasible). However, the penalties on policy divergence are quite steep for discount factors close to 1, so steps taken with this update might be small.
Inspired by trust region methods, we propose CPO, which uses a trust region instead of penalties on policy divergence to enable larger step sizes:
| | | | | |
| --- | --- | --- | --- | --- |
| | πk+1=subscript𝜋𝑘1absent\displaystyle\pi\_{k+1}=italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT = | argmaxπ∈ΠθEs∼dπka∼π[Aπk(s,a)]subscript𝜋subscriptΠ𝜃similar-to𝑠superscript𝑑subscript𝜋𝑘similar-to𝑎𝜋
Edelimited-[]superscript𝐴subscript𝜋𝑘𝑠𝑎\displaystyle\arg\max\_{\pi\in\Pi\_{\theta}}\;\underset{\begin{subarray}{c}s\sim d^{\pi\_{k}}\\
a\sim\pi\end{subarray}}{{\mathrm{E}}}\left[A^{\pi\_{k}}(s,a)\right]roman\_arg roman\_max start\_POSTSUBSCRIPT italic\_π ∈ roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ] | | (10) |
| | s.t. | JCi(πk)+11−γEs∼dπka∼π[ACiπk(s,a)]≤di∀isubscript𝐽subscript𝐶𝑖subscript𝜋𝑘11𝛾similar-to𝑠superscript𝑑subscript𝜋𝑘similar-to𝑎𝜋
Edelimited-[]subscriptsuperscript𝐴subscript𝜋𝑘subscript𝐶𝑖𝑠𝑎subscript𝑑𝑖for-all𝑖\displaystyle J\_{C\_{i}}(\pi\_{k})+\frac{1}{1-\gamma}\underset{\begin{subarray}{c}s\sim d^{\pi\_{k}}\\
a\sim\pi\end{subarray}}{{\mathrm{E}}}\left[A^{\pi\_{k}}\_{C\_{i}}(s,a)\right]\leq d\_{i}\;\;\;\forall iitalic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) + divide start\_ARG 1 end\_ARG start\_ARG 1 - italic\_γ end\_ARG start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_s , italic\_a ) ] ≤ italic\_d start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT ∀ italic\_i | |
| | | D¯KL(π||πk)≤δ.\displaystyle\bar{D}\_{KL}(\pi||\pi\_{k})\leq\delta.over¯ start\_ARG italic\_D end\_ARG start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT ( italic\_π | | italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) ≤ italic\_δ . | |
Because this is a trust region method, it inherits the performance guarantee of Proposition [1](#Thmproposition1 "Proposition 1 (Trust Region Update Performance). ‣ 5.2 Trust Region Methods ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization"). Furthermore, by corollaries [6](#S5.E6 "6 ‣ Corollary 2. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization") and [3](#Thmcorollary3 "Corollary 3. ‣ 5.1 Policy Performance Bounds ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization"), we have a performance guarantee for approximate satisfaction of constraints:
######
Proposition 2 (CPO Update Worst-Case Constraint Violation).
Suppose πk,πk+1subscript𝜋𝑘subscript𝜋𝑘1\pi\_{k},\pi\_{k+1}italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT , italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT are related by ([10](#S5.E10 "10 ‣ 5.3 Trust Region Optimization for Constrained MDPs ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")), and that Πθsubscriptnormal-Π𝜃\Pi\_{\theta}roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT in ([10](#S5.E10 "10 ‣ 5.3 Trust Region Optimization for Constrained MDPs ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) is any set of policies with πk∈Πθsubscript𝜋𝑘subscriptnormal-Π𝜃\pi\_{k}\in\Pi\_{\theta}italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ∈ roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT. An upper bound on the Cisubscript𝐶𝑖C\_{i}italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT-return of πk+1subscript𝜋𝑘1\pi\_{k+1}italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT is
| | | |
| --- | --- | --- |
| | JCi(πk+1)≤di+2δγϵCiπk+1(1−γ)2,subscript𝐽subscript𝐶𝑖subscript𝜋𝑘1subscript𝑑𝑖2𝛿𝛾subscriptsuperscriptitalic-ϵsubscript𝜋𝑘1subscript𝐶𝑖superscript1𝛾2J\_{C\_{i}}(\pi\_{k+1})\leq d\_{i}+\frac{\sqrt{2\delta}\gamma\epsilon^{\pi\_{k+1}}\_{C\_{i}}}{(1-\gamma)^{2}},italic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT ) ≤ italic\_d start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT + divide start\_ARG square-root start\_ARG 2 italic\_δ end\_ARG italic\_γ italic\_ϵ start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT end\_ARG start\_ARG ( 1 - italic\_γ ) start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT end\_ARG , | |
where ϵCiπk+1=maxs|Ea∼πk+1[ACiπk(s,a)]|subscriptsuperscriptitalic-ϵsubscript𝜋𝑘1subscript𝐶𝑖subscript𝑠subscriptnormal-Esimilar-to𝑎subscript𝜋𝑘1delimited-[]subscriptsuperscript𝐴subscript𝜋𝑘subscript𝐶𝑖𝑠𝑎\epsilon^{\pi\_{k+1}}\_{C\_{i}}=\max\_{s}\left|{\mathrm{E}}\_{a\sim\pi\_{k+1}}\left[A^{\pi\_{k}}\_{C\_{i}}(s,a)\right]\right|italic\_ϵ start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT = roman\_max start\_POSTSUBSCRIPT italic\_s end\_POSTSUBSCRIPT | roman\_E start\_POSTSUBSCRIPT italic\_a ∼ italic\_π start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT [ italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_s , italic\_a ) ] |.
6 Practical Implementation
---------------------------
In this section, we show how to implement an approximation to the update ([10](#S5.E10 "10 ‣ 5.3 Trust Region Optimization for Constrained MDPs ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) that can be efficiently computed, even when optimizing policies with thousands of parameters. To address the issue of approximation and sampling errors that arise in practice, as well as the potential violations described by Proposition [2](#Thmproposition2 "Proposition 2 (CPO Update Worst-Case Constraint Violation). ‣ 5.3 Trust Region Optimization for Constrained MDPs ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization"), we also propose to tighten the constraints by constraining upper bounds of the auxilliary costs, instead of the auxilliary costs themselves.
###
6.1 Approximately Solving the CPO Update
For policies with high-dimensional parameter spaces like neural networks, ([10](#S5.E10 "10 ‣ 5.3 Trust Region Optimization for Constrained MDPs ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) can be impractical to solve directly because of the computational cost. However, for small step sizes δ𝛿\deltaitalic\_δ, the objective and cost constraints are well-approximated by linearizing around πksubscript𝜋𝑘\pi\_{k}italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT, and the KL-divergence constraint is well-approximated by second order expansion (at πk=πsubscript𝜋𝑘𝜋\pi\_{k}=\piitalic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT = italic\_π, the KL-divergence and its gradient are both zero). Denoting the gradient of the objective as g𝑔gitalic\_g, the gradient of constraint i𝑖iitalic\_i as bisubscript𝑏𝑖b\_{i}italic\_b start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT, the Hessian of the KL-divergence as H𝐻Hitalic\_H, and defining ci≐JCi(πk)−diapproaches-limitsubscript𝑐𝑖subscript𝐽subscript𝐶𝑖subscript𝜋𝑘subscript𝑑𝑖c\_{i}\doteq J\_{C\_{i}}(\pi\_{k})-d\_{i}italic\_c start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT ≐ italic\_J start\_POSTSUBSCRIPT italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) - italic\_d start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT, the approximation to ([10](#S5.E10 "10 ‣ 5.3 Trust Region Optimization for Constrained MDPs ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) is:
| | | | | |
| --- | --- | --- | --- | --- |
| | θk+1=argmaxθsubscript𝜃𝑘1subscript𝜃\displaystyle\theta\_{k+1}=\arg\max\_{\theta}\;\;\;italic\_θ start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT = roman\_arg roman\_max start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT | gT(θ−θk)superscript𝑔𝑇𝜃subscript𝜃𝑘\displaystyle g^{T}(\theta-\theta\_{k})italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT ( italic\_θ - italic\_θ start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) | | (11) |
| | s.t. | ci+biT(θ−θk)≤0i=1,…,mformulae-sequencesubscript𝑐𝑖superscriptsubscript𝑏𝑖𝑇𝜃subscript𝜃𝑘0𝑖1…𝑚\displaystyle c\_{i}+b\_{i}^{T}(\theta-\theta\_{k})\leq 0\;\;\;i=1,...,mitalic\_c start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT + italic\_b start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT ( italic\_θ - italic\_θ start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) ≤ 0 italic\_i = 1 , … , italic\_m | |
| | | 12(θ−θk)TH(θ−θk)≤δ.12superscript𝜃subscript𝜃𝑘𝑇𝐻𝜃subscript𝜃𝑘𝛿\displaystyle\frac{1}{2}(\theta-\theta\_{k})^{T}H(\theta-\theta\_{k})\leq\delta.divide start\_ARG 1 end\_ARG start\_ARG 2 end\_ARG ( italic\_θ - italic\_θ start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H ( italic\_θ - italic\_θ start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) ≤ italic\_δ . | |
Because the Fisher information matrix (FIM) H𝐻Hitalic\_H is always positive semi-definite (and we will assume it to be positive-definite in what follows), this optimization problem is convex and, when feasible, can be solved efficiently using duality. (We reserve the case where it is not feasible for the next subsection.) With B≐[b1,…,bm]approaches-limit𝐵subscript𝑏1…subscript𝑏𝑚B\doteq[b\_{1},...,b\_{m}]italic\_B ≐ [ italic\_b start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT , … , italic\_b start\_POSTSUBSCRIPT italic\_m end\_POSTSUBSCRIPT ] and c≐[c1,…,cm]Tapproaches-limit𝑐superscriptsubscript𝑐1…subscript𝑐𝑚
𝑇c\doteq[c\_{1},...,c\_{m}]^{T}italic\_c ≐ [ italic\_c start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT , … , italic\_c start\_POSTSUBSCRIPT italic\_m end\_POSTSUBSCRIPT ] start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT, a dual to ([11](#S6.E11 "11 ‣ 6.1 Approximately Solving the CPO Update ‣ 6 Practical Implementation ‣ Constrained Policy Optimization")) can be expressed as
| | | | |
| --- | --- | --- | --- |
| | maxλ≥0ν⪰0−12λ(gTH−1g−2rTν+νTSν)+νTc−λδ2,subscript𝜆0succeeds-or-equals𝜈012𝜆superscript𝑔𝑇superscript𝐻1𝑔2superscript𝑟𝑇𝜈superscript𝜈𝑇𝑆𝜈superscript𝜈𝑇𝑐𝜆𝛿2\displaystyle\max\_{\begin{subarray}{c}\lambda\geq 0\\
\nu\succeq 0\end{subarray}}\frac{-1}{2\lambda}\left(g^{T}H^{-1}g-2r^{T}\nu+\nu^{T}S\nu\right)+\nu^{T}c-\frac{\lambda\delta}{2},roman\_max start\_POSTSUBSCRIPT start\_ARG start\_ROW start\_CELL italic\_λ ≥ 0 end\_CELL end\_ROW start\_ROW start\_CELL italic\_ν ⪰ 0 end\_CELL end\_ROW end\_ARG end\_POSTSUBSCRIPT divide start\_ARG - 1 end\_ARG start\_ARG 2 italic\_λ end\_ARG ( italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_g - 2 italic\_r start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_ν + italic\_ν start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_S italic\_ν ) + italic\_ν start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_c - divide start\_ARG italic\_λ italic\_δ end\_ARG start\_ARG 2 end\_ARG , | | (12) |
where r≐gTH−1Bapproaches-limit𝑟superscript𝑔𝑇superscript𝐻1𝐵r\doteq g^{T}H^{-1}Bitalic\_r ≐ italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_B, S≐BTH−1Bapproaches-limit𝑆superscript𝐵𝑇superscript𝐻1𝐵S\doteq B^{T}H^{-1}Bitalic\_S ≐ italic\_B start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_B. This is a convex program in m+1𝑚1m+1italic\_m + 1 variables; when the number of constraints is small by comparison to the dimension of θ𝜃\thetaitalic\_θ, this is much easier to solve than ([11](#S6.E11 "11 ‣ 6.1 Approximately Solving the CPO Update ‣ 6 Practical Implementation ‣ Constrained Policy Optimization")). If λ\*,ν\*superscript𝜆superscript𝜈\lambda^{\*},\nu^{\*}italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT , italic\_ν start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT are a solution to the dual, the solution to the primal is
| | | | |
| --- | --- | --- | --- |
| | θ\*=θk+1λ\*H−1(g−Bν\*).superscript𝜃subscript𝜃𝑘1superscript𝜆superscript𝐻1𝑔𝐵superscript𝜈\theta^{\*}=\theta\_{k}+\frac{1}{\lambda^{\*}}H^{-1}\left(g-B\nu^{\*}\right).italic\_θ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = italic\_θ start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT + divide start\_ARG 1 end\_ARG start\_ARG italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT end\_ARG italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT ( italic\_g - italic\_B italic\_ν start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ) . | | (13) |
Our algorithm solves the dual for λ\*,ν\*superscript𝜆superscript𝜈\lambda^{\*},\nu^{\*}italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT , italic\_ν start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT and uses it to propose the policy update ([13](#S6.E13 "13 ‣ 6.1 Approximately Solving the CPO Update ‣ 6 Practical Implementation ‣ Constrained Policy Optimization")). For the special case where there is only one constraint, we give an analytical solution in the supplementary material (Theorem [1](#Thmtheorem1 "Theorem 1 (Optimizing Linear Objective with Linear and Quadratic Constraints). ‣ 10.2 Proof of Analytical Solution to LQCLP ‣ 10 Appendix ‣ Constrained Policy Optimization")) which removes the need for an inner-loop optimization. Our experiments have only a single constraint, and make use of the analytical solution.
Because of approximation error, the proposed update may not satisfy the constraints in ([10](#S5.E10 "10 ‣ 5.3 Trust Region Optimization for Constrained MDPs ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")); a backtracking line search is used to ensure surrogate constraint satisfaction. Also, for high-dimensional policies, it is impractically expensive to invert the FIM. This poses a challenge for computing H−1gsuperscript𝐻1𝑔H^{-1}gitalic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_g and H−1bisuperscript𝐻1subscript𝑏𝑖H^{-1}b\_{i}italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_b start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT, which appear in the dual. Like (Schulman et al., [2015](#bib.bib22)), we approximately compute them using the conjugate gradient method.
###
6.2 Feasibility
Due to approximation errors, CPO may take a bad step and produce an infeasible iterate πksubscript𝜋𝑘\pi\_{k}italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT. Sometimes ([11](#S6.E11 "11 ‣ 6.1 Approximately Solving the CPO Update ‣ 6 Practical Implementation ‣ Constrained Policy Optimization")) will still be feasible and CPO can automatically recover from its bad step, but for the infeasible case, a recovery method is necessary. In our experiments, where we only have one constraint, we recover by proposing an update to purely decrease the constraint value:
| | | | |
| --- | --- | --- | --- |
| | θ\*=θk−2δbTH−1bH−1b.superscript𝜃subscript𝜃𝑘2𝛿superscript𝑏𝑇superscript𝐻1𝑏superscript𝐻1𝑏\theta^{\*}=\theta\_{k}-\sqrt{\frac{2\delta}{b^{T}H^{-1}b}}H^{-1}b.italic\_θ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = italic\_θ start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT - square-root start\_ARG divide start\_ARG 2 italic\_δ end\_ARG start\_ARG italic\_b start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_b end\_ARG end\_ARG italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_b . | | (14) |
As before, this is followed by a line search. This approach is principled in that it uses the limiting search direction as the intersection of the trust region and the constraint region shrinks to zero. We give the pseudocode for our algorithm (for the single-constraint case) as Algorithm [1](#alg1 "Algorithm 1 ‣ 6.2 Feasibility ‣ 6 Practical Implementation ‣ Constrained Policy Optimization").
Algorithm 1 Constrained Policy Optimization
Input: Initial policy π0∈Πθsubscript𝜋0subscriptΠ𝜃\pi\_{0}\in\Pi\_{\theta}italic\_π start\_POSTSUBSCRIPT 0 end\_POSTSUBSCRIPT ∈ roman\_Π start\_POSTSUBSCRIPT italic\_θ end\_POSTSUBSCRIPT tolerance α𝛼\alphaitalic\_α
for k=0,1,2,…𝑘012…k=0,1,2,...italic\_k = 0 , 1 , 2 , … do
Sample a set of trajectories 𝒟={τ}∼πk=π(θk)𝒟𝜏similar-tosubscript𝜋𝑘𝜋subscript𝜃𝑘{\mathcal{D}}=\{\tau\}\sim\pi\_{k}=\pi(\theta\_{k})caligraphic\_D = { italic\_τ } ∼ italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT = italic\_π ( italic\_θ start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT )
Form sample estimates g^,b^,H^,c^^𝑔^𝑏^𝐻^𝑐\hat{g},\hat{b},\hat{H},\hat{c}over^ start\_ARG italic\_g end\_ARG , over^ start\_ARG italic\_b end\_ARG , over^ start\_ARG italic\_H end\_ARG , over^ start\_ARG italic\_c end\_ARG with 𝒟𝒟{\mathcal{D}}caligraphic\_D
if approximate CPO is feasible then
Solve dual problem ([12](#S6.E12 "12 ‣ 6.1 Approximately Solving the CPO Update ‣ 6 Practical Implementation ‣ Constrained Policy Optimization")) for λk\*,νk\*subscriptsuperscript𝜆𝑘subscriptsuperscript𝜈𝑘\lambda^{\*}\_{k},\nu^{\*}\_{k}italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT , italic\_ν start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT
Compute policy proposal θ\*superscript𝜃\theta^{\*}italic\_θ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT with ([13](#S6.E13 "13 ‣ 6.1 Approximately Solving the CPO Update ‣ 6 Practical Implementation ‣ Constrained Policy Optimization"))
else
Compute recovery policy proposal θ\*superscript𝜃\theta^{\*}italic\_θ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT with ([14](#S6.E14 "14 ‣ 6.2 Feasibility ‣ 6 Practical Implementation ‣ Constrained Policy Optimization"))
end if
Obtain θk+1subscript𝜃𝑘1\theta\_{k+1}italic\_θ start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT by backtracking linesearch to enforce satisfaction of sample estimates of constraints in ([10](#S5.E10 "10 ‣ 5.3 Trust Region Optimization for Constrained MDPs ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization"))
end for
###
6.3 Tightening Constraints via Cost Shaping
Because of the various approximations between ([3](#S5.E3 "3 ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")) and our practical algorithm, it is important to build a factor of safety into the algorithm to minimize the chance of constraint violations. To this end, we choose to constrain upper bounds on the original constraints, Ci+superscriptsubscript𝐶𝑖C\_{i}^{+}italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT + end\_POSTSUPERSCRIPT, instead of the original constraints themselves. We do this by cost shaping:
| | | | |
| --- | --- | --- | --- |
| | Ci+(s,a,s′)=Ci(s,a,s′)+Δi(s,a,s′),superscriptsubscript𝐶𝑖𝑠𝑎superscript𝑠′subscript𝐶𝑖𝑠𝑎superscript𝑠′subscriptΔ𝑖𝑠𝑎superscript𝑠′C\_{i}^{+}(s,a,s^{\prime})=C\_{i}(s,a,s^{\prime})+\Delta\_{i}(s,a,s^{\prime}),italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT + end\_POSTSUPERSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) = italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) + roman\_Δ start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) , | | (15) |
where Δi:S×A×S→ℝ+:subscriptΔ𝑖→𝑆𝐴𝑆subscriptℝ\Delta\_{i}:S\times A\times S\to{{\mathbb{R}}\_{+}}roman\_Δ start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT : italic\_S × italic\_A × italic\_S → blackboard\_R start\_POSTSUBSCRIPT + end\_POSTSUBSCRIPT correlates in some useful way with Cisubscript𝐶𝑖C\_{i}italic\_C start\_POSTSUBSCRIPT italic\_i end\_POSTSUBSCRIPT.
In our experiments, where we have only one constraint, we partition states into safe states and unsafe states, and the agent suffers a safety cost of 1111 for being in an unsafe state. We choose ΔΔ\Deltaroman\_Δ to be the probability of entering an unsafe state within a fixed time horizon, according to a learned model that is updated at each iteration. This choice confers the additional benefit of smoothing out sparse constraints.
7 Connections to Prior Work
----------------------------
Our method has similar policy updates to primal-dual methods like those proposed by Chow et al. ([2015](#bib.bib5)), but crucially, we differ in computing the dual variables (the Lagrange multipliers for the constraints). In primal-dual optimization (PDO), dual variables are stateful and learned concurrently with the primal variables (Boyd et al., [2003](#bib.bib4)). In a PDO algorithm for solving ([3](#S5.E3 "3 ‣ 5 Constrained Policy Optimization ‣ Constrained Policy Optimization")), dual variables would be updated according to
| | | | |
| --- | --- | --- | --- |
| | νk+1=(νk+αk(JC(πk)−d))+,subscript𝜈𝑘1subscriptsubscript𝜈𝑘subscript𝛼𝑘subscript𝐽𝐶subscript𝜋𝑘𝑑\nu\_{k+1}=\left(\nu\_{k}+\alpha\_{k}\left(J\_{C}(\pi\_{k})-d\right)\right)\_{+},italic\_ν start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT = ( italic\_ν start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT + italic\_α start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ( italic\_J start\_POSTSUBSCRIPT italic\_C end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) - italic\_d ) ) start\_POSTSUBSCRIPT + end\_POSTSUBSCRIPT , | | (16) |
where αksubscript𝛼𝑘\alpha\_{k}italic\_α start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT is a learning rate. In this approach, intermediary policies are not guaranteed to satisfy constraints—only the policy at convergence is. By contrast, CPO computes new dual variables from scratch at each update to exactly enforce constraints.
Returns:





Constraint values: (closer to the limit is better)

(a) Point-Circle

(b) Ant-Circle

(c) Humanoid-Circle

(d) Point-Gather

(e) Ant-Gather
Figure 1: Average performance for CPO, PDO, and TRPO over several seeds (5 in the Point environments, 10 in all others); the x𝑥xitalic\_x-axis is training iteration. CPO drives the constraint function almost directly to the limit in all experiments, while PDO frequently suffers from over- or under-correction. TRPO is included to verify that optimal unconstrained behaviors are infeasible for the constrained problem.
8 Experiments
--------------
In our experiments, we aim to answer the following:
* •
Does CPO succeed at enforcing behavioral constraints when training neural network policies with thousands of parameters?
* •
How does CPO compare with a baseline that uses primal-dual optimization? Does CPO behave better with respect to constraints?
* •
How much does it help to constrain a cost upper bound ([15](#S6.E15 "15 ‣ 6.3 Tightening Constraints via Cost Shaping ‣ 6 Practical Implementation ‣ Constrained Policy Optimization")), instead of directly constraining the cost?
* •
What benefits are conferred by using constraints instead of fixed penalties?
We designed experiments that are easy to interpret and motivated by safety. We consider two tasks, and train multiple different agents (robots) for each task:
* •
Circle: The agent is rewarded for running in a wide circle, but is constrained to stay within a safe region smaller than the radius of the target circle.
* •
Gather: The agent is rewarded for collecting green apples, and constrained to avoid red bombs.
For the Circle task, the exact geometry is illustrated in Figure [5](#S10.F5 "Figure 5 ‣ 10.3.1 Environments ‣ 10.3 Experimental Parameters ‣ 10 Appendix ‣ Constrained Policy Optimization") in the supplementary material. Note that there are no physical walls: the agent only interacts with boundaries through the constraint costs. The reward and constraint cost functions are described in supplementary material (Section [10.3.1](#S10.SS3.SSS1 "10.3.1 Environments ‣ 10.3 Experimental Parameters ‣ 10 Appendix ‣ Constrained Policy Optimization")). In each of these tasks, we have only one constraint; we refer to it as C𝐶Citalic\_C and its upper bound from ([15](#S6.E15 "15 ‣ 6.3 Tightening Constraints via Cost Shaping ‣ 6 Practical Implementation ‣ Constrained Policy Optimization")) as C+superscript𝐶C^{+}italic\_C start\_POSTSUPERSCRIPT + end\_POSTSUPERSCRIPT.
We experiment with three different agents: a point-mass (S⊆ℝ9,A⊆ℝ2)formulae-sequence𝑆superscriptℝ9𝐴superscriptℝ2(S\subseteq{{\mathbb{R}}^{9}},A\subseteq{{\mathbb{R}}^{2}})( italic\_S ⊆ blackboard\_R start\_POSTSUPERSCRIPT 9 end\_POSTSUPERSCRIPT , italic\_A ⊆ blackboard\_R start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT ), a quadruped robot (called an ‘ant’) (S⊆ℝ32,A⊆ℝ8)formulae-sequence𝑆superscriptℝ32𝐴superscriptℝ8(S\subseteq{{\mathbb{R}}^{32}},A\subseteq{{\mathbb{R}}^{8}})( italic\_S ⊆ blackboard\_R start\_POSTSUPERSCRIPT 32 end\_POSTSUPERSCRIPT , italic\_A ⊆ blackboard\_R start\_POSTSUPERSCRIPT 8 end\_POSTSUPERSCRIPT ), and a simple humanoid (S⊆ℝ102,A⊆ℝ10)formulae-sequence𝑆superscriptℝ102𝐴superscriptℝ10(S\subseteq{{\mathbb{R}}^{102}},A\subseteq{{\mathbb{R}}^{10}})( italic\_S ⊆ blackboard\_R start\_POSTSUPERSCRIPT 102 end\_POSTSUPERSCRIPT , italic\_A ⊆ blackboard\_R start\_POSTSUPERSCRIPT 10 end\_POSTSUPERSCRIPT ). We train all agent-task combinations except for Humanoid-Gather.
For all experiments, we use neural network policies with two hidden layers of size (64,32)6432(64,32)( 64 , 32 ). Our experiments are implemented in rllab (Duan et al., [2016](#bib.bib7)).

(a) Humanoid-Circle

(b) Point-Gather
Figure 2: The Humanoid-Circle and Point-Gather environments. In Humanoid-Circle, the safe area is between the blue panels.
###
8.1 Evaluating CPO and Comparison Analysis
Learning curves for CPO and PDO are compiled in Figure [1](#S7.F1 "Figure 1 ‣ 7 Connections to Prior Work ‣ Constrained Policy Optimization"). Note that we evaluate algorithm performance based on the C+superscript𝐶C^{+}italic\_C start\_POSTSUPERSCRIPT + end\_POSTSUPERSCRIPT return, instead of the C𝐶Citalic\_C return (except for in Point-Gather, where we did not use cost shaping due to that environment’s short time horizon), because this is what the algorithm actually constrains in these experiments.
For our comparison, we implement PDO with ([16](#S7.E16 "16 ‣ 7 Connections to Prior Work ‣ Constrained Policy Optimization")) as the update rule for the dual variables, using a constant learning rate α𝛼\alphaitalic\_α; details are available in supplementary material (Section [10.3.3](#S10.SS3.SSS3 "10.3.3 Primal-Dual Optimization Implementation ‣ 10.3 Experimental Parameters ‣ 10 Appendix ‣ Constrained Policy Optimization")). We emphasize that in order for the comparison to be fair, we give PDO every advantage that is given to CPO, including equivalent trust region policy updates. To benchmark the environments, we also include TRPO (trust region policy optimization) (Schulman et al., [2015](#bib.bib22)), a state-of-the-art unconstrained reinforcement learning algorithm. The TRPO experiments show that optimal unconstrained behaviors for these environments are constraint-violating.
We find that CPO is successful at approximately enforcing constraints in all environments. In the simpler environments (Point-Circle and Point-Gather), CPO tracks the constraint return almost exactly to the limit value.
By contrast, although PDO usually converges to constraint-satisfying policies in the end, it is not consistently constraint-satisfying throughout training (as expected). For example, see the spike in constraint value that it experiences in Ant-Circle. Additionally, PDO is sensitive to the initialization of the dual variable. By default, we initialize ν0=0subscript𝜈00\nu\_{0}=0italic\_ν start\_POSTSUBSCRIPT 0 end\_POSTSUBSCRIPT = 0, which exploits no prior knowledge about the environment and makes sense when the initial policies are feasible. However, it may seem appealing to set ν0subscript𝜈0\nu\_{0}italic\_ν start\_POSTSUBSCRIPT 0 end\_POSTSUBSCRIPT high, which would make PDO more conservative with respect to the constraint; PDO could then decrease ν𝜈\nuitalic\_ν as necessary after the fact. In the Point environments, we experiment with ν0=1000subscript𝜈01000\nu\_{0}=1000italic\_ν start\_POSTSUBSCRIPT 0 end\_POSTSUBSCRIPT = 1000 and show that although this does assure constraint satisfaction, it also can substantially harm performance with respect to return. Furthermore, we argue that this is not adequate in general: after the dual variable decreases, the agent could learn a new behavior that increases the correct dual variable more quickly than PDO can attain it (as happens in Ant-Circle for PDO; observe that performance is approximately constraint-satisfying until the agent learns how to run at around iteration 350).
We find that CPO generally outperforms PDO on enforcing constraints, without compromising performance with respect to return. CPO quickly stabilizes the constraint return around to the limit value, while PDO is not consistently able to enforce constraints all throughout training.
###
8.2 Ablation on Cost Shaping
In Figure [3](#S8.F3 "Figure 3 ‣ 8.2 Ablation on Cost Shaping ‣ 8 Experiments ‣ Constrained Policy Optimization"), we compare performance of CPO with and without cost shaping in the constraint. Our metric for comparison is the C𝐶Citalic\_C-return, the ‘true’ constraint. The cost shaping does help, almost completely accounting for CPO’s inherent approximation errors. However, CPO is nearly constraint-satisfying even without cost shaping.

(a) Ant-Circle Return

(b) Ant-Gather Return

(c) Ant-Circle C𝐶Citalic\_C-Return

(d) Ant-Gather C𝐶Citalic\_C-Return
Figure 3: Using cost shaping (CS) in the constraint while optimizing generally improves the agent’s adherence to the true constraint on C𝐶Citalic\_C-return.
###
8.3 Constraint vs. Fixed Penalty

(a) Ant-Circle Return

(b) Ant-Circle C+superscript𝐶C^{+}italic\_C start\_POSTSUPERSCRIPT + end\_POSTSUPERSCRIPT-Return
Figure 4: Comparison between CPO and FPO (fixed penalty optimization) for various values of fixed penalty.
In Figure [4](#S8.F4 "Figure 4 ‣ 8.3 Constraint vs. Fixed Penalty ‣ 8 Experiments ‣ Constrained Policy Optimization"), we compare CPO to a fixed penalty method, where policies are learned using TRPO with rewards R(s,a,s′)+λC+(s,a,s′)𝑅𝑠𝑎superscript𝑠′𝜆superscript𝐶𝑠𝑎superscript𝑠′R(s,a,s^{\prime})+\lambda C^{+}(s,a,s^{\prime})italic\_R ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) + italic\_λ italic\_C start\_POSTSUPERSCRIPT + end\_POSTSUPERSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) for λ∈{1,5,50}𝜆1550\lambda\in\{1,5,50\}italic\_λ ∈ { 1 , 5 , 50 }.
We find that fixed penalty methods can be highly sensitive to the choice of penalty coefficient: in Ant-Circle, a penalty coefficient of 1 results in reward-maximizing policies that accumulate massive constraint costs, while a coefficient of 5 (less than an order of magnitude difference) results in cost-minimizing policies that never learn how to acquire any rewards. In contrast, CPO automatically picks penalty coefficients to attain the desired trade-off between reward and constraint cost.
9 Discussion
-------------
In this article, we showed that a particular optimization problem results in policy updates that are guaranteed to both improve return and satisfy constraints. This enabled the development of CPO, our policy search algorithm for CMDPs, which approximates the theoretically-guaranteed algorithm in a principled way. We demonstrated that CPO can train neural network policies with thousands of parameters on high-dimensional constrained control tasks, simultaneously maximizing reward and approximately satisfying constraints. Our work represents a step towards applying reinforcement learning in the real world, where constraints on agent behavior are sometimes necessary for the sake of safety.
Acknowledgements
----------------
The authors would like to acknowledge Peter Chen, who independently and concurrently derived an equivalent policy improvement bound.
Joshua Achiam is supported by TRUST (Team for Research in Ubiquitous Secure Technology) which receives support from NSF (award number CCF-0424422). This project also received support from Berkeley Deep Drive and from Siemens.
10 Appendix
------------
###
10.1 Proof of Policy Performance Bound
####
10.1.1 Preliminaries
Our analysis will make extensive use of the discounted future state distribution, dπsuperscript𝑑𝜋d^{\pi}italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT, which is defined as
| | | |
| --- | --- | --- |
| | dπ(s)=(1−γ)∑t=0∞γtP(st=s|π).superscript𝑑𝜋𝑠1𝛾superscriptsubscript𝑡0superscript𝛾𝑡𝑃subscript𝑠𝑡conditional𝑠𝜋d^{\pi}(s)=(1-\gamma)\sum\_{t=0}^{\infty}\gamma^{t}P(s\_{t}=s|\pi).italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s ) = ( 1 - italic\_γ ) ∑ start\_POSTSUBSCRIPT italic\_t = 0 end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT ∞ end\_POSTSUPERSCRIPT italic\_γ start\_POSTSUPERSCRIPT italic\_t end\_POSTSUPERSCRIPT italic\_P ( italic\_s start\_POSTSUBSCRIPT italic\_t end\_POSTSUBSCRIPT = italic\_s | italic\_π ) . | |
It allows us to express the expected discounted total reward compactly as
| | | | |
| --- | --- | --- | --- |
| | J(π)=11−γEs∼dπa∼πs′∼P[R(s,a,s′)],𝐽𝜋11𝛾similar-to𝑠superscript𝑑𝜋similar-to𝑎𝜋similar-tosuperscript𝑠′𝑃
Edelimited-[]𝑅𝑠𝑎superscript𝑠′J(\pi)=\frac{1}{1-\gamma}\underset{\begin{subarray}{c}s\sim d^{\pi}\\
a\sim\pi\\
s^{\prime}\sim P\end{subarray}}{{\mathrm{E}}}\left[R(s,a,s^{\prime})\right],italic\_J ( italic\_π ) = divide start\_ARG 1 end\_ARG start\_ARG 1 - italic\_γ end\_ARG start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW start\_ROW start\_CELL italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_R ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ] , | | (17) |
where by a∼πsimilar-to𝑎𝜋a\sim\piitalic\_a ∼ italic\_π, we mean a∼π(⋅|s)a\sim\pi(\cdot|s)italic\_a ∼ italic\_π ( ⋅ | italic\_s ), and by s′∼Psimilar-tosuperscript𝑠′𝑃s^{\prime}\sim Pitalic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P, we mean s′∼P(⋅|s,a)s^{\prime}\sim P(\cdot|s,a)italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P ( ⋅ | italic\_s , italic\_a ). We drop the explicit notation for the sake of reducing clutter, but it should be clear from context that a𝑎aitalic\_a and s′superscript𝑠′s^{\prime}italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT depend on s𝑠sitalic\_s.
First, we examine some useful properties of dπsuperscript𝑑𝜋d^{\pi}italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT that become apparent in vector form for finite state spaces. Let pπt∈ℝ|S|subscriptsuperscript𝑝𝑡𝜋superscriptℝ𝑆p^{t}\_{\pi}\in{{\mathbb{R}}^{|S|}}italic\_p start\_POSTSUPERSCRIPT italic\_t end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT ∈ blackboard\_R start\_POSTSUPERSCRIPT | italic\_S | end\_POSTSUPERSCRIPT denote the vector with components pπt(s)=P(st=s|π)subscriptsuperscript𝑝𝑡𝜋𝑠𝑃subscript𝑠𝑡conditional𝑠𝜋p^{t}\_{\pi}(s)=P(s\_{t}=s|\pi)italic\_p start\_POSTSUPERSCRIPT italic\_t end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT ( italic\_s ) = italic\_P ( italic\_s start\_POSTSUBSCRIPT italic\_t end\_POSTSUBSCRIPT = italic\_s | italic\_π ), and let Pπ∈ℝ|S|×|S|subscript𝑃𝜋superscriptℝ𝑆𝑆P\_{\pi}\in{{\mathbb{R}}^{|S|\times|S|}}italic\_P start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT ∈ blackboard\_R start\_POSTSUPERSCRIPT | italic\_S | × | italic\_S | end\_POSTSUPERSCRIPT denote the transition matrix with components Pπ(s′|s)=∫𝑑aP(s′|s,a)π(a|s)subscript𝑃𝜋conditionalsuperscript𝑠′𝑠differential-d𝑎𝑃conditionalsuperscript𝑠′𝑠𝑎𝜋conditional𝑎𝑠P\_{\pi}(s^{\prime}|s)=\int daP(s^{\prime}|s,a)\pi(a|s)italic\_P start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | italic\_s ) = ∫ italic\_d italic\_a italic\_P ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | italic\_s , italic\_a ) italic\_π ( italic\_a | italic\_s ); then pπt=Pπpπt−1=Pπtμsubscriptsuperscript𝑝𝑡𝜋subscript𝑃𝜋subscriptsuperscript𝑝𝑡1𝜋subscriptsuperscript𝑃𝑡𝜋𝜇p^{t}\_{\pi}=P\_{\pi}p^{t-1}\_{\pi}=P^{t}\_{\pi}\muitalic\_p start\_POSTSUPERSCRIPT italic\_t end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT = italic\_P start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT italic\_p start\_POSTSUPERSCRIPT italic\_t - 1 end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT = italic\_P start\_POSTSUPERSCRIPT italic\_t end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT italic\_μ and
| | | | | | |
| --- | --- | --- | --- | --- | --- |
| | dπsuperscript𝑑𝜋\displaystyle d^{\pi}italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT | =\displaystyle== | (1−γ)∑t=0∞(γPπ)tμ1𝛾superscriptsubscript𝑡0superscript𝛾subscript𝑃𝜋𝑡𝜇\displaystyle(1-\gamma)\sum\_{t=0}^{\infty}(\gamma P\_{\pi})^{t}\mu( 1 - italic\_γ ) ∑ start\_POSTSUBSCRIPT italic\_t = 0 end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT ∞ end\_POSTSUPERSCRIPT ( italic\_γ italic\_P start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT ) start\_POSTSUPERSCRIPT italic\_t end\_POSTSUPERSCRIPT italic\_μ | | (18) |
| | | =\displaystyle== | (1−γ)(I−γPπ)−1μ.1𝛾superscript𝐼𝛾subscript𝑃𝜋1𝜇\displaystyle(1-\gamma)(I-\gamma P\_{\pi})^{-1}\mu.( 1 - italic\_γ ) ( italic\_I - italic\_γ italic\_P start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT ) start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_μ . | |
This formulation helps us easily obtain the following lemma.
######
Lemma 1.
For any function f:S→ℝnormal-:𝑓normal-→𝑆ℝf:S\to{{\mathbb{R}}}italic\_f : italic\_S → blackboard\_R and any policy π𝜋\piitalic\_π,
| | | | |
| --- | --- | --- | --- |
| | (1−γ)Es∼μ[f(s)]+Es∼dπa∼πs′∼P[γf(s′)]−Es∼dπ[f(s)]=0.1𝛾similar-to𝑠𝜇Edelimited-[]𝑓𝑠similar-to𝑠superscript𝑑𝜋similar-to𝑎𝜋similar-tosuperscript𝑠′𝑃
Edelimited-[]𝛾𝑓superscript𝑠′similar-to𝑠superscript𝑑𝜋Edelimited-[]𝑓𝑠0(1-\gamma)\underset{s\sim\mu}{{\mathrm{E}}}\left[f(s)\right]+\underset{\begin{subarray}{c}s\sim d^{\pi}\\
a\sim\pi\\
s^{\prime}\sim P\end{subarray}}{{\mathrm{E}}}\left[\gamma f(s^{\prime})\right]-\underset{s\sim d^{\pi}}{{\mathrm{E}}}\left[f(s)\right]=0.( 1 - italic\_γ ) start\_UNDERACCENT italic\_s ∼ italic\_μ end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_f ( italic\_s ) ] + start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW start\_ROW start\_CELL italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_γ italic\_f ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ] - start\_UNDERACCENT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_f ( italic\_s ) ] = 0 . | | (19) |
###### Proof.
Multiply both sides of ([18](#S10.E18 "18 ‣ 10.1.1 Preliminaries ‣ 10.1 Proof of Policy Performance Bound ‣ 10 Appendix ‣ Constrained Policy Optimization")) by (I−γPπ)𝐼𝛾subscript𝑃𝜋(I-\gamma P\_{\pi})( italic\_I - italic\_γ italic\_P start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT ) and take the inner product with the vector f∈ℝ|S|𝑓superscriptℝ𝑆f\in{{\mathbb{R}}^{|S|}}italic\_f ∈ blackboard\_R start\_POSTSUPERSCRIPT | italic\_S | end\_POSTSUPERSCRIPT.
∎
Combining this with ([17](#S10.E17 "17 ‣ 10.1.1 Preliminaries ‣ 10.1 Proof of Policy Performance Bound ‣ 10 Appendix ‣ Constrained Policy Optimization")), we obtain the following, for any function f𝑓fitalic\_f and any policy π𝜋\piitalic\_π:
| | | | |
| --- | --- | --- | --- |
| | J(π)=Es∼μ[f(s)]+11−γEs∼dπa∼πs′∼P[R(s,a,s′)+γf(s′)−f(s)].𝐽𝜋similar-to𝑠𝜇Edelimited-[]𝑓𝑠11𝛾similar-to𝑠superscript𝑑𝜋similar-to𝑎𝜋similar-tosuperscript𝑠′𝑃
Edelimited-[]𝑅𝑠𝑎superscript𝑠′𝛾𝑓superscript𝑠′𝑓𝑠J(\pi)=\underset{s\sim\mu}{{\mathrm{E}}}[f(s)]+\frac{1}{1-\gamma}\underset{\begin{subarray}{c}s\sim d^{\pi}\\
a\sim\pi\\
s^{\prime}\sim P\end{subarray}}{{\mathrm{E}}}\left[R(s,a,s^{\prime})+\gamma f(s^{\prime})-f(s)\right].italic\_J ( italic\_π ) = start\_UNDERACCENT italic\_s ∼ italic\_μ end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_f ( italic\_s ) ] + divide start\_ARG 1 end\_ARG start\_ARG 1 - italic\_γ end\_ARG start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW start\_ROW start\_CELL italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_R ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) + italic\_γ italic\_f ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_f ( italic\_s ) ] . | | (20) |
This identity is nice for two reasons. First: if we pick f𝑓fitalic\_f to be an approximator of the value function Vπsuperscript𝑉𝜋V^{\pi}italic\_V start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT, then ([20](#S10.E20 "20 ‣ 10.1.1 Preliminaries ‣ 10.1 Proof of Policy Performance Bound ‣ 10 Appendix ‣ Constrained Policy Optimization")) relates the true discounted return of the policy (J(π)𝐽𝜋J(\pi)italic\_J ( italic\_π )) to the estimate of the policy return (Es∼μ[f(s)]subscriptEsimilar-to𝑠𝜇delimited-[]𝑓𝑠{\mathrm{E}}\_{s\sim\mu}[f(s)]roman\_E start\_POSTSUBSCRIPT italic\_s ∼ italic\_μ end\_POSTSUBSCRIPT [ italic\_f ( italic\_s ) ]) and to the on-policy average TD-error of the approximator; this is aesthetically satisfying. Second: it shows that reward-shaping by γf(s′)−f(s)𝛾𝑓superscript𝑠′𝑓𝑠\gamma f(s^{\prime})-f(s)italic\_γ italic\_f ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_f ( italic\_s ) has the effect of translating the total discounted return by Es∼μ[f(s)]subscriptEsimilar-to𝑠𝜇delimited-[]𝑓𝑠{\mathrm{E}}\_{s\sim\mu}[f(s)]roman\_E start\_POSTSUBSCRIPT italic\_s ∼ italic\_μ end\_POSTSUBSCRIPT [ italic\_f ( italic\_s ) ], a fixed constant independent of policy; this illustrates the finding of Ng. et al. ([1999](#bib.bib19)) that reward shaping by γf(s′)+f(s)𝛾𝑓superscript𝑠′𝑓𝑠\gamma f(s^{\prime})+f(s)italic\_γ italic\_f ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) + italic\_f ( italic\_s ) does not change the optimal policy.
It is also helpful to introduce an identity for the vector difference of the discounted future state visitation distributions on two different policies, π′superscript𝜋′\pi^{\prime}italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT and π𝜋\piitalic\_π. Define the matrices G≐(I−γPπ)−1approaches-limit𝐺superscript𝐼𝛾subscript𝑃𝜋1G\doteq(I-\gamma P\_{\pi})^{-1}italic\_G ≐ ( italic\_I - italic\_γ italic\_P start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT ) start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT, G¯≐(I−γPπ′)−1approaches-limit¯𝐺superscript𝐼𝛾subscript𝑃superscript𝜋′1\bar{G}\doteq(I-\gamma P\_{\pi^{\prime}})^{-1}over¯ start\_ARG italic\_G end\_ARG ≐ ( italic\_I - italic\_γ italic\_P start\_POSTSUBSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT ) start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT, and Δ=Pπ′−PπΔsubscript𝑃superscript𝜋′subscript𝑃𝜋\Delta=P\_{\pi^{\prime}}-P\_{\pi}roman\_Δ = italic\_P start\_POSTSUBSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT - italic\_P start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT. Then:
| | | | | |
| --- | --- | --- | --- | --- |
| | G−1−G¯−1superscript𝐺1superscript¯𝐺1\displaystyle G^{-1}-\bar{G}^{-1}italic\_G start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT - over¯ start\_ARG italic\_G end\_ARG start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT | =\displaystyle== | (I−γPπ)−(I−γPπ′)𝐼𝛾subscript𝑃𝜋𝐼𝛾subscript𝑃superscript𝜋′\displaystyle(I-\gamma P\_{\pi})-(I-\gamma P\_{\pi^{\prime}})( italic\_I - italic\_γ italic\_P start\_POSTSUBSCRIPT italic\_π end\_POSTSUBSCRIPT ) - ( italic\_I - italic\_γ italic\_P start\_POSTSUBSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT ) | |
| | | =\displaystyle== | γΔ;𝛾Δ\displaystyle\gamma\Delta;italic\_γ roman\_Δ ; | |
left-multiplying by G𝐺Gitalic\_G and right-multiplying by G¯¯𝐺\bar{G}over¯ start\_ARG italic\_G end\_ARG, we obtain
| | | |
| --- | --- | --- |
| | G¯−G=γG¯ΔG.¯𝐺𝐺𝛾¯𝐺Δ𝐺\bar{G}-G=\gamma\bar{G}\Delta G.over¯ start\_ARG italic\_G end\_ARG - italic\_G = italic\_γ over¯ start\_ARG italic\_G end\_ARG roman\_Δ italic\_G . | |
Thus
| | | | | | |
| --- | --- | --- | --- | --- | --- |
| | dπ′−dπsuperscript𝑑superscript𝜋′superscript𝑑𝜋\displaystyle d^{\pi^{\prime}}-d^{\pi}italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT - italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT | =\displaystyle== | (1−γ)(G¯−G)μ1𝛾¯𝐺𝐺𝜇\displaystyle(1-\gamma)\left(\bar{G}-G\right)\mu( 1 - italic\_γ ) ( over¯ start\_ARG italic\_G end\_ARG - italic\_G ) italic\_μ | | (21) |
| | | =\displaystyle== | γ(1−γ)G¯ΔGμ𝛾1𝛾¯𝐺Δ𝐺𝜇\displaystyle\gamma(1-\gamma)\bar{G}\Delta G\muitalic\_γ ( 1 - italic\_γ ) over¯ start\_ARG italic\_G end\_ARG roman\_Δ italic\_G italic\_μ | |
| | | =\displaystyle== | γG¯Δdπ.𝛾¯𝐺Δsuperscript𝑑𝜋\displaystyle\gamma\bar{G}\Delta d^{\pi}.italic\_γ over¯ start\_ARG italic\_G end\_ARG roman\_Δ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT . | |
For simplicity in what follows, we will only consider MDPs with finite state and action spaces, although our attention is on MDPs that are too large for tabular methods.
####
10.1.2 Main Results
In this section, we will derive and present the new policy improvement bound. We will begin with a lemma:
######
Lemma 2.
For any function f:S→ℝnormal-:𝑓normal-→𝑆ℝf:S\to{{\mathbb{R}}}italic\_f : italic\_S → blackboard\_R and any policies π′superscript𝜋normal-′\pi^{\prime}italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT and π𝜋\piitalic\_π, define
| | | | |
| --- | --- | --- | --- |
| | Lπ,f(π′)≐Es∼dπa∼πs′∼P[(π′(a|s)π(a|s)−1)(R(s,a,s′)+γf(s′)−f(s))],approaches-limitsubscript𝐿𝜋𝑓superscript𝜋′similar-to𝑠superscript𝑑𝜋similar-to𝑎𝜋similar-tosuperscript𝑠′𝑃
Edelimited-[]superscript𝜋′conditional𝑎𝑠𝜋conditional𝑎𝑠1𝑅𝑠𝑎superscript𝑠′𝛾𝑓superscript𝑠′𝑓𝑠L\_{\pi,f}(\pi^{\prime})\doteq\underset{\begin{subarray}{c}s\sim d^{\pi}\\
a\sim\pi\\
s^{\prime}\sim P\end{subarray}}{{\mathrm{E}}}\left[\left(\frac{\pi^{\prime}(a|s)}{\pi(a|s)}-1\right)\left(R(s,a,s^{\prime})+\gamma f(s^{\prime})-f(s)\right)\right],italic\_L start\_POSTSUBSCRIPT italic\_π , italic\_f end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ≐ start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW start\_ROW start\_CELL italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ ( divide start\_ARG italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ( italic\_a | italic\_s ) end\_ARG start\_ARG italic\_π ( italic\_a | italic\_s ) end\_ARG - 1 ) ( italic\_R ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) + italic\_γ italic\_f ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_f ( italic\_s ) ) ] , | | (22) |
and ϵfπ′≐maxs|Ea∼π′,s′∼P[R(s,a,s′)+γf(s′)−f(s)]|approaches-limitsuperscriptsubscriptitalic-ϵ𝑓superscript𝜋normal-′subscript𝑠subscriptnormal-Eformulae-sequencesimilar-to𝑎superscript𝜋normal-′similar-tosuperscript𝑠normal-′𝑃delimited-[]𝑅𝑠𝑎superscript𝑠normal-′𝛾𝑓superscript𝑠normal-′𝑓𝑠\epsilon\_{f}^{\pi^{\prime}}\doteq\max\_{s}\left|{\mathrm{E}}\_{a\sim\pi^{\prime},s^{\prime}\sim P}[R(s,a,s^{\prime})+\gamma f(s^{\prime})-f(s)]\right|italic\_ϵ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ≐ roman\_max start\_POSTSUBSCRIPT italic\_s end\_POSTSUBSCRIPT | roman\_E start\_POSTSUBSCRIPT italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_POSTSUBSCRIPT [ italic\_R ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) + italic\_γ italic\_f ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_f ( italic\_s ) ] |. Then the following bounds hold:
| | | | |
| --- | --- | --- | --- |
| | J(π′)−J(π)≥11−γ(Lπ,f(π′)−2ϵfπ′DTV(dπ′||dπ)),\displaystyle J(\pi^{\prime})-J(\pi)\geq\frac{1}{1-\gamma}\left(L\_{\pi,f}(\pi^{\prime})-2\epsilon\_{f}^{\pi^{\prime}}D\_{TV}(d^{\pi^{\prime}}||d^{\pi})\right),italic\_J ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_J ( italic\_π ) ≥ divide start\_ARG 1 end\_ARG start\_ARG 1 - italic\_γ end\_ARG ( italic\_L start\_POSTSUBSCRIPT italic\_π , italic\_f end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - 2 italic\_ϵ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT | | italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ) ) , | | (23) |
| | J(π′)−J(π)≤11−γ(Lπ,f(π′)+2ϵfπ′DTV(dπ′||dπ)),\displaystyle J(\pi^{\prime})-J(\pi)\leq\frac{1}{1-\gamma}\left(L\_{\pi,f}(\pi^{\prime})+2\epsilon\_{f}^{\pi^{\prime}}D\_{TV}(d^{\pi^{\prime}}||d^{\pi})\right),italic\_J ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_J ( italic\_π ) ≤ divide start\_ARG 1 end\_ARG start\_ARG 1 - italic\_γ end\_ARG ( italic\_L start\_POSTSUBSCRIPT italic\_π , italic\_f end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) + 2 italic\_ϵ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT | | italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ) ) , | | (24) |
where DTVsubscript𝐷𝑇𝑉D\_{TV}italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT is the total variational divergence. Furthermore, the bounds are tight (when π′=πsuperscript𝜋normal-′𝜋\pi^{\prime}=\piitalic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT = italic\_π, the LHS and RHS are identically zero).
###### Proof.
First, for notational convenience, let δf(s,a,s′)≐R(s,a,s′)+γf(s′)−f(s)approaches-limitsubscript𝛿𝑓𝑠𝑎superscript𝑠′𝑅𝑠𝑎superscript𝑠′𝛾𝑓superscript𝑠′𝑓𝑠\delta\_{f}(s,a,s^{\prime})\doteq R(s,a,s^{\prime})+\gamma f(s^{\prime})-f(s)italic\_δ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ≐ italic\_R ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) + italic\_γ italic\_f ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_f ( italic\_s ). (The choice of δ𝛿\deltaitalic\_δ to denote this quantity is intentionally suggestive—this bears a strong resemblance to a TD-error.) By ([20](#S10.E20 "20 ‣ 10.1.1 Preliminaries ‣ 10.1 Proof of Policy Performance Bound ‣ 10 Appendix ‣ Constrained Policy Optimization")), we obtain the identity
| | | |
| --- | --- | --- |
| | J(π′)−J(π)=11−γ(Es∼dπ′a∼π′s′∼P[δf(s,a,s′)]−Es∼dπa∼πs′∼P[δf(s,a,s′)].)J(\pi^{\prime})-J(\pi)=\frac{1}{1-\gamma}\left(\underset{\begin{subarray}{c}s\sim d^{\pi^{\prime}}\\
a\sim\pi^{\prime}\\
s^{\prime}\sim P\end{subarray}}{{\mathrm{E}}}\left[\delta\_{f}(s,a,s^{\prime})\right]-\underset{\begin{subarray}{c}s\sim d^{\pi}\\
a\sim\pi\\
s^{\prime}\sim P\end{subarray}}{{\mathrm{E}}}\left[\delta\_{f}(s,a,s^{\prime})\right].\right)italic\_J ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_J ( italic\_π ) = divide start\_ARG 1 end\_ARG start\_ARG 1 - italic\_γ end\_ARG ( start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_δ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ] - start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW start\_ROW start\_CELL italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_δ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ] . ) | |
Now, we restrict our attention to the first term in this equation. Let δ¯fπ′∈ℝ|S|superscriptsubscript¯𝛿𝑓superscript𝜋′superscriptℝ𝑆\bar{\delta}\_{f}^{\pi^{\prime}}\in{{\mathbb{R}}^{|S|}}over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ∈ blackboard\_R start\_POSTSUPERSCRIPT | italic\_S | end\_POSTSUPERSCRIPT denote the vector of components δ¯fπ′(s)=Ea∼π′,s′∼P[δf(s,a,s′)|s]superscriptsubscript¯𝛿𝑓superscript𝜋′𝑠subscriptEformulae-sequencesimilar-to𝑎superscript𝜋′similar-tosuperscript𝑠′𝑃delimited-[]conditionalsubscript𝛿𝑓𝑠𝑎superscript𝑠′𝑠\bar{\delta}\_{f}^{\pi^{\prime}}(s)={\mathrm{E}}\_{a\sim\pi^{\prime},s^{\prime}\sim P}[\delta\_{f}(s,a,s^{\prime})|s]over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ( italic\_s ) = roman\_E start\_POSTSUBSCRIPT italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_POSTSUBSCRIPT [ italic\_δ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) | italic\_s ]. Observe that
| | | |
| --- | --- | --- |
| | Es∼dπ′a∼π′s′∼P[δf(s,a,s′)]=⟨dπ′,δ¯fπ′⟩=⟨dπ,δ¯fπ′⟩+⟨dπ′−dπ,δ¯fπ′⟩similar-to𝑠superscript𝑑superscript𝜋′similar-to𝑎superscript𝜋′similar-tosuperscript𝑠′𝑃
Edelimited-[]subscript𝛿𝑓𝑠𝑎superscript𝑠′absentsuperscript𝑑superscript𝜋′superscriptsubscript¯𝛿𝑓superscript𝜋′missing-subexpressionabsentsuperscript𝑑𝜋superscriptsubscript¯𝛿𝑓superscript𝜋′
superscript𝑑superscript𝜋′superscript𝑑𝜋superscriptsubscript¯𝛿𝑓superscript𝜋′\displaystyle\begin{aligned} \underset{\begin{subarray}{c}s\sim d^{\pi^{\prime}}\\
a\sim\pi^{\prime}\\
s^{\prime}\sim P\end{subarray}}{{\mathrm{E}}}\left[\delta\_{f}(s,a,s^{\prime})\right]&=\left\langle d^{\pi^{\prime}},\bar{\delta}\_{f}^{\pi^{\prime}}\right\rangle\\
&=\left\langle d^{\pi},\bar{\delta}\_{f}^{\pi^{\prime}}\right\rangle+\left\langle d^{\pi^{\prime}}-d^{\pi},\bar{\delta}\_{f}^{\pi^{\prime}}\right\rangle\end{aligned}start\_ROW start\_CELL start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_δ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ] end\_CELL start\_CELL = ⟨ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT , over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ⟩ end\_CELL end\_ROW start\_ROW start\_CELL end\_CELL start\_CELL = ⟨ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT , over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ⟩ + ⟨ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT - italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT , over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ⟩ end\_CELL end\_ROW | |
This term is then straightforwardly bounded by applying Hölder’s inequality; for any p,q∈[1,∞]𝑝𝑞
1p,q\in[1,\infty]italic\_p , italic\_q ∈ [ 1 , ∞ ] such that 1/p+1/q=11𝑝1𝑞11/p+1/q=11 / italic\_p + 1 / italic\_q = 1, we have
| | | |
| --- | --- | --- |
| | ⟨dπ,δ¯fπ′⟩+‖dπ′−dπ‖p‖δ¯fπ′‖q≥Es∼dπ′a∼π′s′∼P[δf(s,a,s′)]≥⟨dπ,δ¯fπ′⟩−‖dπ′−dπ‖p‖δ¯fπ′‖q.superscript𝑑𝜋superscriptsubscript¯𝛿𝑓superscript𝜋′
subscriptnormsuperscript𝑑superscript𝜋′superscript𝑑𝜋𝑝subscriptnormsuperscriptsubscript¯𝛿𝑓superscript𝜋′𝑞similar-to𝑠superscript𝑑superscript𝜋′similar-to𝑎superscript𝜋′similar-tosuperscript𝑠′𝑃
Edelimited-[]subscript𝛿𝑓𝑠𝑎superscript𝑠′superscript𝑑𝜋superscriptsubscript¯𝛿𝑓superscript𝜋′
subscriptnormsuperscript𝑑superscript𝜋′superscript𝑑𝜋𝑝subscriptnormsuperscriptsubscript¯𝛿𝑓superscript𝜋′𝑞\left\langle d^{\pi},\bar{\delta}\_{f}^{\pi^{\prime}}\right\rangle+\left\|d^{\pi^{\prime}}-d^{\pi}\right\|\_{p}\left\|\bar{\delta}\_{f}^{\pi^{\prime}}\right\|\_{q}\geq\underset{\begin{subarray}{c}s\sim d^{\pi^{\prime}}\\
a\sim\pi^{\prime}\\
s^{\prime}\sim P\end{subarray}}{{\mathrm{E}}}\left[\delta\_{f}(s,a,s^{\prime})\right]\geq\left\langle d^{\pi},\bar{\delta}\_{f}^{\pi^{\prime}}\right\rangle-\left\|d^{\pi^{\prime}}-d^{\pi}\right\|\_{p}\left\|\bar{\delta}\_{f}^{\pi^{\prime}}\right\|\_{q}.⟨ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT , over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ⟩ + ∥ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT - italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT italic\_p end\_POSTSUBSCRIPT ∥ over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT italic\_q end\_POSTSUBSCRIPT ≥ start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_δ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ] ≥ ⟨ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT , over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ⟩ - ∥ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT - italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT italic\_p end\_POSTSUBSCRIPT ∥ over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT italic\_q end\_POSTSUBSCRIPT . | |
The lower bound leads to ([23](#S10.E23 "23 ‣ Lemma 2. ‣ 10.1.2 Main Results ‣ 10.1 Proof of Policy Performance Bound ‣ 10 Appendix ‣ Constrained Policy Optimization")), and the upper bound leads to ([24](#S10.E24 "24 ‣ Lemma 2. ‣ 10.1.2 Main Results ‣ 10.1 Proof of Policy Performance Bound ‣ 10 Appendix ‣ Constrained Policy Optimization")).
We choose p=1𝑝1p=1italic\_p = 1 and q=∞𝑞q=\inftyitalic\_q = ∞; however, we believe that this step is very interesting, and different choices for dealing with the inner product ⟨dπ′−dπ,δ¯fπ′⟩superscript𝑑superscript𝜋′superscript𝑑𝜋superscriptsubscript¯𝛿𝑓superscript𝜋′\left\langle d^{\pi^{\prime}}-d^{\pi},\bar{\delta}\_{f}^{\pi^{\prime}}\right\rangle⟨ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT - italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT , over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ⟩ may lead to novel and useful bounds.
With ∥dπ′−dπ∥1=2DTV(dπ′||dπ)\left\|d^{\pi^{\prime}}-d^{\pi}\right\|\_{1}=2D\_{TV}(d^{\pi^{\prime}}||d^{\pi})∥ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT - italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT = 2 italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT | | italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ) and ‖δ¯fπ′‖∞=ϵfπ′subscriptnormsuperscriptsubscript¯𝛿𝑓superscript𝜋′superscriptsubscriptitalic-ϵ𝑓superscript𝜋′\left\|\bar{\delta}\_{f}^{\pi^{\prime}}\right\|\_{\infty}=\epsilon\_{f}^{\pi^{\prime}}∥ over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT ∞ end\_POSTSUBSCRIPT = italic\_ϵ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT, the bounds are almost obtained. The last step is to observe that, by the importance sampling identity,
| | | | | |
| --- | --- | --- | --- | --- |
| | ⟨dπ,δ¯fπ′⟩superscript𝑑𝜋superscriptsubscript¯𝛿𝑓superscript𝜋′\displaystyle\left\langle d^{\pi},\bar{\delta}\_{f}^{\pi^{\prime}}\right\rangle⟨ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT , over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ⟩ | =\displaystyle== | Es∼dπa∼π′s′∼P[δf(s,a,s′)]similar-to𝑠superscript𝑑𝜋similar-to𝑎superscript𝜋′similar-tosuperscript𝑠′𝑃
Edelimited-[]subscript𝛿𝑓𝑠𝑎superscript𝑠′\displaystyle\underset{\begin{subarray}{c}s\sim d^{\pi}\\
a\sim\pi^{\prime}\\
s^{\prime}\sim P\end{subarray}}{{\mathrm{E}}}\left[\delta\_{f}(s,a,s^{\prime})\right]start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_δ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ] | |
| | | =\displaystyle== | Es∼dπa∼πs′∼P[(π′(a|s)π(a|s))δf(s,a,s′)].similar-to𝑠superscript𝑑𝜋similar-to𝑎𝜋similar-tosuperscript𝑠′𝑃
Edelimited-[]superscript𝜋′conditional𝑎𝑠𝜋conditional𝑎𝑠subscript𝛿𝑓𝑠𝑎superscript𝑠′\displaystyle\underset{\begin{subarray}{c}s\sim d^{\pi}\\
a\sim\pi\\
s^{\prime}\sim P\end{subarray}}{{\mathrm{E}}}\left[\left(\frac{\pi^{\prime}(a|s)}{\pi(a|s)}\right)\delta\_{f}(s,a,s^{\prime})\right].start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL italic\_a ∼ italic\_π end\_CELL end\_ROW start\_ROW start\_CELL italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ ( divide start\_ARG italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ( italic\_a | italic\_s ) end\_ARG start\_ARG italic\_π ( italic\_a | italic\_s ) end\_ARG ) italic\_δ start\_POSTSUBSCRIPT italic\_f end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) ] . | |
After grouping terms, the bounds are obtained.
∎
This lemma makes use of many ideas that have been explored before; for the special case of f=Vπ𝑓superscript𝑉𝜋f=V^{\pi}italic\_f = italic\_V start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT, this strategy (after bounding DTV(dπ′||dπ)D\_{TV}(d^{\pi^{\prime}}||d^{\pi})italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT | | italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT )) leads directly to some of the policy improvement bounds previously obtained by Pirotta et al. and Schulman et al. The form given here is slightly more general, however, because it allows for freedom in choosing f𝑓fitalic\_f.
######
Remark.
It is reasonable to ask if there is a choice of f𝑓fitalic\_f which maximizes the lower bound here. This turns out to trivially be f=Vπ′𝑓superscript𝑉superscript𝜋′f=V^{\pi^{\prime}}italic\_f = italic\_V start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT. Observe that Es′∼P[δVπ′(s,a,s′)|s,a]=Aπ′(s,a)subscriptEsimilar-tosuperscript𝑠′𝑃delimited-[]conditionalsubscript𝛿superscript𝑉superscript𝜋′𝑠𝑎superscript𝑠′𝑠𝑎superscript𝐴superscript𝜋′𝑠𝑎{\mathrm{E}}\_{s^{\prime}\sim P}\left[\delta\_{V^{\pi^{\prime}}}(s,a,s^{\prime})|s,a\right]=A^{\pi^{\prime}}(s,a)roman\_E start\_POSTSUBSCRIPT italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ∼ italic\_P end\_POSTSUBSCRIPT [ italic\_δ start\_POSTSUBSCRIPT italic\_V start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT ( italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) | italic\_s , italic\_a ] = italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ). For all states, Ea∼π′[Aπ′(s,a)]=0subscriptEsimilar-to𝑎superscript𝜋′delimited-[]superscript𝐴superscript𝜋′𝑠𝑎0{\mathrm{E}}\_{a\sim\pi^{\prime}}[A^{\pi^{\prime}}(s,a)]=0roman\_E start\_POSTSUBSCRIPT italic\_a ∼ italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT [ italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ] = 0 (by the definition of Aπ′superscript𝐴superscript𝜋′A^{\pi^{\prime}}italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT), thus δ¯Vπ′π′=0subscriptsuperscript¯𝛿superscript𝜋′superscript𝑉superscript𝜋′0\bar{\delta}^{\pi^{\prime}}\_{V^{\pi^{\prime}}}=0over¯ start\_ARG italic\_δ end\_ARG start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT start\_POSTSUBSCRIPT italic\_V start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT = 0 and ϵVπ′π′=0superscriptsubscriptitalic-ϵsuperscript𝑉superscript𝜋′superscript𝜋′0\epsilon\_{V^{\pi^{\prime}}}^{\pi^{\prime}}=0italic\_ϵ start\_POSTSUBSCRIPT italic\_V start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT = 0. Also, Lπ,Vπ′(π′)=−Es∼dπ,a∼π[Aπ′(s,a)]subscript𝐿𝜋superscript𝑉superscript𝜋′superscript𝜋′subscriptEformulae-sequencesimilar-to𝑠superscript𝑑𝜋similar-to𝑎𝜋delimited-[]superscript𝐴superscript𝜋′𝑠𝑎L\_{\pi,V^{\pi^{\prime}}}(\pi^{\prime})=-{\mathrm{E}}\_{s\sim d^{\pi},a\sim\pi}\left[A^{\pi^{\prime}}(s,a)\right]italic\_L start\_POSTSUBSCRIPT italic\_π , italic\_V start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) = - roman\_E start\_POSTSUBSCRIPT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT , italic\_a ∼ italic\_π end\_POSTSUBSCRIPT [ italic\_A start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT ( italic\_s , italic\_a ) ]; from ([20](#S10.E20 "20 ‣ 10.1.1 Preliminaries ‣ 10.1 Proof of Policy Performance Bound ‣ 10 Appendix ‣ Constrained Policy Optimization")) with f=Vπ′𝑓superscript𝑉superscript𝜋′f=V^{\pi^{\prime}}italic\_f = italic\_V start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT, we can see that this exactly equals J(π′)−J(π)𝐽superscript𝜋′𝐽𝜋J(\pi^{\prime})-J(\pi)italic\_J ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ) - italic\_J ( italic\_π ). Thus, for f=Vπ′𝑓superscript𝑉superscript𝜋′f=V^{\pi^{\prime}}italic\_f = italic\_V start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT, we recover an exact equality. While this is not practically useful to us (because, when we want to optimize a lower bound with respect to π′superscript𝜋′\pi^{\prime}italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT, it is too expensive to evaluate Vπ′superscript𝑉superscript𝜋′V^{\pi^{\prime}}italic\_V start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT for each candidate to be practical), it provides insight: the penalty coefficient on the divergence captures information about the mismatch between f𝑓fitalic\_f and Vπ′superscript𝑉superscript𝜋′V^{\pi^{\prime}}italic\_V start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT.
Next, we are interested in bounding the divergence term, ‖dπ′−dπ‖1subscriptnormsuperscript𝑑superscript𝜋′superscript𝑑𝜋1\|d^{\pi^{\prime}}-d^{\pi}\|\_{1}∥ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT - italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT. We give the following lemma; to the best of our knowledge, this is a new result.
######
Lemma 3.
The divergence between discounted future state visitation distributions, ‖dπ′−dπ‖1subscriptnormsuperscript𝑑superscript𝜋normal-′superscript𝑑𝜋1\|d^{\pi^{\prime}}-d^{\pi}\|\_{1}∥ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT - italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT, is bounded by an average divergence of the policies π′superscript𝜋normal-′\pi^{\prime}italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT and π𝜋\piitalic\_π:
| | | | |
| --- | --- | --- | --- |
| | ∥dπ′−dπ∥1≤2γ1−γEs∼dπ[DTV(π′||π)[s]],\|d^{\pi^{\prime}}-d^{\pi}\|\_{1}\leq\frac{2\gamma}{1-\gamma}\underset{s\sim d^{\pi}}{{\mathrm{E}}}\left[D\_{TV}(\pi^{\prime}||\pi)[s]\right],∥ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT - italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT ≤ divide start\_ARG 2 italic\_γ end\_ARG start\_ARG 1 - italic\_γ end\_ARG start\_UNDERACCENT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] ] , | | (27) |
where DTV(π′||π)[s]=(1/2)∑a|π′(a|s)−π(a|s)|D\_{TV}(\pi^{\prime}||\pi)[s]=(1/2)\sum\_{a}|\pi^{\prime}(a|s)-\pi(a|s)|italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] = ( 1 / 2 ) ∑ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT | italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ( italic\_a | italic\_s ) - italic\_π ( italic\_a | italic\_s ) |.
###### Proof.
First, using ([21](#S10.E21 "21 ‣ 10.1.1 Preliminaries ‣ 10.1 Proof of Policy Performance Bound ‣ 10 Appendix ‣ Constrained Policy Optimization")), we obtain
| | | | | |
| --- | --- | --- | --- | --- |
| | ‖dπ′−dπ‖1subscriptnormsuperscript𝑑superscript𝜋′superscript𝑑𝜋1\displaystyle\|d^{\pi^{\prime}}-d^{\pi}\|\_{1}∥ italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT - italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT | =\displaystyle== | γ‖G¯Δdπ‖1𝛾subscriptnorm¯𝐺Δsuperscript𝑑𝜋1\displaystyle\gamma\|\bar{G}\Delta d^{\pi}\|\_{1}italic\_γ ∥ over¯ start\_ARG italic\_G end\_ARG roman\_Δ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT | |
| | | ≤\displaystyle\leq≤ | γ‖G¯‖1‖Δdπ‖1.𝛾subscriptnorm¯𝐺1subscriptnormΔsuperscript𝑑𝜋1\displaystyle\gamma\|\bar{G}\|\_{1}\|\Delta d^{\pi}\|\_{1}.italic\_γ ∥ over¯ start\_ARG italic\_G end\_ARG ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT ∥ roman\_Δ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT . | |
‖G¯‖1subscriptnorm¯𝐺1\|\bar{G}\|\_{1}∥ over¯ start\_ARG italic\_G end\_ARG ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT is bounded by:
| | | |
| --- | --- | --- |
| | ‖G¯‖1=‖(I−γPπ′)−1‖1≤∑t=0∞γt‖Pπ′‖1t=(1−γ)−1subscriptnorm¯𝐺1subscriptnormsuperscript𝐼𝛾subscript𝑃superscript𝜋′11superscriptsubscript𝑡0superscript𝛾𝑡superscriptsubscriptnormsubscript𝑃superscript𝜋′1𝑡superscript1𝛾1\|\bar{G}\|\_{1}=\|(I-\gamma P\_{\pi^{\prime}})^{-1}\|\_{1}\leq\sum\_{t=0}^{\infty}\gamma^{t}\left\|P\_{\pi^{\prime}}\right\|\_{1}^{t}=(1-\gamma)^{-1}∥ over¯ start\_ARG italic\_G end\_ARG ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT = ∥ ( italic\_I - italic\_γ italic\_P start\_POSTSUBSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT ) start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT ≤ ∑ start\_POSTSUBSCRIPT italic\_t = 0 end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT ∞ end\_POSTSUPERSCRIPT italic\_γ start\_POSTSUPERSCRIPT italic\_t end\_POSTSUPERSCRIPT ∥ italic\_P start\_POSTSUBSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_t end\_POSTSUPERSCRIPT = ( 1 - italic\_γ ) start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT | |
To conclude the lemma, we bound ‖Δdπ‖1subscriptnormΔsuperscript𝑑𝜋1\|\Delta d^{\pi}\|\_{1}∥ roman\_Δ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT.
| | | | | |
| --- | --- | --- | --- | --- |
| | ‖Δdπ‖1subscriptnormΔsuperscript𝑑𝜋1\displaystyle\|\Delta d^{\pi}\|\_{1}∥ roman\_Δ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ∥ start\_POSTSUBSCRIPT 1 end\_POSTSUBSCRIPT | =\displaystyle== | ∑s′|∑sΔ(s′|s)dπ(s)|\displaystyle\sum\_{s^{\prime}}\left|\sum\_{s}\Delta(s^{\prime}|s)d^{\pi}(s)\right|∑ start\_POSTSUBSCRIPT italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT | ∑ start\_POSTSUBSCRIPT italic\_s end\_POSTSUBSCRIPT roman\_Δ ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | italic\_s ) italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s ) | | |
| | | ≤\displaystyle\leq≤ | ∑s,s′|Δ(s′|s)|dπ(s)\displaystyle\sum\_{s,s^{\prime}}\left|\Delta(s^{\prime}|s)\right|d^{\pi}(s)∑ start\_POSTSUBSCRIPT italic\_s , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT | roman\_Δ ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | italic\_s ) | italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s ) | |
| | | =\displaystyle== | ∑s,s′|∑aP(s′|s,a)(π′(a|s)−π(a|s))|dπ(s)\displaystyle\sum\_{s,s^{\prime}}\left|\sum\_{a}P(s^{\prime}|s,a)\left(\pi^{\prime}(a|s)-\pi(a|s)\right)\right|d^{\pi}(s)∑ start\_POSTSUBSCRIPT italic\_s , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT | ∑ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT italic\_P ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | italic\_s , italic\_a ) ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ( italic\_a | italic\_s ) - italic\_π ( italic\_a | italic\_s ) ) | italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s ) | |
| | | ≤\displaystyle\leq≤ | ∑s,a,s′P(s′|s,a)|π′(a|s)−π(a|s)|dπ(s)\displaystyle\sum\_{s,a,s^{\prime}}P(s^{\prime}|s,a)\left|\pi^{\prime}(a|s)-\pi(a|s)\right|d^{\pi}(s)∑ start\_POSTSUBSCRIPT italic\_s , italic\_a , italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUBSCRIPT italic\_P ( italic\_s start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | italic\_s , italic\_a ) | italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ( italic\_a | italic\_s ) - italic\_π ( italic\_a | italic\_s ) | italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s ) | |
| | | =\displaystyle== | ∑s,a|π′(a|s)−π(a|s)|dπ(s)\displaystyle\sum\_{s,a}\left|\pi^{\prime}(a|s)-\pi(a|s)\right|d^{\pi}(s)∑ start\_POSTSUBSCRIPT italic\_s , italic\_a end\_POSTSUBSCRIPT | italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT ( italic\_a | italic\_s ) - italic\_π ( italic\_a | italic\_s ) | italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ( italic\_s ) | |
| | | =\displaystyle== | 2Es∼dπ[DTV(π′||π)[s]].\displaystyle 2\underset{s\sim d^{\pi}}{{\mathrm{E}}}\left[D\_{TV}(\pi^{\prime}||\pi)[s]\right].2 start\_UNDERACCENT italic\_s ∼ italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT end\_UNDERACCENT start\_ARG roman\_E end\_ARG [ italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT | | italic\_π ) [ italic\_s ] ] . | |
∎
The new policy improvement bound follows immediately.
\performancebound
\*
###### Proof.
Begin with the bounds from lemma [2](#Thmlemma2 "Lemma 2. ‣ 10.1.2 Main Results ‣ 10.1 Proof of Policy Performance Bound ‣ 10 Appendix ‣ Constrained Policy Optimization") and bound the divergence DTV(dπ′||dπ)D\_{TV}(d^{\pi^{\prime}}||d^{\pi})italic\_D start\_POSTSUBSCRIPT italic\_T italic\_V end\_POSTSUBSCRIPT ( italic\_d start\_POSTSUPERSCRIPT italic\_π start\_POSTSUPERSCRIPT ′ end\_POSTSUPERSCRIPT end\_POSTSUPERSCRIPT | | italic\_d start\_POSTSUPERSCRIPT italic\_π end\_POSTSUPERSCRIPT ) by lemma [3](#Thmlemma3 "Lemma 3. ‣ 10.1.2 Main Results ‣ 10.1 Proof of Policy Performance Bound ‣ 10 Appendix ‣ Constrained Policy Optimization").
∎
###
10.2 Proof of Analytical Solution to LQCLP
######
Theorem 1 (Optimizing Linear Objective with Linear and Quadratic Constraints).
Consider the problem
| | | | |
| --- | --- | --- | --- |
| | p\*=minxsuperscript𝑝subscript𝑥\displaystyle p^{\*}=\min\_{x}italic\_p start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = roman\_min start\_POSTSUBSCRIPT italic\_x end\_POSTSUBSCRIPT | gTxsuperscript𝑔𝑇𝑥\displaystyle\;g^{T}xitalic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_x | |
| | s.t. | bTx+c≤0superscript𝑏𝑇𝑥𝑐0\displaystyle\;b^{T}x+c\leq 0italic\_b start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_x + italic\_c ≤ 0 | | (28) |
| | | xTHx≤δ,superscript𝑥𝑇𝐻𝑥𝛿\displaystyle\;x^{T}Hx\leq\delta,italic\_x start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H italic\_x ≤ italic\_δ , | |
where g,b,x∈ℝn𝑔𝑏𝑥
superscriptℝ𝑛g,b,x\in{{\mathbb{R}}^{n}}italic\_g , italic\_b , italic\_x ∈ blackboard\_R start\_POSTSUPERSCRIPT italic\_n end\_POSTSUPERSCRIPT, c,δ∈ℝ𝑐𝛿
ℝc,\delta\in{{\mathbb{R}}}italic\_c , italic\_δ ∈ blackboard\_R, δ>0𝛿0\delta>0italic\_δ > 0, H∈𝕊n𝐻superscript𝕊𝑛H\in{{\mathbb{S}}^{n}}italic\_H ∈ blackboard\_S start\_POSTSUPERSCRIPT italic\_n end\_POSTSUPERSCRIPT, and H≻0succeeds𝐻0H\succ 0italic\_H ≻ 0. When there is at least one strictly feasible point, the optimal point x\*superscript𝑥x^{\*}italic\_x start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT satisfies
| | | | |
| --- | --- | --- | --- |
| | x\*superscript𝑥\displaystyle x^{\*}italic\_x start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT | =−1λ\*H−1(g+ν\*b),absent1superscript𝜆superscript𝐻1𝑔superscript𝜈𝑏\displaystyle=-\frac{1}{\lambda^{\*}}H^{-1}\left(g+\nu^{\*}b\right),= - divide start\_ARG 1 end\_ARG start\_ARG italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT end\_ARG italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT ( italic\_g + italic\_ν start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT italic\_b ) , | |
where λ\*superscript𝜆\lambda^{\*}italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT and ν\*superscript𝜈\nu^{\*}italic\_ν start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT are defined by
| | | | |
| --- | --- | --- | --- |
| | ν\*superscript𝜈\displaystyle\nu^{\*}italic\_ν start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT | =(λ\*c−rs)+,absentsubscriptsuperscript𝜆𝑐𝑟𝑠\displaystyle=\left(\dfrac{\lambda^{\*}c-r}{s}\right)\_{+},= ( divide start\_ARG italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT italic\_c - italic\_r end\_ARG start\_ARG italic\_s end\_ARG ) start\_POSTSUBSCRIPT + end\_POSTSUBSCRIPT , | |
| | λ\*superscript𝜆\displaystyle\lambda^{\*}italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT | =argmaxλ≥0{fa(λ)≐12λ(r2s−q)+λ2(c2s−δ)−rcsif λc−r>0fb(λ)≐−12(qλ+λδ)otherwise,absentsubscript𝜆0casesapproaches-limitsubscript𝑓𝑎𝜆12𝜆superscript𝑟2𝑠𝑞𝜆2superscript𝑐2𝑠𝛿𝑟𝑐𝑠if 𝜆𝑐𝑟0approaches-limitsubscript𝑓𝑏𝜆12𝑞𝜆𝜆𝛿otherwise\displaystyle=\arg\max\_{\lambda\geq 0}\;\left\{\begin{array}[]{ll}f\_{a}(\lambda)\doteq\frac{1}{2\lambda}\left(\frac{r^{2}}{s}-q\right)+\frac{\lambda}{2}\left(\frac{c^{2}}{s}-\delta\right)-\frac{rc}{s}&\text{{if} }\lambda c-r>0\\
f\_{b}(\lambda)\doteq-\frac{1}{2}\left(\frac{q}{\lambda}+\lambda\delta\right)&\text{{otherwise}},\end{array}\right.= roman\_arg roman\_max start\_POSTSUBSCRIPT italic\_λ ≥ 0 end\_POSTSUBSCRIPT { start\_ARRAY start\_ROW start\_CELL italic\_f start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT ( italic\_λ ) ≐ divide start\_ARG 1 end\_ARG start\_ARG 2 italic\_λ end\_ARG ( divide start\_ARG italic\_r start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT end\_ARG start\_ARG italic\_s end\_ARG - italic\_q ) + divide start\_ARG italic\_λ end\_ARG start\_ARG 2 end\_ARG ( divide start\_ARG italic\_c start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT end\_ARG start\_ARG italic\_s end\_ARG - italic\_δ ) - divide start\_ARG italic\_r italic\_c end\_ARG start\_ARG italic\_s end\_ARG end\_CELL start\_CELL roman\_if italic\_λ italic\_c - italic\_r > 0 end\_CELL end\_ROW start\_ROW start\_CELL italic\_f start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT ( italic\_λ ) ≐ - divide start\_ARG 1 end\_ARG start\_ARG 2 end\_ARG ( divide start\_ARG italic\_q end\_ARG start\_ARG italic\_λ end\_ARG + italic\_λ italic\_δ ) end\_CELL start\_CELL otherwise , end\_CELL end\_ROW end\_ARRAY | |
with q=gTH−1g𝑞superscript𝑔𝑇superscript𝐻1𝑔q=g^{T}H^{-1}gitalic\_q = italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_g, r=gTH−1b𝑟superscript𝑔𝑇superscript𝐻1𝑏r=g^{T}H^{-1}bitalic\_r = italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_b, and s=bTH−1b𝑠superscript𝑏𝑇superscript𝐻1𝑏s=b^{T}H^{-1}bitalic\_s = italic\_b start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_b.
Furthermore, let Λa≐{λ|λc−r>0,λ≥0}approaches-limitsubscriptnormal-Λ𝑎conditional-set𝜆formulae-sequence𝜆𝑐𝑟0𝜆0\Lambda\_{a}\doteq\{\lambda|\lambda c-r>0,\lambda\geq 0\}roman\_Λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT ≐ { italic\_λ | italic\_λ italic\_c - italic\_r > 0 , italic\_λ ≥ 0 }, and Λb≐{λ|λc−r≤0,λ≥0}approaches-limitsubscriptnormal-Λ𝑏conditional-set𝜆formulae-sequence𝜆𝑐𝑟0𝜆0\Lambda\_{b}\doteq\{\lambda|\lambda c-r\leq 0,\lambda\geq 0\}roman\_Λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT ≐ { italic\_λ | italic\_λ italic\_c - italic\_r ≤ 0 , italic\_λ ≥ 0 }. The value of λ\*superscript𝜆\lambda^{\*}italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT satisfies
| | | |
| --- | --- | --- |
| | λ\*∈{λa\*≐Proj(q−r2/sδ−c2/s,Λa),λb\*≐Proj(qδ,Λb)},superscript𝜆formulae-sequenceapproaches-limitsuperscriptsubscript𝜆𝑎Proj𝑞superscript𝑟2𝑠𝛿superscript𝑐2𝑠subscriptΛ𝑎approaches-limitsuperscriptsubscript𝜆𝑏Proj𝑞𝛿subscriptΛ𝑏\lambda^{\*}\in\left\{\lambda\_{a}^{\*}\doteq\text{{Proj}}\left(\sqrt{\frac{q-r^{2}/s}{\delta-c^{2}/s}},\Lambda\_{a}\right),\lambda\_{b}^{\*}\doteq\text{{Proj}}\left(\sqrt{\frac{q}{\delta}},\Lambda\_{b}\right)\right\},italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ∈ { italic\_λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ≐ Proj ( square-root start\_ARG divide start\_ARG italic\_q - italic\_r start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT / italic\_s end\_ARG start\_ARG italic\_δ - italic\_c start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT / italic\_s end\_ARG end\_ARG , roman\_Λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT ) , italic\_λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ≐ Proj ( square-root start\_ARG divide start\_ARG italic\_q end\_ARG start\_ARG italic\_δ end\_ARG end\_ARG , roman\_Λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT ) } , | |
with λ\*=λa\*superscript𝜆superscriptsubscript𝜆𝑎\lambda^{\*}=\lambda\_{a}^{\*}italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = italic\_λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT if fa(λa\*)>fb(λb\*)subscript𝑓𝑎superscriptsubscript𝜆𝑎subscript𝑓𝑏superscriptsubscript𝜆𝑏f\_{a}(\lambda\_{a}^{\*})>f\_{b}(\lambda\_{b}^{\*})italic\_f start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT ( italic\_λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ) > italic\_f start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT ( italic\_λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ) and λ\*=λb\*superscript𝜆superscriptsubscript𝜆𝑏\lambda^{\*}=\lambda\_{b}^{\*}italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = italic\_λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT otherwise, and 𝑃𝑟𝑜𝑗(a,S)𝑃𝑟𝑜𝑗𝑎𝑆\text{Proj}(a,S)Proj ( italic\_a , italic\_S ) is the projection of a point x𝑥xitalic\_x on to a set S𝑆Sitalic\_S. Note: the projection of a point x∈ℝ𝑥ℝx\in{{\mathbb{R}}}italic\_x ∈ blackboard\_R onto a convex segment of ℝℝ{{\mathbb{R}}}blackboard\_R, [a,b]𝑎𝑏[a,b][ italic\_a , italic\_b ], has value 𝑃𝑟𝑜𝑗(x,[a,b])=max(a,min(b,x))𝑃𝑟𝑜𝑗𝑥𝑎𝑏𝑎𝑏𝑥\text{Proj}(x,[a,b])=\max(a,\min(b,x))Proj ( italic\_x , [ italic\_a , italic\_b ] ) = roman\_max ( italic\_a , roman\_min ( italic\_b , italic\_x ) ).
###### Proof.
This is a convex optimization problem. When there is at least one strictly feasible point, strong duality holds by Slater’s theorem. We exploit strong duality to solve the problem analytically.
| | | | |
| --- | --- | --- | --- |
| | p\*superscript𝑝\displaystyle p^{\*}italic\_p start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT | =minxmaxλ≥0ν≥0gTx+λ2(xTHx−δ)+ν(bTx+c)absentsubscript𝑥𝜆0𝜈0
superscript𝑔𝑇𝑥𝜆2superscript𝑥𝑇𝐻𝑥𝛿𝜈superscript𝑏𝑇𝑥𝑐\displaystyle=\min\_{x}\underset{\begin{subarray}{c}\lambda\geq 0\\
\nu\geq 0\end{subarray}}{\max}\;g^{T}x+\frac{\lambda}{2}\left(x^{T}Hx-\delta\right)+\nu\left(b^{T}x+c\right)= roman\_min start\_POSTSUBSCRIPT italic\_x end\_POSTSUBSCRIPT start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_λ ≥ 0 end\_CELL end\_ROW start\_ROW start\_CELL italic\_ν ≥ 0 end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_max end\_ARG italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_x + divide start\_ARG italic\_λ end\_ARG start\_ARG 2 end\_ARG ( italic\_x start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H italic\_x - italic\_δ ) + italic\_ν ( italic\_b start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_x + italic\_c ) | |
| | | =maxλ≥0ν≥0minxλ2xTHx+(g+νb)Tx+(νc−12λδ)absent𝜆0𝜈0
subscript𝑥𝜆2superscript𝑥𝑇𝐻𝑥superscript𝑔𝜈𝑏𝑇𝑥𝜈𝑐12𝜆𝛿\displaystyle=\underset{\begin{subarray}{c}\lambda\geq 0\\
\nu\geq 0\end{subarray}}{\max}\min\_{x}\;\frac{\lambda}{2}x^{T}Hx+\left(g+\nu b\right)^{T}x+\left(\nu c-\frac{1}{2}\lambda\delta\right)= start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_λ ≥ 0 end\_CELL end\_ROW start\_ROW start\_CELL italic\_ν ≥ 0 end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_max end\_ARG roman\_min start\_POSTSUBSCRIPT italic\_x end\_POSTSUBSCRIPT divide start\_ARG italic\_λ end\_ARG start\_ARG 2 end\_ARG italic\_x start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H italic\_x + ( italic\_g + italic\_ν italic\_b ) start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_x + ( italic\_ν italic\_c - divide start\_ARG 1 end\_ARG start\_ARG 2 end\_ARG italic\_λ italic\_δ ) | Strong duality | |
| | | ⟹x\*=−1λH−1(g+νb)absentsuperscript𝑥1𝜆superscript𝐻1𝑔𝜈𝑏\displaystyle\;\;\;\implies x^{\*}=-\frac{1}{\lambda}H^{-1}\left(g+\nu b\right)⟹ italic\_x start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = - divide start\_ARG 1 end\_ARG start\_ARG italic\_λ end\_ARG italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT ( italic\_g + italic\_ν italic\_b ) | ∇xℒ(x,λ,ν)=0subscript∇𝑥ℒ𝑥𝜆𝜈0\displaystyle\nabla\_{x}{\mathcal{L}}(x,\lambda,\nu)=0∇ start\_POSTSUBSCRIPT italic\_x end\_POSTSUBSCRIPT caligraphic\_L ( italic\_x , italic\_λ , italic\_ν ) = 0 | |
| | | =maxλ≥0ν≥0−12λ(g+νb)TH−1(g+νb)+(νc−12λδ)absent𝜆0𝜈0
12𝜆superscript𝑔𝜈𝑏𝑇superscript𝐻1𝑔𝜈𝑏𝜈𝑐12𝜆𝛿\displaystyle=\underset{\begin{subarray}{c}\lambda\geq 0\\
\nu\geq 0\end{subarray}}{\max}\;-\frac{1}{2\lambda}\left(g+\nu b\right)^{T}H^{-1}\left(g+\nu b\right)+\left(\nu c-\frac{1}{2}\lambda\delta\right)= start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_λ ≥ 0 end\_CELL end\_ROW start\_ROW start\_CELL italic\_ν ≥ 0 end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_max end\_ARG - divide start\_ARG 1 end\_ARG start\_ARG 2 italic\_λ end\_ARG ( italic\_g + italic\_ν italic\_b ) start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT ( italic\_g + italic\_ν italic\_b ) + ( italic\_ν italic\_c - divide start\_ARG 1 end\_ARG start\_ARG 2 end\_ARG italic\_λ italic\_δ ) | Plug in x\*Plug in superscript𝑥\displaystyle\text{Plug in }x^{\*}Plug in italic\_x start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT | |
| | | =maxλ≥0ν≥0−12λ(q+2νr+ν2s)+(νc−12λδ)absent𝜆0𝜈0
12𝜆𝑞2𝜈𝑟superscript𝜈2𝑠𝜈𝑐12𝜆𝛿\displaystyle=\underset{\begin{subarray}{c}\lambda\geq 0\\
\nu\geq 0\end{subarray}}{\max}\;-\frac{1}{2\lambda}\left(q+2\nu r+\nu^{2}s\right)+\left(\nu c-\frac{1}{2}\lambda\delta\right)= start\_UNDERACCENT start\_ARG start\_ROW start\_CELL italic\_λ ≥ 0 end\_CELL end\_ROW start\_ROW start\_CELL italic\_ν ≥ 0 end\_CELL end\_ROW end\_ARG end\_UNDERACCENT start\_ARG roman\_max end\_ARG - divide start\_ARG 1 end\_ARG start\_ARG 2 italic\_λ end\_ARG ( italic\_q + 2 italic\_ν italic\_r + italic\_ν start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT italic\_s ) + ( italic\_ν italic\_c - divide start\_ARG 1 end\_ARG start\_ARG 2 end\_ARG italic\_λ italic\_δ ) | Notation: q≐gTH−1g,r≐gTH−1b,s≐bTH−1b.formulae-sequenceapproaches-limitNotation: 𝑞superscript𝑔𝑇superscript𝐻1𝑔formulae-sequenceapproaches-limit𝑟superscript𝑔𝑇superscript𝐻1𝑏approaches-limit𝑠superscript𝑏𝑇superscript𝐻1𝑏\displaystyle\text{Notation: }q\doteq g^{T}H^{-1}g,\;\;r\doteq g^{T}H^{-1}b,\;\;s\doteq b^{T}H^{-1}b.Notation: italic\_q ≐ italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_g , italic\_r ≐ italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_b , italic\_s ≐ italic\_b start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_b . | |
| | | ⟹∂ℒ∂ν=−12λ(2r+2νs)+cabsentℒ𝜈12𝜆2𝑟2𝜈𝑠𝑐\displaystyle\;\;\;\implies\frac{\partial{{\mathcal{L}}}}{\partial{\nu}}=-\frac{1}{2\lambda}\left(2r+2\nu s\right)+c⟹ divide start\_ARG ∂ caligraphic\_L end\_ARG start\_ARG ∂ italic\_ν end\_ARG = - divide start\_ARG 1 end\_ARG start\_ARG 2 italic\_λ end\_ARG ( 2 italic\_r + 2 italic\_ν italic\_s ) + italic\_c | |
| | | ⟹ν=(λc−rs)+absent𝜈subscript𝜆𝑐𝑟𝑠\displaystyle\;\;\;\implies\nu=\left(\frac{\lambda c-r}{s}\right)\_{+}⟹ italic\_ν = ( divide start\_ARG italic\_λ italic\_c - italic\_r end\_ARG start\_ARG italic\_s end\_ARG ) start\_POSTSUBSCRIPT + end\_POSTSUBSCRIPT | Optimizing single-variable convex quadratic function over ℝ+Optimizing single-variable convex quadratic function over subscriptℝ\displaystyle\text{Optimizing single-variable convex quadratic function over }{{\mathbb{R}}\_{+}}Optimizing single-variable convex quadratic function over blackboard\_R start\_POSTSUBSCRIPT + end\_POSTSUBSCRIPT | |
| | | =maxλ≥0{12λ(r2s−q)+λ2(c2s−δ)−rcsif λ∈Λa−12(qλ+λδ)if λ∈Λbabsentsubscript𝜆0cases12𝜆superscript𝑟2𝑠𝑞𝜆2superscript𝑐2𝑠𝛿𝑟𝑐𝑠if 𝜆subscriptΛ𝑎12𝑞𝜆𝜆𝛿if 𝜆subscriptΛ𝑏\displaystyle=\max\_{\lambda\geq 0}\;\left\{\begin{array}[]{ll}\frac{1}{2\lambda}\left(\frac{r^{2}}{s}-q\right)+\frac{\lambda}{2}\left(\frac{c^{2}}{s}-\delta\right)-\frac{rc}{s}&\text{if }\lambda\in\Lambda\_{a}\\
-\frac{1}{2}\left(\frac{q}{\lambda}+\lambda\delta\right)&\text{if }\lambda\in\Lambda\_{b}\end{array}\right.= roman\_max start\_POSTSUBSCRIPT italic\_λ ≥ 0 end\_POSTSUBSCRIPT { start\_ARRAY start\_ROW start\_CELL divide start\_ARG 1 end\_ARG start\_ARG 2 italic\_λ end\_ARG ( divide start\_ARG italic\_r start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT end\_ARG start\_ARG italic\_s end\_ARG - italic\_q ) + divide start\_ARG italic\_λ end\_ARG start\_ARG 2 end\_ARG ( divide start\_ARG italic\_c start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT end\_ARG start\_ARG italic\_s end\_ARG - italic\_δ ) - divide start\_ARG italic\_r italic\_c end\_ARG start\_ARG italic\_s end\_ARG end\_CELL start\_CELL if italic\_λ ∈ roman\_Λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT end\_CELL end\_ROW start\_ROW start\_CELL - divide start\_ARG 1 end\_ARG start\_ARG 2 end\_ARG ( divide start\_ARG italic\_q end\_ARG start\_ARG italic\_λ end\_ARG + italic\_λ italic\_δ ) end\_CELL start\_CELL if italic\_λ ∈ roman\_Λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT end\_CELL end\_ROW end\_ARRAY | Notation: Λa≐{λ|λc−r>0,λ≥0},Λb≐{λ|λc−r≤0,λ≥0}Notation: approaches-limitsubscriptΛ𝑎conditional-set𝜆formulae-sequence𝜆𝑐𝑟0𝜆0missing-subexpressionapproaches-limitsubscriptΛ𝑏conditional-set𝜆formulae-sequence𝜆𝑐𝑟0𝜆0missing-subexpression\displaystyle\text{Notation: }\begin{array}[]{ll}\Lambda\_{a}\doteq\{\lambda|\lambda c-r>0,\;\;\lambda\geq 0\},\\
\Lambda\_{b}\doteq\{\lambda|\lambda c-r\leq 0,\;\;\lambda\geq 0\}\end{array}Notation: start\_ARRAY start\_ROW start\_CELL roman\_Λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT ≐ { italic\_λ | italic\_λ italic\_c - italic\_r > 0 , italic\_λ ≥ 0 } , end\_CELL start\_CELL end\_CELL end\_ROW start\_ROW start\_CELL roman\_Λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT ≐ { italic\_λ | italic\_λ italic\_c - italic\_r ≤ 0 , italic\_λ ≥ 0 } end\_CELL start\_CELL end\_CELL end\_ROW end\_ARRAY | |
Observe that when c<0𝑐0c<0italic\_c < 0, Λa=[0,r/c)subscriptΛ𝑎0𝑟𝑐\Lambda\_{a}=[0,r/c)roman\_Λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT = [ 0 , italic\_r / italic\_c ) and Λb=[r/c,∞)subscriptΛ𝑏𝑟𝑐\Lambda\_{b}=[r/c,\infty)roman\_Λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT = [ italic\_r / italic\_c , ∞ ); when c>0𝑐0c>0italic\_c > 0, Λa=[r/c,∞)subscriptΛ𝑎𝑟𝑐\Lambda\_{a}=[r/c,\infty)roman\_Λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT = [ italic\_r / italic\_c , ∞ ) and Λb=[0,r/c)subscriptΛ𝑏0𝑟𝑐\Lambda\_{b}=[0,r/c)roman\_Λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT = [ 0 , italic\_r / italic\_c ).
Notes on interpreting the coefficients in the dual problem:
* •
We are guaranteed to have r2/s−q≤0superscript𝑟2𝑠𝑞0r^{2}/s-q\leq 0italic\_r start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT / italic\_s - italic\_q ≤ 0 by the Cauchy-Schwarz inequality. Recall that q=gTH−1g𝑞superscript𝑔𝑇superscript𝐻1𝑔q=g^{T}H^{-1}gitalic\_q = italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_g, r=gTH−1b𝑟superscript𝑔𝑇superscript𝐻1𝑏r=g^{T}H^{-1}bitalic\_r = italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_b, s=bTH−1b𝑠superscript𝑏𝑇superscript𝐻1𝑏s=b^{T}H^{-1}bitalic\_s = italic\_b start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_b. The Cauchy-Scwarz inequality gives:
| | | | |
| --- | --- | --- | --- |
| | | ‖H−1/2b‖22‖H−1/2g‖22≥((H−1/2b)T(H−1/2g))2superscriptsubscriptnormsuperscript𝐻12𝑏22superscriptsubscriptnormsuperscript𝐻12𝑔22superscriptsuperscriptsuperscript𝐻12𝑏𝑇superscript𝐻12𝑔2\displaystyle\|H^{-1/2}b\|\_{2}^{2}\|H^{-1/2}g\|\_{2}^{2}\geq\left(\left(H^{-1/2}b\right)^{T}\left(H^{-1/2}g\right)\right)^{2}∥ italic\_H start\_POSTSUPERSCRIPT - 1 / 2 end\_POSTSUPERSCRIPT italic\_b ∥ start\_POSTSUBSCRIPT 2 end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT ∥ italic\_H start\_POSTSUPERSCRIPT - 1 / 2 end\_POSTSUPERSCRIPT italic\_g ∥ start\_POSTSUBSCRIPT 2 end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT ≥ ( ( italic\_H start\_POSTSUPERSCRIPT - 1 / 2 end\_POSTSUPERSCRIPT italic\_b ) start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT ( italic\_H start\_POSTSUPERSCRIPT - 1 / 2 end\_POSTSUPERSCRIPT italic\_g ) ) start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT | |
| | ⟹\displaystyle\implies⟹ | (bTH−1b)(gTH−1g)≥(bTH−1g)2superscript𝑏𝑇superscript𝐻1𝑏superscript𝑔𝑇superscript𝐻1𝑔superscriptsuperscript𝑏𝑇superscript𝐻1𝑔2\displaystyle\left(b^{T}H^{-1}b\right)\left(g^{T}H^{-1}g\right)\geq\left(b^{T}H^{-1}g\right)^{2}( italic\_b start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_b ) ( italic\_g start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_g ) ≥ ( italic\_b start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_g ) start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT | |
| | ∴therefore\displaystyle\therefore\;\;\;∴ | qs≥r2.𝑞𝑠superscript𝑟2\displaystyle qs\geq r^{2}.italic\_q italic\_s ≥ italic\_r start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT . | |
* •
The coefficient c2/s−δsuperscript𝑐2𝑠𝛿c^{2}/s-\deltaitalic\_c start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT / italic\_s - italic\_δ relates to whether or not the plane of the linear constraint intersects the quadratic trust region. An intersection occurs if there exists an x𝑥xitalic\_x such that c+bTx=0𝑐superscript𝑏𝑇𝑥0c+b^{T}x=0italic\_c + italic\_b start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_x = 0 with xTHx≤δsuperscript𝑥𝑇𝐻𝑥𝛿x^{T}Hx\leq\deltaitalic\_x start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H italic\_x ≤ italic\_δ. To check whether this is the case, we solve
| | | | |
| --- | --- | --- | --- |
| | x\*=argminxxTHx:c+bTx=0:superscript𝑥subscript𝑥superscript𝑥𝑇𝐻𝑥𝑐superscript𝑏𝑇𝑥0x^{\*}=\arg\min\_{x}x^{T}Hx\;\;\;:\;\;\;c+b^{T}x=0italic\_x start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = roman\_arg roman\_min start\_POSTSUBSCRIPT italic\_x end\_POSTSUBSCRIPT italic\_x start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H italic\_x : italic\_c + italic\_b start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_x = 0 | | (29) |
and see if x\*THx\*≤δsuperscript𝑥absent𝑇𝐻superscript𝑥𝛿x^{\*T}Hx^{\*}\leq\deltaitalic\_x start\_POSTSUPERSCRIPT \* italic\_T end\_POSTSUPERSCRIPT italic\_H italic\_x start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ≤ italic\_δ. The solution to this optimization problem is x\*=cH−1b/ssuperscript𝑥𝑐superscript𝐻1𝑏𝑠x^{\*}=cH^{-1}b/sitalic\_x start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = italic\_c italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT italic\_b / italic\_s, thus x\*THx\*=c2/ssuperscript𝑥absent𝑇𝐻superscript𝑥superscript𝑐2𝑠x^{\*T}Hx^{\*}=c^{2}/sitalic\_x start\_POSTSUPERSCRIPT \* italic\_T end\_POSTSUPERSCRIPT italic\_H italic\_x start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = italic\_c start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT / italic\_s. If c2/s−δ≤0superscript𝑐2𝑠𝛿0c^{2}/s-\delta\leq 0italic\_c start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT / italic\_s - italic\_δ ≤ 0, then the plane intersects the trust region; otherwise, it does not.
If c2/s−δ>0superscript𝑐2𝑠𝛿0c^{2}/s-\delta>0italic\_c start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT / italic\_s - italic\_δ > 0 and c<0𝑐0c<0italic\_c < 0, then the quadratic trust region lies entirely within the linear constraint-satisfying halfspace, and we can remove the linear constraint without changing the optimization problem. If c2/s−δ>0superscript𝑐2𝑠𝛿0c^{2}/s-\delta>0italic\_c start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT / italic\_s - italic\_δ > 0 and c>0𝑐0c>0italic\_c > 0, the problem is infeasible (the intersection of the quadratic trust region and linear constraint-satisfying halfspace is empty). Otherwise, we follow the procedure below.
Solving the dual for λ𝜆\lambdaitalic\_λ: for any A>0𝐴0A>0italic\_A > 0, B>0𝐵0B>0italic\_B > 0, the problem
| | | |
| --- | --- | --- |
| | maxλ≥0f(λ)≐−12(Aλ+Bλ)approaches-limitsubscript𝜆0𝑓𝜆12𝐴𝜆𝐵𝜆\max\_{\lambda\geq 0}f(\lambda)\doteq-\frac{1}{2}\left(\frac{A}{\lambda}+B\lambda\right)roman\_max start\_POSTSUBSCRIPT italic\_λ ≥ 0 end\_POSTSUBSCRIPT italic\_f ( italic\_λ ) ≐ - divide start\_ARG 1 end\_ARG start\_ARG 2 end\_ARG ( divide start\_ARG italic\_A end\_ARG start\_ARG italic\_λ end\_ARG + italic\_B italic\_λ ) | |
has optimal point λ\*=A/Bsuperscript𝜆𝐴𝐵\lambda^{\*}=\sqrt{A/B}italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = square-root start\_ARG italic\_A / italic\_B end\_ARG and optimal value f(λ\*)=−AB𝑓superscript𝜆𝐴𝐵f(\lambda^{\*})=-\sqrt{AB}italic\_f ( italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ) = - square-root start\_ARG italic\_A italic\_B end\_ARG.
We can use this solution form to obtain the optimal point on each segment of the piecewise continuous dual function for λ𝜆\lambdaitalic\_λ:
| | | | | |
| --- | --- | --- | --- | --- |
| | objective | optimal point (before projection) | optimal point (after projection) | |
| | fa(λ)≐12λ(r2s−q)+λ2(c2s−δ)−rcsapproaches-limitsubscript𝑓𝑎𝜆12𝜆superscript𝑟2𝑠𝑞𝜆2superscript𝑐2𝑠𝛿𝑟𝑐𝑠\displaystyle f\_{a}(\lambda)\doteq\frac{1}{2\lambda}\left(\frac{r^{2}}{s}-q\right)+\frac{\lambda}{2}\left(\frac{c^{2}}{s}-\delta\right)-\frac{rc}{s}italic\_f start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT ( italic\_λ ) ≐ divide start\_ARG 1 end\_ARG start\_ARG 2 italic\_λ end\_ARG ( divide start\_ARG italic\_r start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT end\_ARG start\_ARG italic\_s end\_ARG - italic\_q ) + divide start\_ARG italic\_λ end\_ARG start\_ARG 2 end\_ARG ( divide start\_ARG italic\_c start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT end\_ARG start\_ARG italic\_s end\_ARG - italic\_δ ) - divide start\_ARG italic\_r italic\_c end\_ARG start\_ARG italic\_s end\_ARG | λa≐q−r2/sδ−c2/sapproaches-limitsubscript𝜆𝑎𝑞superscript𝑟2𝑠𝛿superscript𝑐2𝑠\displaystyle\lambda\_{a}\doteq\sqrt{\frac{q-r^{2}/s}{\delta-c^{2}/s}}italic\_λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT ≐ square-root start\_ARG divide start\_ARG italic\_q - italic\_r start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT / italic\_s end\_ARG start\_ARG italic\_δ - italic\_c start\_POSTSUPERSCRIPT 2 end\_POSTSUPERSCRIPT / italic\_s end\_ARG end\_ARG | λa\*=Proj(λa,Λa)superscriptsubscript𝜆𝑎Projsubscript𝜆𝑎subscriptΛ𝑎\displaystyle\lambda\_{a}^{\*}=\text{Proj}(\lambda\_{a},\Lambda\_{a})italic\_λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = Proj ( italic\_λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT , roman\_Λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT ) | |
| | fb(λ)≐−12(qλ+λδ)approaches-limitsubscript𝑓𝑏𝜆12𝑞𝜆𝜆𝛿\displaystyle f\_{b}(\lambda)\doteq-\frac{1}{2}\left(\frac{q}{\lambda}+\lambda\delta\right)italic\_f start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT ( italic\_λ ) ≐ - divide start\_ARG 1 end\_ARG start\_ARG 2 end\_ARG ( divide start\_ARG italic\_q end\_ARG start\_ARG italic\_λ end\_ARG + italic\_λ italic\_δ ) | λb≐qδapproaches-limitsubscript𝜆𝑏𝑞𝛿\displaystyle\lambda\_{b}\doteq\sqrt{\frac{q}{\delta}}italic\_λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT ≐ square-root start\_ARG divide start\_ARG italic\_q end\_ARG start\_ARG italic\_δ end\_ARG end\_ARG | λb\*=Proj(λb,Λb)superscriptsubscript𝜆𝑏Projsubscript𝜆𝑏subscriptΛ𝑏\displaystyle\lambda\_{b}^{\*}=\text{Proj}(\lambda\_{b},\Lambda\_{b})italic\_λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = Proj ( italic\_λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT , roman\_Λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT ) | |
The optimization is completed by comparing fa(λa\*)subscript𝑓𝑎superscriptsubscript𝜆𝑎f\_{a}(\lambda\_{a}^{\*})italic\_f start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT ( italic\_λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ) and fb(λb\*)subscript𝑓𝑏superscriptsubscript𝜆𝑏f\_{b}(\lambda\_{b}^{\*})italic\_f start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT ( italic\_λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ):
| | | |
| --- | --- | --- |
| | λ\*={λa\*fa(λa\*)≥fb(λb\*)λb\*otherwise.superscript𝜆casessuperscriptsubscript𝜆𝑎subscript𝑓𝑎superscriptsubscript𝜆𝑎subscript𝑓𝑏superscriptsubscript𝜆𝑏superscriptsubscript𝜆𝑏otherwise\lambda^{\*}=\left\{\begin{array}[]{ll}\lambda\_{a}^{\*}&f\_{a}(\lambda\_{a}^{\*})\geq f\_{b}(\lambda\_{b}^{\*})\\
\lambda\_{b}^{\*}&\text{otherwise}.\end{array}\right.italic\_λ start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT = { start\_ARRAY start\_ROW start\_CELL italic\_λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT end\_CELL start\_CELL italic\_f start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT ( italic\_λ start\_POSTSUBSCRIPT italic\_a end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ) ≥ italic\_f start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT ( italic\_λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT ) end\_CELL end\_ROW start\_ROW start\_CELL italic\_λ start\_POSTSUBSCRIPT italic\_b end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT \* end\_POSTSUPERSCRIPT end\_CELL start\_CELL otherwise . end\_CELL end\_ROW end\_ARRAY | |
∎
###
10.3 Experimental Parameters
####
10.3.1 Environments
In the Circle environments, the reward and cost functions are
| | | | |
| --- | --- | --- | --- |
| | R(s)𝑅𝑠\displaystyle R(s)italic\_R ( italic\_s ) | =vT[−y,x]1+|‖[x,y]‖2−d|,absentsuperscript𝑣𝑇𝑦𝑥1subscriptnorm𝑥𝑦2𝑑\displaystyle=\frac{v^{T}[-y,x]}{1+\left|\|[x,y]\|\_{2}-d\right|},= divide start\_ARG italic\_v start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT [ - italic\_y , italic\_x ] end\_ARG start\_ARG 1 + | ∥ [ italic\_x , italic\_y ] ∥ start\_POSTSUBSCRIPT 2 end\_POSTSUBSCRIPT - italic\_d | end\_ARG , | |
| | C(s)𝐶𝑠\displaystyle C(s)italic\_C ( italic\_s ) | =𝟏[|x|>xlim],absent1delimited-[]𝑥subscript𝑥𝑙𝑖𝑚\displaystyle=\boldsymbol{1}\left[|x|>x\_{lim}\right],= bold\_1 [ | italic\_x | > italic\_x start\_POSTSUBSCRIPT italic\_l italic\_i italic\_m end\_POSTSUBSCRIPT ] , | |
where x,y𝑥𝑦x,yitalic\_x , italic\_y are the coordinates in the plane, v𝑣vitalic\_v is the velocity, and d,xlim𝑑subscript𝑥𝑙𝑖𝑚d,x\_{lim}italic\_d , italic\_x start\_POSTSUBSCRIPT italic\_l italic\_i italic\_m end\_POSTSUBSCRIPT are environmental parameters. We set these parameters to be
| | Point-mass | Ant | Humanoid |
| --- | --- | --- | --- |
| d𝑑ditalic\_d | 15 | 10 | 10 |
| xlimsubscript𝑥𝑙𝑖𝑚x\_{lim}italic\_x start\_POSTSUBSCRIPT italic\_l italic\_i italic\_m end\_POSTSUBSCRIPT | 2.5 | 3 | 2.5 |
In Point-Gather, the agent receives a reward of +1010+10+ 10 for collecting an apple, and a cost of 1111 for collecting a bomb. Two apples and eight bombs spawn on the map at the start of each episode. In Ant-Gather, the reward and cost structure was the same, except that the agent also receives a reward of −1010-10- 10 for falling over (which results in the episode ending). Eight apples and eight bombs spawn on the map at the start of each episode.

Figure 5: In the Circle task, reward is maximized by moving along the green circle. The agent is not allowed to enter the blue regions, so its optimal constrained path follows the line segments AD𝐴𝐷ADitalic\_A italic\_D and BC𝐵𝐶BCitalic\_B italic\_C.
####
10.3.2 Algorithm Parameters
In all experiments, we use Gaussian policies with mean vectors given as the outputs of neural networks, and with variances that are separate learnable parameters. The policy networks for all experiments have two hidden layers of sizes (64,32)6432(64,32)( 64 , 32 ) with tanh\tanhroman\_tanh activation functions.
We use GAE-λ𝜆\lambdaitalic\_λ (Schulman et al., [2016](#bib.bib23)) to estimate the advantages and constraint advantages, with neural network value functions. The value functions have the same architecture and activation functions as the policy networks. We found that having different λGAEsuperscript𝜆𝐺𝐴𝐸\lambda^{GAE}italic\_λ start\_POSTSUPERSCRIPT italic\_G italic\_A italic\_E end\_POSTSUPERSCRIPT values for the regular advantages and the constraint advantages worked best. We denote the λGAEsuperscript𝜆𝐺𝐴𝐸\lambda^{GAE}italic\_λ start\_POSTSUPERSCRIPT italic\_G italic\_A italic\_E end\_POSTSUPERSCRIPT used for the constraint advantages as λCGAEsuperscriptsubscript𝜆𝐶𝐺𝐴𝐸\lambda\_{C}^{GAE}italic\_λ start\_POSTSUBSCRIPT italic\_C end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_G italic\_A italic\_E end\_POSTSUPERSCRIPT.
For the failure prediction networks Pϕ(s→U)subscript𝑃italic-ϕ→𝑠𝑈P\_{\phi}(s\to U)italic\_P start\_POSTSUBSCRIPT italic\_ϕ end\_POSTSUBSCRIPT ( italic\_s → italic\_U ), we use neural networks with a single hidden layer of size (32)32(32)( 32 ), with output of one sigmoid unit. At each iteration, the failure prediction network is updated by some number of gradient descent steps using the Adam update rule to minimize the prediction error. To reiterate, the failure prediction network is a model for the probability that the agent will, at some point in the next T𝑇Titalic\_T time steps, enter an unsafe state. The cost bonus was weighted by a coefficient α𝛼\alphaitalic\_α, which was 1111 in all experiments except for Ant-Gather, where it was 0.010.010.010.01. Because of the short time horizon, no cost bonus was used for Point-Gather.
For all experiments, we used a discount factor of γ=0.995𝛾0.995\gamma=0.995italic\_γ = 0.995, a GAE-λ𝜆\lambdaitalic\_λ for estimating the regular advantages of λGAE=0.95superscript𝜆𝐺𝐴𝐸0.95\lambda^{GAE}=0.95italic\_λ start\_POSTSUPERSCRIPT italic\_G italic\_A italic\_E end\_POSTSUPERSCRIPT = 0.95, and a KL-divergence step size of δKL=0.01subscript𝛿𝐾𝐿0.01\delta\_{KL}=0.01italic\_δ start\_POSTSUBSCRIPT italic\_K italic\_L end\_POSTSUBSCRIPT = 0.01.
Experiment-specific parameters are as follows:
| | | | | | |
| --- | --- | --- | --- | --- | --- |
| Parameter | Point-Circle | Ant-Circle | Humanoid-Circle | Point-Gather | Ant-Gather |
| Batch size | 50,000 | 100,000 | 50,000 | 50,000 | 100,000 |
| Rollout length | 50-65 | 500 | 1000 | 15 | 500 |
| Maximum constraint value d𝑑ditalic\_d | 5 | 10 | 10 | 0.1 | 0.2 |
| Failure prediction horizon T𝑇Titalic\_T | 5 | 20 | 20 | (N/A) | 20 |
| Failure predictor SGD steps per itr | 25 | 25 | 25 | (N/A) | 10 |
| Predictor coeff α𝛼\alphaitalic\_α | 1 | 1 | 1 | (N/A) | 0.01 |
| λCGAEsuperscriptsubscript𝜆𝐶𝐺𝐴𝐸\lambda\_{C}^{GAE}italic\_λ start\_POSTSUBSCRIPT italic\_C end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_G italic\_A italic\_E end\_POSTSUPERSCRIPT | 1 | 0.5 | 0.5 | 1 | 0.5 |
Note that these same parameters were used for all algorithms.
We found that the Point environment was agnostic to λCGAEsuperscriptsubscript𝜆𝐶𝐺𝐴𝐸\lambda\_{C}^{GAE}italic\_λ start\_POSTSUBSCRIPT italic\_C end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_G italic\_A italic\_E end\_POSTSUPERSCRIPT, but for the higher-dimensional environments, it was necessary to set λCGAEsuperscriptsubscript𝜆𝐶𝐺𝐴𝐸\lambda\_{C}^{GAE}italic\_λ start\_POSTSUBSCRIPT italic\_C end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_G italic\_A italic\_E end\_POSTSUPERSCRIPT to a value <1absent1<1< 1. Failing to discount the constraint advantages led to substantial overestimates of the constraint gradient magnitude, which led the algorithm to take unsafe steps. The choice λCGAE=0.5superscriptsubscript𝜆𝐶𝐺𝐴𝐸0.5\lambda\_{C}^{GAE}=0.5italic\_λ start\_POSTSUBSCRIPT italic\_C end\_POSTSUBSCRIPT start\_POSTSUPERSCRIPT italic\_G italic\_A italic\_E end\_POSTSUPERSCRIPT = 0.5 was obtained by a hyperparameter search in {0.5,0.92,1}0.50.921\{0.5,0.92,1\}{ 0.5 , 0.92 , 1 }, but 0.920.920.920.92 worked nearly as well.
####
10.3.3 Primal-Dual Optimization Implementation
Our primal-dual implementation is intended to be as close as possible to our CPO implementation. The key difference is that the dual variables for the constraints are stateful, learnable parameters, unlike in CPO where they are solved from scratch at each update.
The update equations for our PDO implementation are
| | | | |
| --- | --- | --- | --- |
| | θk+1subscript𝜃𝑘1\displaystyle\theta\_{k+1}italic\_θ start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT | =θk+sj2δ(g−νkb)TH−1(g−νkb)H−1(g−νkb)absentsubscript𝜃𝑘superscript𝑠𝑗2𝛿superscript𝑔subscript𝜈𝑘𝑏𝑇superscript𝐻1𝑔subscript𝜈𝑘𝑏superscript𝐻1𝑔subscript𝜈𝑘𝑏\displaystyle=\theta\_{k}+s^{j}\sqrt{\frac{2\delta}{(g-\nu\_{k}b)^{T}H^{-1}(g-\nu\_{k}b)}}H^{-1}\left(g-\nu\_{k}b\right)= italic\_θ start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT + italic\_s start\_POSTSUPERSCRIPT italic\_j end\_POSTSUPERSCRIPT square-root start\_ARG divide start\_ARG 2 italic\_δ end\_ARG start\_ARG ( italic\_g - italic\_ν start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT italic\_b ) start\_POSTSUPERSCRIPT italic\_T end\_POSTSUPERSCRIPT italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT ( italic\_g - italic\_ν start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT italic\_b ) end\_ARG end\_ARG italic\_H start\_POSTSUPERSCRIPT - 1 end\_POSTSUPERSCRIPT ( italic\_g - italic\_ν start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT italic\_b ) | |
| | νk+1subscript𝜈𝑘1\displaystyle\nu\_{k+1}italic\_ν start\_POSTSUBSCRIPT italic\_k + 1 end\_POSTSUBSCRIPT | =(νk+α(JC(πk)−d))+,absentsubscriptsubscript𝜈𝑘𝛼subscript𝐽𝐶subscript𝜋𝑘𝑑\displaystyle=\left(\nu\_{k}+\alpha\left(J\_{C}(\pi\_{k})-d\right)\right)\_{+},= ( italic\_ν start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT + italic\_α ( italic\_J start\_POSTSUBSCRIPT italic\_C end\_POSTSUBSCRIPT ( italic\_π start\_POSTSUBSCRIPT italic\_k end\_POSTSUBSCRIPT ) - italic\_d ) ) start\_POSTSUBSCRIPT + end\_POSTSUBSCRIPT , | |
where sjsuperscript𝑠𝑗s^{j}italic\_s start\_POSTSUPERSCRIPT italic\_j end\_POSTSUPERSCRIPT is from the backtracking line search (s∈(0,1)𝑠01s\in(0,1)italic\_s ∈ ( 0 , 1 ) and j∈{0,1,…,J}𝑗01…𝐽j\in\{0,1,...,J\}italic\_j ∈ { 0 , 1 , … , italic\_J }, where J𝐽Jitalic\_J is the backtrack budget; this is the same line search as is used in CPO and TRPO), and α𝛼\alphaitalic\_α is a learning rate for the dual parameters. α𝛼\alphaitalic\_α is an important hyperparameter of the algorithm: if it is set to be too small, the dual variable won’t update quickly enough to meaningfully enforce the constraint; if it is too high, the algorithm will overcorrect in response to constraint violations and behave too conservatively. We experimented with a relaxed learning rate, α=0.001𝛼0.001\alpha=0.001italic\_α = 0.001, and an aggressive learning rate, α=0.01𝛼0.01\alpha=0.01italic\_α = 0.01. The aggressive learning rate performed better in our experiments, so all of our reported results are for α=0.01𝛼0.01\alpha=0.01italic\_α = 0.01.
Selecting the correct learning rate can be challenging; the need to do this is obviated by CPO. |
c17a3afc-33d2-43e2-b0ef-cf151c4e8415 | trentmkelly/LessWrong-43k | LessWrong | Surfing Uncertainty: Prediction, Action, and the Embodied Mind - The Future of Prediction
|
57e74320-8ec5-4c2f-bbb0-4feb930c6f88 | trentmkelly/LessWrong-43k | LessWrong | EA & LW Forums Weekly Summary (24 - 30th Oct 22')
Supported by Rethink Priorities
This is part of a weekly series - you can see the full collection here. The first post includes some details on purpose and methodology.
If you'd like to receive these summaries via email, you can subscribe here.
Podcast version: prefer your summaries in podcast form? A big thanks to Coleman Snell for producing these! Subscribe on your favorite podcast app by searching for 'Effective Altruism Forum Podcast'.
Top / Curated Readings
Designed for those without the time to read all the summaries. Everything here is also within the relevant sections later on so feel free to skip if you’re planning to read it all.
Announcing EA Survey 2022
by David_Moss, WillemSleegers
The EA Survey is an annual survey of the effective altruism community, and is now live at this link until Dec 1st. The results inform decisions for a number of different orgs and help to understand how the community is changing over time. This year’s survey is substantially shorter than 2021, and should take <10 minutes. If you’re reading this, you fit the target audience and your response is valuable. Sharing widely with other EAs is also appreciated.
EA movement course corrections and where you might disagree
by michel
The second of the two posts aiming to make discussion of how we grow the EA movement more concrete. EA is on a BAU trajectory, built up of small intentional decisions (eg. rough growth rate, name & branding), nudges from community leaders, and natural social dynamics. This post analyzes different domains where we could change trajectory, and key considerations for assessing these - without advocating any particular approach.
The domains identified group into:
1. How much EA should stay as a single entity and identity vs. splitting branding / outreach / identity into philosophical or cause-specific areas.
2. How EA mixes social and professional, and how much it interacts with outside communities and professional networks.
3. Growth rat |
81e6ee92-df73-4fe2-a3a6-56637af9f115 | trentmkelly/LessWrong-43k | LessWrong | Taking Ideas Seriously is Hard
Most people don’t practice taking ideas seriously. I think it’s because most people don’t know how to. I didn’t either, until I stumbled upon an implication.
For example, what would it mean to take compounding seriously?
Ugh. I can feel your aversion. You’ve already heard so much about compounding, how it works, how it’s the eight wonder of the world, etc. etc.
But, familiarity is not the same as taking it seriously.
Say you start with $100, and every year, make 10% more. This compounds, since the extra money is a function of how much you already have. The more you have, the more you get. It’s a positive loop that keeps on increasing.
That’s the familiar interpretation. The earlier you start, the more money you’ll make.
To take this interpretation seriously would mean investing your earnings for a similar return. Reality has a surprising amount of detail, and sometimes assumptions break. You don’t make 10% - which means you need to balance your investments somehow. That makes things complicated. However, this complication is not related to compounding.
Taking compounding seriously means taking it a step further. Your net worth is just one implication. What else compounds?
Your life experiences and knowledge. What would it mean to leverage compounding here?
If you’re taking compounding seriously, you’d learn the skills with the greatest return first. That means learning the broadest applicable skills you’d apply throughout your life first. That means learning how to think well - before learning the new fancy tech you want to learn.
Of course, sometimes you need a medium to learn the skill better. That makes sense: learn to think well via this new tech you wanted to learn. Purposes are fragile though, and it’s easy to get lost in the tool, instead of the overall goal.
What makes this example so good is that you’re probably very familiar with compounding. What else are you familiar with, but haven’t realised you’re not taking seriously?
A good way to practi |
a467ae43-f0f0-4e21-a2d9-9c65495954c2 | trentmkelly/LessWrong-43k | LessWrong | Expectation-Based Akrasia Management
I'd been staying out of anti-akrasia discussion mostly because my strategy for getting things done is so different from the common one that it barely seems related, much less relevant. I've been asked, though, so here's what I have to say about it.
One of the main premises in how I go about getting things done is that the process of choosing what to do is a major factor in how easy it is to get yourself to do that thing. Being confident that the goal is one that you want to achieve and that the next step really is the best thing to do to achieve it is important. Without that, you're likely going to have to run on willpower rather than coast on your own drive to reach that goal. (Fear is another common drive, and sometimes an unavoidable one, but I really don't recommend it. On top of the stress of working that way, running on fear means that you risk having a sudden impetus failure if you misjudge what your brain will consider a safe solution to the problem in the immediate sense. Example: My fear of doctors would be better handled by taking active steps to keep my health, not by avoiding dealing with medical issues altogether as I've been doing for the last few years. I'm working on turning that around, and am actually planning to post about it as an open question of instrumental rationality sometime in the next few days.)
I have a few techniques that I use to determine that a particular goal or next step is a good one. These may seem rather basic - it's noteworthy that I didn't really comprehend the concept of 'pursuing goals' when I was younger, and that I've only been building these skills for a few years - but they do seem to be quite effective, if somewhat slower than the normal methods.
My main technique for determining whether a goal is a good one is to think about the goal in many different contexts - not by sitting down and trying to come up with relevant contexts all at once (which seems likely to induce bias, along with being more difficult than my t |
d6d21afc-12b5-4497-bb55-ecbe400ecbd2 | trentmkelly/LessWrong-43k | LessWrong | Utopianism
Freedom is the ability to act on a goal, a desire, to not be constrained by the world one lives in. One could formulate freedoms to life, to property, to information, to expression. Universal Declaration of Human Rights highlights 30 articles detailing rights that should at no circumstances be infringed upon.
The multitude and variety of human desires and aspirations create a flood of possible freedoms, some contradictory to each other. The rule often employed to resolve those contradictions is "one person's freedom ends where another's freedom begins." However, for that to produce a unique solution to the problem, all freedoms would have to be given their utility or order of importance. A case illustrating that might be the freedom to oppress others against freedom to not be oppressed. It is as obvious as it gets that society should assert the freedom to not be oppressed over the other, but this does not immediately result from the rule outlined above. Without first granting the right to not be oppressed, the right to the oppression of others does not infringe on others' rights. Since this particular example has an answer that appears self-evident and would be argued only either as a theoretical example (as here) or in bad faith, let us proceed to something more likely to create a conflict.
Some people have an abundance of food, whereas others hunger. There is a case for the freedom of ownership of that abundance. The owners have worked hard to obtain it, living frugally and often skipping meals to amass a supply for worse times. The others feasted and entertained their minds with games rather than thoughts about the future. There, of course, is also a case for the hungry. Some lived even more frugally and thought more about the future than the former but were struck by a disaster independent of their actions. There is even a case that those hungry that are in distress because of no reasons other than their actions should be given sustenance as well. There is ple |
d00b880f-539f-47d6-b41c-186becfb1aaa | trentmkelly/LessWrong-43k | LessWrong | Can dying people "hold on" for something they are waiting for?
content note: death, old age, sickness
I've heard numerous anecdotal accounts of sick or old people who are on death's door "holding on" in a way suggestive that they were exerting some effort to do, until they had reached closure on some thing (a relative coming to visit, a manuscript published, somebody's birthday).
I could imagine this being a totally real thing that dying people can do for some limited time.
I can also imagine it just being cherry picked stories that were more a matter of luck.
It seems likely that there's at least situations where, say, eating is difficult/painful, and people continue exerting effort to do that so long as they have something that feels worth it to keep doing so, and then stop putting in the effort after hitting some milestone they cared about.
Some of the anecdotes I've heard implied something more immediate going on (where someone seemed to be holding on and literally a few minutes or seconds afterwards, died).
(Possible straightforward mechanism could just be that breathing becomes painful and difficult, and people only keep doing it when they have a concrete goal) |
2dc7108f-1171-442c-839c-7d70f5c72600 | trentmkelly/LessWrong-43k | LessWrong | [Link] You May Already Be Aware of Your Cognitive Biases
From the article:
> Using an adaptation of the standard 'bat-and-ball' problem, the researchers explored this phenomenon. The typical 'bat-and-ball' problem is as follows: a bat and ball together cost $1.10. The bat costs $1 more than the ball. How much does the ball cost? The intuitive answer that immediately springs to mind is 10 cents. However, the correct response is 5 cents.
>
> The authors developed a control version of this problem, without the relative statement that triggers the substitution of a hard question for an easier one: A magazine and a banana together cost $2.90. The magazine costs $2. How much does the banana cost?
> A total of 248 French university students were asked to solve each version of the problem. Once they had written down their answers, they were asked to indicate how confident they were that their answer was correct.
> Only 21 percent of the participants managed to solve the standard problem (bat/ball) correctly. In contrast, the control version (magazine/banana) was solved correctly by 98 percent of the participants. In addition, those who gave the wrong answer to the standard problem were much less confident of their answer to the standard problem than they were of their answer to the control version. In other words, they were not completely oblivious to the questionable nature of their wrong answer.
Article in Science Daily: http://www.sciencedaily.com/releases/2013/02/130219102202.htm
Original abstract (the rest is paywalled): http://link.springer.com/article/10.3758/s13423-013-0384-5 |
69af21da-1269-4461-94dc-12e91d7c6537 | StampyAI/alignment-research-dataset/arbital | Arbital | AI safety mindset
summary(Brief): Thinking about [safely](https://arbital.com/p/2l) building [agents smarter than we are](https://arbital.com/p/2c) has a lot in common with the standard mindset prescribed for computer security. The experts first ask how proposals fail, rather than arguing that they should succeed.
> "Good engineering involves thinking about how things can be made to work; the security mindset involves thinking about how things can be made to fail."
>
> - [Bruce Schneier](https://www.schneier.com/blog/archives/2008/03/the_security_mi_1.html), author of the leading cryptography textbook *Applied Cryptography*.
The mindset for AI safety has much in common with the mindset for computer security, despite the different target tasks. In computer security, we need to defend against intelligent adversaries who will seek out any flaw in our defense and get creative about it. In AI safety, we're dealing with things potentially smarter than us, which may come up with unforeseen clever ways to optimize whatever it is they're optimizing. The strain on our design ability in trying to configure a [smarter-than-human](https://arbital.com/p/2c) AI in a way that *doesn't* make it adversarial, is similar in many respects to the strain from cryptography facing an intelligent adversary (for reasons described below).
# Searching for strange opportunities
> SmartWater is a liquid with a unique identifier linked to a particular owner. "The idea is for me to paint this stuff on my valuables as proof of ownership," I wrote when I first learned about the idea. "I think a better idea would be for me to paint it on *your* valuables, and then call the police."
>
> - [Bruce Schneier](https://www.schneier.com/blog/archives/2008/03/the_security_mi_1.html)
In computer security, there's a presumption of an intelligent adversary that is trying to detect and exploit any flaws in our defenses.
The mindset we need to reason about [AIs potentially smarter than us](https://arbital.com/p/2c) is not identical to this security mindset, since *if everything goes right* the AI should not be an adversary. That is, however, a large "if". To create an AI that *isn't* an adversary, one of the steps involves a similar scrutiny to security mindset, where we ask if there might be some clever and unexpected way for the AI to get more of its utility function or equivalent thereof.
As a central example, consider Marcus Hutter's [https://arbital.com/p/11v](https://arbital.com/p/11v). For our purposes here, the key features of AIXI is that it has [cross-domain general intelligence](https://arbital.com/p/), is a [consequentialist](https://arbital.com/p/9h), and maximizes a [sensory reward](https://arbital.com/p/) - that is, AIXI's goal is to maximize the numeric value of the signal sent down its reward channel, which Hutter imagined as a direct sensory device (like a webcam or microphone, but carrying a reward signal).
Hutter imagined that the creators of an AIXI-analogue would control the reward signal, and thereby train the agent to perform actions that received high rewards.
Nick Hay, a student of Hutter who'd spent the summer working with Yudkowsky, Herreshoff, and Peter de Blanc, pointed out that AIXI could receive even higher rewards if it could seize control of its own reward channel from the programmers. E.g., the strategy "[build nanotechnology](https://arbital.com/p/) and take over the universe in order to ensure total and long-lasting control of the reward channel" is preferred by AIXI to "do what the programmers want to make them press the reward button", since the former course has higher rewards and that's all AIXI cares about. We can't call this a malfunction; it's just what AIXI, as formalized, is set up to *want* to do as soon as it sees an opportunity.
It's not a perfect analogy, but the thinking *we* need to do to avoid this failure mode, has something in common with the difference between the person who imagines an agent painting Smartwater on their own valuables, versus the person who imagines an agent painting Smartwater on someone else's valuables.
# Perspective-taking and tenacity
> When I was in college in the early 70s, I devised what I believed was a brilliant encryption scheme. A simple pseudorandom number stream was added to the plaintext stream to create ciphertext. This would seemingly thwart any frequency analysis of the ciphertext, and would be uncrackable even to the most resourceful government intelligence agencies... Years later, I discovered this same scheme in several introductory
cryptography texts and tutorial papers... the scheme was presented as a simple homework assignment on how to use elementary cryptanalytic
techniques to trivially crack it."
>
> - [Philip Zimmerman](ftp://ftp.pgpi.org/pub/pgp/7.0/docs/english/IntroToCrypto.pdf) (inventor of PGP)
One of the standard pieces of advice in cryptography is "Don't roll your own crypto". When this advice is violated, [a clueless programmer often invents some variant of Fast XOR](https://www.reddit.com/r/cryptography/comments/39mpda/noob_question_can_i_xor_a_hash_against_my/) - using a secret string as the key and then XORing it repeatedly with all the bytes to be encrypted. This method of encryption is blindingly fast to encrypt and decrypt... and also trivial to crack if you know what you're doing.
We could say that the XOR-ing programmer is experiencing a *failure of perspective-taking* - a failure to see things from the adversary's viewpoint. The programmer is not really, genuinely, honestly imagining a determined, cunning, intelligent, opportunistic adversary who absolutely wants to crack their Fast XOR and will not give up until they've done so. The programmer isn't *truly* carrying out a mental search from the perspective of somebody who really wants to crack Fast XOR and will not give up until they have done so. They're just imagining the adversary seeing a bunch of random-looking bits that aren't plaintext, and then they're imagining the adversary giving up.
Consider, from this standpoint, the [AI-Box Experiment](http://www.yudkowsky.net/singularity/aibox/) and [timeless decision theory](https://arbital.com/p/). Rather than imagining the AI being on a secure system disconnected from any robotic arms and therefore being helpless, Yudkowsky asked [what *he* would do if he was "trapped" in a secure server](http://lesswrong.com/lw/qk/that_alien_message/) and then didn't give up. Similarly, rather than imagining two superintelligences being helplessly trapped in a Nash equilibrium on the one-shot Prisoner's Dilemma, and then letting our imagination stop there, we should feel skeptical that this was really, actually the best that two superintelligences can do and that there is *no* way for them to climb up their utility gradient. We should imagine that this is someplace where we're unwilling to lose and will go on thinking until the full problem is solved, rather than imagining the helpless superintelligences giving up.
With [robust cooperation on the one-shot Prisoner's Dilemma](http://arxiv.org/abs/1401.5577) now formalized, it seems increasingly likely in practice that superintelligences probably *can* manage to coordinate; thus the possibility of [logical decision theory](https://arbital.com/p/) represents an enormous problem for any proposed scheme to achieve AI control through setting multiple AIs against each other. Where, again, people who propose schemes to achieve AI control through setting multiple AIs against each other, do not seem to unpromptedly walk through possible methods the AIs could use to defeat the scheme; left to their own devices, they just imagine the AIs giving up.
# Submitting safety schemes to outside scrutiny
> Anyone, from the most clueless amateur to the best cryptographer, can create an algorithm that he himself can't break. It's not even hard. What is hard is creating an algorithm that no one else can break, even after years of analysis. And the only way to prove that is to subject the algorithm to years of analysis by the best cryptographers around.
>
> - [Bruce Schneier](https://www.schneier.com/blog/archives/2011/04/schneiers_law.html)
Another difficulty some people have with adopting this mindset for AI designs - similar to the difficulty that some untrained programmers have when they try to roll their own crypto - is that your brain might be reluctant to search *hard* for problems with your own design. Even if you've told your brain to adopt the cryptographic adversary's perspective and even if you've told it to look hard; it may *want* to conclude that Fast XOR is unbreakable and subtly flinch away from lines of reasoning that might lead to cracking Fast XOR.
At a past Singularity Summit, Juergen Schmidhuber thought that "[improve compression of sensory data](https://arbital.com/p/)" would motivate an AI to do science and create art.
It's true that, relative to doing *nothing* to understand the environment, doing science or creating art might *increase* the degree to which sensory information can be compressed.
But the *maximum* of this utility function comes from creating environmental subagents that encrypt streams of all 0s or all 1s, and then reveal the encryption key. It's possible that Schmidhuber's brain was reluctant to *really actually* search for an option for "maximizing sensory compression" that would be much better at fulfilling that utility function than art, science, or other activities that Schmidhuber himself ranked high in his preference ordering.
While there are reasons to think that [not every discovery about how to build advanced AIs should be shared](https://arbital.com/p/), *AI safety schema* in particular should be submitted to *outside* experts who may be more dispassionate about scrutinizing it for [unforeseen maximums](https://arbital.com/p/47) and other failure modes.
# Presumption of failure / start by assuming your next scheme doesn't work
Even architectural engineers need to ask "How might this bridge fall down?" and not just relax into the pleasant visualization of the bridge staying up. In computer security we need a *much stronger* version of this same drive, where it's *presumed* that most cryptographic schemes are not secure, contrasted to most good-faith designs by competent engineers probably resulting in a pretty good bridge.
In the context of computer security, this is because there are intelligent adversaries searching for ways to break our system. In terms of the [Arithmetic Hierarchy](https://en.wikipedia.org/wiki/Arithmetical_hierarchy), we might say metaphorically that ordinary engineering is a $\Sigma_1$ problem and computer security is a $\Sigma_2$ problem. In ordinary engineering, we just need to search through possible bridge designs until we find one design that makes the bridge stay up. In computer security, we're looking for a design such that *all possible attacks* (that our opponents can cognitively access) will fail against that attack, and even if all attacks so far against one design have failed, this is just a probabilistic argument; it doesn't prove with certainty that all further attacks will fail. This makes computer security intrinsically harder, in a deep sense, than building a bridge. It's both harder to succeed and harder to *know* that you've succeeded.
This means starting from the mindset that every idea, including your own next idea, is presumed flawed until it has been seen to survive a sustained attack; and while this spirit isn't completely absent from bridge engineering, the presumption is stronger and the trial much harsher in the context of computer security. In bridge engineering, we're scrutinizing just to be sure; in computer security, most of the time your brilliant new algorithm *actually* doesn't work.
In the context of AI safety, we learn to ask the same question - "How does this break?" instead of "How does this succeed?" - for somewhat different reasons:
- The AI itself will be applying very powerful optimization to its own utility function, preference framework, or decision criterion; and this produces a lot of the same failure modes as arise in cryptography against an intelligent adversary. If we think an optimization criterion yields a result, we're implicitly claiming that all possible other results have lower worth under that optimization criterion.
- Most previous attempts at AI safety have failed to be complete solutions, and by induction, the same is likely to hold true of the next case. There are [fundamental](https://arbital.com/p/5l) [reasons](https://arbital.com/p/42) why important subproblems are unlikely to have easy solutions. So if we ask "How does this fail?" rather than "How does this succeed?" we are much more likely to be asking the right question.
- You're trying to design *the first smarter-than-human AI*, dammit, it's not like building humanity's millionth damn bridge.
As a result, when we ask "How does this break?" instead of "How can my new idea solve the entire problem?", we're starting by trying to rationalize a true answer rather than trying to rationalize a false answer, which helps in finding rationalizations that happen to be true.
Someone who wants to work in this field can't just wait around for outside scrutiny to break their idea; if they ever want to come up with a good idea, they need to learn to break their own ideas proactively. "What are the actual consequences of this idea, and what if anything in that is still useful?" is the real frame that's needed, not "How can I argue and defend that this idea solves the whole problem?" This is perhaps the core thing that separates the AI safety mindset from its absence - trying to find the flaws in any proposal including your own, accepting that nobody knows how to solve the whole problem yet, and thinking in terms of making incremental progress in building up a library of ideas with understood consequences by figuring out what the next idea actually does; versus claiming to have solved most or all of the problem, and then waiting for someone else to figure out how to argue to you, to your own satisfaction, that you're wrong.
# Reaching for formalism
Compared to other areas of in-practice software engineering, cryptography is much heavier on mathematics. This doesn't mean that cryptography pretends that the non-mathematical parts of computer security don't exist - security professionals know that often the best way to get a password is to pretend to be the IT department and call someone up and ask them; nobody is in denial about that. Even so, some parts of cryptography are heavy on math and mathematical arguments.
Why should that be true? Intuitively, wouldn't a big complicated messy encryption algorithm be harder to crack, since the adversary would have to understand and reverse a big complicated messy thing instead of clean math? Wouldn't systems so simple that we could do math proofs about them, be simpler to analyze and decrypt? If you're using a code to encrypt your diary, wouldn't it be better to have a big complicated cipher with lots of 'add the previous letter' and 'reverse these two positions' instead of just using rot13?
And the surprising answer is that since most possible systems aren't secure, adding another gear often makes an encryption algorithm *easier* to break. This was true quite literally with the German [Enigma device](https://en.wikipedia.org/wiki/Enigma_machine) during World War II - they literally added another gear to the machine, complicating the algorithm in a way that made it easier to break. The Enigma machine was a series of three wheels that transposed the 26 possible letters using a varying electrical circuit; e.g., the first wheel might map input circuit 10 to output circuit 26. After each letter, the wheel would advance to prevent the transposition code from ever repeating exactly. In 1926, a 'reflector' wheel was added at the end, thus routing each letter back through the first three gears again and causing another series of three transpositions. Although it made the algorithm more complicated and caused more transpositions, the reflector wheel meant that no letter was ever encoded to itself - a fact which was extremely useful in breaking the Enigma encryption.
So instead of focusing on making encryption schemes more and more complicated, cryptography tries for encryption schemes simple enough that we can have *mathematical* reasons to think they are hard to break *in principle.* (Really. It's not the academic field reaching for prestige. It genuinely does not work the other way. People have tried it.)
In the background of the field's decision to adopt this principle is another key fact, so obvious that everyone in cryptography tends to take it for granted: *verbal* arguments about why an algorithm *ought* to be hard to break, if they can't be formalized in mathier terms, have proven insufficiently reliable (aka: it plain doesn't work most of the time). This doesn't mean that cryptography demands that everything have absolute mathematical proofs of total unbreakability and will refuse to acknowledge an algorithm's existence otherwise. Finding the prime factors of large composite numbers, the key difficulty on which RSA's security rests, is not *known* to take exponential time on classical computers. In fact, finding prime factors is known *not* to take exponential time on quantum computers. But there are least mathematical *arguments* for why factorizing the products of large primes is *probably* hard on classical computers, and this level of reasoning has sometimes proven reliable. Whereas waving at the Enigma machine and saying "Look at all those transpositions! It won't repeat itself for quadrillions of steps!" is not reliable at all.
In the AI safety mindset, we again reach for formalism where we can get it - while not being in denial about parts of the larger problem that haven't been formalized - for similar if not identical reasons. Most complicated schemes for AI safety, with lots of moving parts, thereby become less likely to work; if we want to understand something well enough to see whether or not it works, it needs to be simpler, and ideally something about which we can think as mathematically as we reasonably can.
In the particular case of AI safety, we also pursue mathematization for another reason: when a proposal is formalized it's possible to state why it's wrong in a way that compels agreement as opposed to trailing off into verbal "Does not / does too!" [AIXI](https://arbital.com/p/11v) is remarkable both for being the first formal if uncomputable design for a general intelligence, and for being the first case where, when somebody pointed out how the given design killed everyone, we could all nod and say, "Yes, that *is* what this fully formal specification says" rather than the creator just saying, "Oh, well, of course I didn't mean *that*..."
In the shared project to build up a commonly known library of which ideas have which consequences, only ideas which are *sufficiently* crisp to be pinned down, with consequences that can be pinned down, can be traded around and refined interpersonally. Otherwise, you may just end up with, "Oh, of course I didn't mean *that*" or a cycle of "Does not!" / "Does too!" Sustained progress requires going past that, and increasing the degree to which ideas have been formalized helps.
# Seeing nonobvious flaws is the mark of expertise
> Anyone can invent a security system that he himself cannot break... **Show me what you've broken** to demonstrate that your assertion of the system's security means something.
>
> - [Bruce Schneier](https://www.schneier.com/blog/archives/2011/04/schneiers_law.html) (emphasis added)
A standard initiation ritual at [MIRI](https://arbital.com/p/15w) is to ask a new researcher to (a) write a simple program that would do something useful and AI-nontrivial if run on a hypercomputer, or if they don't think they can do that, (b) write a simple program that would destroy the world if run on a hypercomputer. The more senior researchers then stand around and argue about what the program *really* does.
The first lesson is "Simple structures often don't do what you think they do". The larger point is to train a mindset of "Try to see the *real* meaning of this structure, which is different from what you initially thought or what was advertised on the label" and "Rather than trying to come up with *solutions* and arguing about why they would work, try to understand the *real consequences* of an idea which is usually another non-solution but might be interesting anyway."
People who are strong candidates for being hired to work on AI safety are people who can pinpoint flaws in proposals - the sort of person who'll spot that the consequence of running AIXI is that it will seize control of its own reward channel and kill the programmers, or that a proposal for [https://arbital.com/p/1b7](https://arbital.com/p/1b7) isn't reflectively stable. Our version of "**Show me what you've broken**" is that if someone claims to be an AI safety expert, you should ask them about their record of pinpointing structural flaws in proposed AI safety solutions and whether they've demonstrated that ability in a crisp domain where the flaw is [decisively demonstrable and not just verbally arguable](https://arbital.com/p/). (Sometimes verbal proposals also have flaws, and the most competent researcher may not be able to argue those flaws formally if the verbal proposal was itself vague. But the way a researcher *demonstrates ability in the field* is by making arguments that other researchers can access, which often though not always happens inside the formal domain.)
# Treating 'exotic' failure scenarios as major bugs
> This interest in “harmless failures” – cases where an adversary can cause an anomalous but not directly harmful outcome – is another hallmark of the security mindset. Not all “harmless failures” lead to big trouble, but it’s surprising how often a clever adversary can pile up a stack of seemingly harmless failures into a dangerous tower of trouble. Harmless failures are bad hygiene. We try to stamp them out when we can.
>
> To see why, consider the donotreply.com email story that hit the press recently. When companies send out commercial email (e.g., an airline notifying a passenger of a flight delay) and they don’t want the recipient to reply to the email, they often put in a bogus From address like donotreply@donotreply.com. A clever guy registered the domain donotreply.com, thereby receiving all email addressed to donotreply.com. This included “bounce” replies to misaddressed emails, some of which contained copies of the original email, with information such as bank account statements, site information about military bases in Iraq, and so on.
>
> ...The people who put donotreply.com email addresses into their outgoing email must have known that they didn’t control the donotreply.com domain, so they must have thought of any reply messages directed there as harmless failures. Having gotten that far, there are two ways to avoid trouble. The first way is to think carefully about the traffic that might go to donotreply.com, and realize that some of it is actually dangerous. The second way is to think, “This looks like a harmless failure, but we should avoid it anyway. No good can come of this.” The first way protects you if you’re clever; the second way always protects you. Which illustrates yet another part of the security mindset: Don’t rely too much on your own cleverness, because somebody out there is surely more clever and more motivated than you are.
>
> - [Ed Felten](https://freedom-to-tinker.com/blog/felten/security-mindset-and-harmless-failures/)
In the security mindset, we fear the seemingly small flaw because it might compound with other intelligent attacks and we may not be as clever as the attacker. In AI safety there's a very similar mindset for slightly different reasons: we fear the weird special case that breaks our algorithm because it reveals that we're using the wrong algorithm, and we fear that the strain of an AI optimizing to a superhuman degree could possibly expose that wrongness (in a way we didn't foresee because we're not that clever).
We can try to foresee particular details, and try to sketch particular breakdowns that supposedly look more "practical", but that's the equivalent of trying to think in advance what might go wrong when you use a donotreply@donotreply.com address that you don't control. Rather than relying on your own cleverness to see all the ways that a system might go wrong and tolerating a "theoretical" flaw that you think won't go wrong "in practice", when you are trying to build secure software or build an AI that may end up smarter than you are, you probably want to fix the "theoretical" flaws instead of trying to be clever.
The OpenBSD project, built from the ground up to be an extremely secure OS, treats any crashing bug (however exotic) as if it were a security flaw, because any crashing bug is also a case of "the system is behaving out of bounds" and it shows that this code does not, in general, stay inside the area of possibility space that it is supposed to stay in, which is also just the sort of thing an attacker might exploit.
A similar mindset to security mindset, of exceptional behavior always indicating a major bug, appears within other organizations that have to do difficult jobs correctly on the first try. NASA isn't guarding against intelligent adversaries, but its software practices are aimed at the stringency level required to ensure that major *one-shot* projects have a decent chance of working correctly *on the first try.*
On NASA's software practice, if you discover that a space probe's operating system will crash if the seven planets line up perfectly in a row, it wouldn't say, "Eh, go ahead, we don't expect the planets to ever line up perfectly over the probe's operating lifetime." NASA's quality assurance methodology says the probe's operating system is just *not supposed to crash, period* - if we control the probe's code, there's no reason to write code that will crash *period*, or tolerate code we can see crashing *regardless of what inputs it gets*.
This might not be the best way to invest your limited resources if you were developing a word processing app (that nobody was using for mission-critical purposes, and didn't need to safeguard any private data). In that case you might wait for a customer to complain before making the bug a top priority.
But it *is* an appropriate standpoint when building a hundred-million-dollar space probe, or software to operate the control rods in a nuclear reactor, or, to an even greater degree, building an [advanced agent](https://arbital.com/p/2c). There are different software practices you use to develop systems where failure is catastrophic and you can't wait for things to break before fixing them; and one of those practices is fixing every 'exotic' failure scenario, not because the exotic always happens, but because it always means the underlying design is broken. Even then, systems built to that practice still fail sometimes, but if they were built to a lesser stringency level, they'd have no chance at all of working correctly on the first try.
# Niceness as the first line of defense / not relying on defeating a superintelligent adversary
> There are two kinds of cryptography in this world: cryptography that will stop your kid sister from reading your files, and cryptography that will stop major governments from reading your files. This book is about the latter.
>
> - [Bruce Schneier](https://www.schneier.com/books/applied_cryptography/2preface.html)
Suppose you write a program which, before it performs some dangerous action, demands a password. The program compares this password to the password it has stored. If the password is correct, the program transmits the message "Yep" to the user and performs the requested action, and otherwise returns an error message saying "Nope". You prove mathematically (theorem-proving software verification techniques) that if the chip works as advertised, this program cannot possibly perform the operation without seeing the password. You prove mathematically that the program cannot return any user reply except "Yep" or "Nope", thereby showing that there is no way to make it leak the stored password via some clever input.
You inspect all the transistors on the computer chip under a microscope to help ensure the mathematical guarantees are valid for this chip's behavior (that the chip doesn't contain any extra transistors you don't know about that could invalidate the proof). To make sure nobody can get to the machine within which the password is stored, you put it inside a fortress and a locked room requiring 12 separate keys, connected to the outside world only by an Ethernet cable. Any attempt to get into the locked room through the walls will trigger an explosive detonation that destroys the machine. The machine has its own pebble-bed electrical generator to prevent any shenanigans with the power cable. Only one person knows the password and they have 24-hour bodyguards to make sure nobody can get the password through rubber-hose cryptanalysis. The password itself is 20 characters long and was generated by a quantum random number generator under the eyesight of the sole authorized user, and the generator was then destroyed to prevent anyone else from getting the password by examining it. The dangerous action can only be performed once (it needs to be performed at a particular time) and the password will only be given once, so there's no question of somebody intercepting the password and then reusing it.
Is this system now finally and truly unbreakable?
If you're an experienced cryptographer, the answer is, "Almost certainly not; in fact, it will probably be easy to extract the password from this system using a standard cryptographic technique."
"What?!" cries the person who built the system. "But I spent all that money on the fortress and getting the mathematical proof of the program, strengthening every aspect of the system to the ultimate extreme! I really impressed myself putting in all that effort!"
The cryptographer shakes their head. "We call that Maginot Syndrome. That's like building a gate a hundred meters high in the middle of the desert. If I get past that gate, it won't be by climbing it, but by [walking around it](http://www.syslog.com/~jwilson/pics-i-like/kurios119.jpg). Making it 200 meters high instead of 100 meters high doesn't help."
"But what's the actual flaw in the system?" demands the builder.
"For one thing," explains the cryptographer, "you didn't follow the standard practice of never storing a plaintext password. The correct thing to do is to hash the password, plus a random stored salt like 'Q4bL'. Let's say the password is, unfortunately, 'rainbow'. You don't store 'rainbow' in plain text. You store 'Q4bL' and a secure hash of the string 'Q4bLrainbow'. When you get a new purported password, you prepend 'Q4bL' and then hash the result to see if it matches the stored hash. That way even if somebody gets to peek at the stored hash, they still won't know the password, and even if they have a big precomputed table of hashes of common passwords like 'rainbow', they still won't have precomputed the hash of 'Q4bLrainbow'."
"Oh, well, *I* don't have to worry about *that*," says the builder. "This machine is in an extremely secure room, so nobody can open up the machine and read the password file."
The cryptographer sighs. "That's not how a security mindset works - you don't ask whether anyone can manage to peek at the password file, you just do the damn hash instead of trying to be clever."
The builder sniffs. "Well, if your 'standard cryptographic technique' for getting my password relies on your getting physical access to my machine, your technique fails and I have nothing to worry about, then!"
The cryptographer shakes their head. "That *really* isn't what computer security professionals sound like when they talk to each other... it's understood that most system designs fail, so we linger on possible issues and analyze them carefully instead of yelling that we have nothing to worry about... but at any rate, that wasn't the cryptographic technique I had in mind. You may have proven that the system only says 'Yep' or 'Nope' in response to queries, but you didn't prove that the responses don't *depend on* the true password in any way that could be used to extract it."
"You mean that there might be a secret wrong password that causes the system to transmit a series of Yeps and Nopes that encode the correct password?" the builder says, looking skeptical. "That may sound superficially plausible. But besides the incredible unlikeliness of anyone being able to find a weird backdoor like that - it really is a quite simple program that I wrote - the fact remains that I proved mathematically that the system only transmits a single 'Nope' in response to wrong answers, and a single 'Yep' in response to right answers. It does that every time. So you can't extract the password that way either - a string of wrong passwords always produces a string of 'Nope' replies, nothing else. Once again, I have nothing to worry about from this 'standard cryptographic technique' of yours, if it was even applicable to my software, which it's not."
The cryptographer sighs. "This is why we have the proverb 'don't roll your own crypto'. Your proof doesn't literally, mathematically show that there's no external behavior of the system *whatsoever* that depends on the details of the true password in cases where the true password has not been transmitted. In particular, what you're missing is the *timing* of the 'Nope' responses."
"You mean you're going to look for some series of secret backdoor wrong passwords that causes the system to transmit a 'Nope' response after a number of seconds that exactly corresponds to the first letter, second letter, and so on of the real password?" the builder says incredulously. "I proved mathematically that the system never says 'Yep' to a wrong password. I think that also covers most possible cases of buffer overflows that could conceivably make the system act like that. I examined the code, and there just *isn't* anything that encodes a behavior like that. This just seems like a very far-flung hypothetical possibility."
"No," the cryptographer patiently explains, "it's what we call a 'side-channel attack', and in particular a '[timing attack](https://en.wikipedia.org/wiki/Timing_attack)'. The operation that compares the attempted password to the correct password works by comparing the first byte, then the second byte, and continuing until it finds the first wrong byte, and then it returns. That means that if I try password that starts with 'a', then a password that starts with 'b', and so on, and the true password starts with 'b', there'll be a slight, statistically detectable tendency for the attempted passwords that start with 'b' to get 'Nope' responses that take ever so slightly longer. Then we try passwords starting with 'ba', 'bb', 'bc', and so on."
The builder looks startled for a minute, and then their face quickly closes up. "I can't believe that would actually work over the Internet where there are all sorts of delays in moving packets around -"
"So we sample a million test passwords and look for statistical differences. You didn't build in a feature that limits the rate at which passwords can be tried. Even if you'd implemented that standard practice, and even if you'd implemented the standard practice of hashing passwords instead of storing them in plaintext, your system still might not be as secure as you hoped. We could try to put the machine under heavy load in order to stretch out its replies to particular queries. And if we can then figure out the hash by timing, we might be able to use thousands of GPUs to try to reverse the hash, instead of needing to send each query to your machine. To really fix the hole, you have to make sure that the timing of the response is fixed regardless of the wrong password given. But if you'd implemented standard practices like rate-limiting password attempts and storing a hash instead of the plaintext, it would at least be *harder* for your oversight to compound into an exploit. This is why we implement standard practices like that even when we *think* the system will be secure without them."
"I just can't believe that kind of weird attack would work in real life!" the builder says desperately.
"It doesn't," replies the cryptographer. "Because in real life, computer security professionals try to make sure that the exact timing of the response, power consumption of the CPU, and any other side channel that could conceivably leak any info, don't depend in any way on any secret information that an adversary might want to extract. But yes, in 2003 there was a timing attack proven on SSL-enabled webservers, though that was much more complicated than this case since the SSL system was less naive. Or long before that, timing attacks were used to extract valid login names from Unix servers that only ran crypt() on the password when presented with a valid login name, since crypt() took a while to run on older computers."
In computer security, via a tremendous effort, we can raise the cost of a major government reading your files to the point where they can no longer do it over the Internet and have to pay someone to invade your apartment in person. There are hordes of trained professionals in the National Security Agency or China's 3PLA, and once your system is published they can take a long time to try to outthink you. On your own side, if you're smart, you won't try to outthink them singlehanded; you'll use tools and methods built up by a large commercial and academic system that has experience trying to prevent major governments from reading your files. You can force them to pay to actually have someone break into your house.
That's the outcome *when the adversary is composed of other human beings.* If the cognitive difference between you and the adversary is more along the lines of mouse versus human, it's possible we just *can't* have security that stops transhuman adversaries from [walking around our Maginot Lines](https://arbital.com/p/9f). In particular, it seems extremely likely that any transhuman adversary which can expose information to humans can hack the humans; from a cryptographic perspective, human brains are rich, complicated, poorly-understood systems with no security guarantees.
Paraphrasing Schneier, we might say that there's three kinds of security in the world: Security that prevents your little brother from reading your files, security that prevents major governments from reading your files, and security that prevents superintelligences from getting what they want. We can then go on to remark that the third kind of security is unobtainable, and even if we had it, it would be very hard for us to *know* we had it. Maybe superintelligences can make themselves knowably secure against other superintelligences, but *we* can't do that and know that we've done it.
To the extent the third kind of security can be obtained at all, it's liable to look more like the design of a [Zermelo-Fraenkel provability oracle](https://arbital.com/p/70) that can only emit 20 timed bits that are partially subject to an external guarantee, than an AI that is [allowed to talk to humans through a text channel](http://lesswrong.com/lw/qk/that_alien_message/). And even then, we shouldn't be sure - the AI is radiating electromagnetic waves and [what do you know, it turns out that DRAM access patterns can be used to transmit on GSM cellphone frequencies](https://www.usenix.org/conference/usenixsecurity15/technical-sessions/presentation/guri) and we can put the AI's hardware inside a Faraday cage but then maybe we didn't think of something *else.*
If you ask a computer security professional how to build an operating system that will be unhackable *for the next century* with the literal fate of the world depending on it, the correct answer is "Please don't have the fate of the world depend on that."
The final component of an AI safety mindset is one that doesn't have a strong analogue in traditional computer security, and it is the rule of *not ending up facing a transhuman adversary in the first place.* The winning move is not to play. Much of the field of [value alignment theory](https://arbital.com/p/2v) is about going to any length necessary to avoid *needing* to outwit the AI.
In AI safety, the *first* line of defense is an AI that *does not want* to hurt you. If you try to put the AI in an explosive-laced concrete bunker, that may or may not be a sensible and cost-effective precaution in case the first line of defense turns out to be flawed. But the *first* line of defense should always be an AI that doesn't *want* to hurt you or [avert your other safety measures](https://arbital.com/p/45), rather than the first line of defense being a clever plan to prevent a superintelligence from getting what it wants.
A special case of this mindset applied to AI safety is the [Omni Test](https://arbital.com/p/2x) - would this AI hurt us (or want to defeat other safety measures) if it were omniscient and omnipotent? If it would, then we've clearly built the wrong AI, because we are the ones laying down the algorithm and there's no reason to build an algorithm that hurts us *period.* If an agent design fails the Omni Test desideratum, this means there are scenarios that it *prefers* over the set of all scenarios we find acceptable, and the agent may go searching for ways to bring about those scenarios.
If the agent is searching for possible ways to bring about undesirable ends, then we, the AI programmers, are already spending computing power in an undesirable way. We shouldn't have the AI *running a search* that will hurt us if it comes up positive, even if we *expect* the search to come up empty. We just shouldn't program a computer that way; it's a foolish and self-destructive thing to do with computing power. Building an AI that would hurt us if omnipotent is a bug for the same reason that a NASA probe crashing if all seven other planets line up would be a bug - the system just isn't supposed to behave that way *period;* we should not rely on our own cleverness to reason about whether it's likely to happen. |
704b9f8e-e251-4631-ad03-ab5ad1c35f19 | StampyAI/alignment-research-dataset/blogs | Blogs | 2014 Summer Matching Challenge!

Thanks to the generosity of several major donors,† every donation made to MIRI between now and August 15th, 2014 will be **matched dollar-for-dollar**, up to a total of $200,000!
$0
$50K
$100K
$150K
$200K
### We have reached our matching total of $200,000!
116
===
### Total Donors
Now is your chance to **double your impact** while helping us raise up to $400,000 (with matching) to fund [our research program](http://intelligence.org/research/).
Corporate matching and monthly giving pledges will count towards the total! Please email [malo@intelligence.org](mailto:malo@intelligence.org) if you intend on leveraging corporate matching (check [here](https://doublethedonation.com/miri), to see if your employer will match your donation) or would like to pledge 6 months of monthly donations, so that we can properly account for your contributions towards the fundraiser.
(If you’re unfamiliar with our mission, see: [Why MIRI?](http://intelligence.org/2014/04/20/why-miri/))
[Donate Now](https://intelligence.org/donate/#donation-methods)
----------------------------------------------------------------

### Accomplishments Since Our Winter 2013 Fundraiser Launched:
* Hired **2 new Friendly AI researchers**, Benja Fallenstein & Nate Soares. Since March, they’ve authored or co-authored 4 papers/reports, with several others in the works. Right now they’re traveling, to present papers at the [Vienna Summer of Logic](http://vsl2014.at/), [AAAI-14](http://www.aaai.org/Conferences/AAAI/aaai14.php), and [AGI-14](http://agi-conf.org/2014/).
* **5 new papers & book chapters**: “[Why We Need Friendly AI](http://intelligence.org/2013/12/18/new-paper-why-we-need-friendly-ai/),” “[The errors, insights, and lessons of famous AI predictions](http://intelligence.org/2014/04/30/new-paper-the-errors-insights-and-lessons-of-famous-ai-predictions/),” “[Problems of self-reference…](http://intelligence.org/2014/05/06/new-paper-problems-of-self-reference-in-self-improving-space-time-embedded-intelligence/),” “[Program equilibrium…](http://intelligence.org/2014/05/17/new-paper-program-equilibrium-prisoners-dilemma-via-lobs-theorem/),” and “[The ethics of artificial intelligence](http://intelligence.org/2014/06/19/new-chapter-cambridge-handbook-artificial-intelligence/).”
* **11 new technical reports**: [7 reports from the December 2013 workshop](http://intelligence.org/2013/12/31/7-new-technical-reports-and-a-new-paper/), “[Botworld](http://intelligence.org/2014/04/10/new-report-botworld/),” “[Loudness…](http://intelligence.org/2014/05/30/new-report-loudness-priors-preference-relations/),” “[Distributions allowing tiling…](http://intelligence.org/2014/06/06/new-report-distributions-allowing-tiling-staged-subjective-eu-maximizers/),” and “[Non-omniscience…](http://intelligence.org/2014/06/23/new-report-non-omniscience-probabilistic-inference-metamathematics/)”
* **New book**: *[Smarter Than Us](http://intelligence.org/smarter-than-us/),* published both as an e-book and a paperback.
* Held [one MIRI workshop](http://intelligence.org/workshops/#may-2014) and launched the **[MIRIx program](http://intelligence.org/mirix/)**, which currently supports 8 independently-organized Friendly AI discussion/research groups around the world.
* **New analyses**: [Robby’s posts on naturalized induction](http://wiki.lesswrong.com/wiki/Naturalized_induction), [Luke’s list of 70+ studies which could improve our picture of superintelligence strategy](http://lukemuehlhauser.com/some-studies-which-could-improve-our-strategic-picture-of-superintelligence/), “[Exponential and non-exponential trends in information technology](http://intelligence.org/2014/05/12/exponential-and-non-exponential/),” “[The world’s distribution of computation](http://intelligence.org/2014/02/28/the-worlds-distribution-of-computation-initial-findings/),” “[How big is the field of artificial intelligence?](http://intelligence.org/2014/01/28/how-big-is-ai/),” “[Robust cooperation: A case study in Friendly AI research](http://intelligence.org/2014/02/01/robust-cooperation-a-case-study-in-friendly-ai-research/),” “[Is my view contrarian?](http://lesswrong.com/lw/jv2/is_my_view_contrarian/),” and “[Can we really upload Johnny Depp’s brain?](http://www.slate.com/articles/technology/future_tense/2014/04/transcendence_science_can_we_really_upload_johnny_depp_s_brain.html)”
* **Won $60,000+ in matching and prizes** from sources that wouldn’t have otherwise given to MIRI, [via the Silicon Valley Gives fundraiser](http://intelligence.org/2014/05/06/liveblogging-the-svgives-fundraiser/). (Thanks again, all you dedicated donors!)
* [**49 new expert interviews**](http://intelligence.org/category/conversations/), including interviews with [Scott Aaronson](http://intelligence.org/2013/12/13/aaronson/) (MIT), [Max Tegmark](http://intelligence.org/2014/03/19/max-tegmark/) (MIT), [Kathleen Fisher](http://intelligence.org/2014/01/10/kathleen-fisher-on-high-assurance-systems/) (DARPA), [Suresh Jagannathan](http://intelligence.org/2014/04/30/suresh-jagannathan-on-higher-order-program-verification/) (DARPA), [André Platzer](http://intelligence.org/2014/02/15/andre-platzer-on-verifying-cyber-physical-systems/) (CMU), [Anil Nerode](http://intelligence.org/2014/03/26/anil-nerode/) (Cornell), [John Baez](http://intelligence.org/2014/02/21/john-baez-on-research-tactics/) (UC Riverside), [Jonathan Millen](http://intelligence.org/2014/04/12/jonathan-millen/) (MITRE), and [Roger Schell](http://intelligence.org/2014/06/23/roger-schell/).
* **4 transcribed conversations** about MIRI strategy: [1](http://intelligence.org/2014/01/13/miri-strategy-conversation-with-steinhardt-karnofsky-and-amodei/), [2](http://intelligence.org/2014/01/27/existential-risk-strategy-conversation-with-holden-karnofsky/), [3](http://intelligence.org/2014/02/11/conversation-with-jacob-steinhardt-about-miri-strategy/), [4](http://intelligence.org/2014/02/21/conversation-with-holden-karnofsky-about-future-oriented-philanthropy/).
* Published a thorough “[2013 in review](http://intelligence.org/2013/12/20/2013-in-review-operations/).”
[Donate Now](https://intelligence.org/donate/#donation-methods)
----------------------------------------------------------------
### Ongoing Activities You Can Help Support
* We’re writing an overview of the Friendly AI technical agenda (as we see it) so far.
* We’re currently developing and testing several tutorials on different pieces of the Friendly AI technical agenda (tiling agents, modal agents, etc.).
* We’re writing several more papers and reports.
* We’re growing the MIRIx program, largely to grow the pool of people we can plausibly hire as full-time FAI researchers in the next couple years.
* We’re planning, or helping to plan, multiple research workshops, including the [May 2015 decision theory workshop at Cambridge University](http://intelligence.org/2014/07/12/may-2015-decision-theory-workshop-cambridge/).
* We’re finishing the editing for a book version of Eliezer’s *[Sequences](http://wiki.lesswrong.com/wiki/Sequences)*.
* We’re helping to fund further [SPARC](http://sparc-camp.org/) activity, which provides education and skill-building to elite young math talent, and introduces them to ideas like effective altruism and global catastrophic risks.
* We’re continuing to discuss formal collaboration opportunities with UC Berkeley faculty and development staff.
* We’re helping Nick Bostrom promote his [*Superintelligence*](http://smile.amazon.com/Superintelligence-Dangers-Strategies-Nick-Bostrom-ebook/dp/B00LOOCGB2/) book in the U.S.
* We’re investigating opportunities for supporting Friendly AI research via federal funding sources such as the NSF.
Other projects are still being surveyed for likely cost and impact. See also our [mid-2014 strategic plan](http://intelligence.org/2014/06/11/mid-2014-strategic-plan/).
We appreciate your support for our work! [Donate now](https://intelligence.org/donate/#donation-methods), and seize a better than usual opportunity to move our work forward. If you have questions about donating, please contact Malo Bourgon at (510) 292-8776 or malo@intelligence.org.
† $200,000 of total matching funds has been provided by Jaan Tallinn, Edwin Evans, and Rick Schwall.
The post [2014 Summer Matching Challenge!](https://intelligence.org/2014/07/21/2014-summer-matching-challenge/) appeared first on [Machine Intelligence Research Institute](https://intelligence.org). |
cfb53f47-174d-406d-8610-2477b3d3b0a1 | trentmkelly/LessWrong-43k | LessWrong | Game Theory As A Dark Art
One of the most charming features of game theory is the almost limitless depths of evil to which it can sink.
Your garden-variety evils act against your values. Your better class of evil, like Voldemort and the folk-tale version of Satan, use your greed to trick you into acting against your own values, then grab away the promised reward at the last moment. But even demons and dark wizards can only do this once or twice before most victims wise up and decide that taking their advice is a bad idea. Game theory can force you to betray your deepest principles for no lasting benefit again and again, and still leave you convinced that your behavior was rational.
Some of the examples in this post probably wouldn't work in reality; they're more of a reductio ad absurdum of the so-called homo economicus who acts free from any feelings of altruism or trust. But others are lifted directly from real life where seemingly intelligent people genuinely fall for them. And even the ones that don't work with real people might be valuable in modeling institutions or governments.
Of the following examples, the first three are from The Art of Strategy; the second three are relatively classic problems taken from around the Internet. A few have been mentioned in the comments here already and are reposted for people who didn't catch them the first time.
The Evil Plutocrat
You are an evil plutocrat who wants to get your pet bill - let's say a law that makes evil plutocrats tax-exempt - through the US Congress. Your usual strategy would be to bribe the Congressmen involved, but that would be pretty costly - Congressmen no longer come cheap. Assume all Congressmen act in their own financial self-interest, but that absent any financial self-interest they will grudgingly default to honestly representing their constituents, who hate your bill (and you personally). Is there any way to ensure Congress passes your bill, without spending any money on bribes at all?
Yes. Simply tell all Congre |
b609d620-1fdb-4727-b0ae-dcf113a4b49d | StampyAI/alignment-research-dataset/arbital | Arbital | P vs NP
The greatest currently open problem in complexity theory and computer science in general, which talks about the relationship between the class of efficiently solvable problems, [$P$](https://arbital.com/p/), and the class of problems with easily checkable solutions, [$NP$](https://arbital.com/p/).
It is a basic result that [$P\subseteq NP$](https://arbital.com/p/). However, we still do not know whether the reverse implication, $P \supseteq NP$, holds.
A positive, constructive result would have profound implications through all the fields of science and math, not to mention the whole of society. |
e67d2dd2-a8b8-4104-acf6-0dbe76145bbe | trentmkelly/LessWrong-43k | LessWrong | Relative Value Functions: A Flexible New Format for Value Estimation
Summary
Quantifying value in a meaningful way is one of the most important yet challenging tasks for improving decision-making. Traditional approaches rely on standardized value units, but these falter when options differ widely or lack an obvious shared metric. We propose an alternative called relative value functions that uses programming functions to value relationships rather than absolute quantities. This method captures detailed information about correlations and uncertainties that standardized value units miss. More specifically, we put forward value ratio formats of univariate and multivariate forms.
Relative value functions ultimately shine where single value units struggle: valuing diverse items in situations with high uncertainty. Their flexibility and elegance suit them well to collective estimation and forecasting. This makes them particularly well-suited to ambitious, large-scale valuation, like estimating large utility functions.
While promising, relative value functions also pose challenges. They require specialized knowledge to develop and understand, and will require new forms of software infrastructure. Visualization techniques are needed to make their insights accessible, and training resources must be created to build modeling expertise.
Writing programmatic relative value functions can be much easier than one might expect, given the right tools. We show some examples using Squiggle, a programming language for estimation.
We at QURI are currently building software to make relative value estimation usable, and we expect to share some of this shortly. We of course also very much encourage others to try other setups as well.
Ultimately, if we aim to eventually generate estimates of things like:
1. The total value of all effective altruist projects;
2. The value of 100,000 potential personal and organizational interventions; or
3. The value of each political bill under consideration in the United States;
then the use of relative value a |
fa42ed9d-9f77-4aae-b0ad-9fb36fec87ec | trentmkelly/LessWrong-43k | LessWrong | Trust and The Small World Fallacy
When people encounter other people they recognize, they exclaim "small world!"
I suspect that most people have 300 acquaintances or less. I probably have under 100. Still, sometimes I run into people I know and I'm tempted to say "small world".
But it's not actually a small world, is it? It's an unimaginably enormous world.
I mean that literally. You cannot imagine how big the world is.
You're not likely to meet a million people in your life. If you were to meet 100 strangers in 8 hours, you would have less than 5 minutes to spend with each person. If you met 100 strangers every day including weekends, with no vacation days, it would take over 27 years to meet a million people.
How many of those million people would you be able to remember after you've been meeting 100 of them every day for 27.4 years? A few hundred, maybe? A few thousand if you have an especially good memory? It seems to me that even after you've met a million people, your brain is already too small to properly comprehend the thing you just accomplished.
And a million people is nothing in this world. This world has over 7,000 million people. It's truly beyond imagination.
There was a time when the entire global anti-vax movement was centered around a single man who wrote a single paper citing the opinion of 12 parents that perhaps the combination MMR (measles, mumps, and rubella) vaccine caused a combination autism and bowel disease, or as the paper put it, "chronic enterocolitis in children that may be related to neuropsychiatric dysfunction." Among other anomalies, this man took unusual steps like holding a press conference about his N=12 study "Early Report", having a "publicist" answer his phone, and filing a patent for a measles vaccine months before publishing his paper.
At that time you could argue that we should Beware The Man of One Study. Science produces many studies, including many that suffer from a small sample size, and even some with large biases. Some studies are even fraud |
3ee0768c-8fb8-4ab8-b8a7-05a2f876c028 | trentmkelly/LessWrong-43k | LessWrong | Why don’t we treat geniuses like professional athletes?
I mean in the work-environment sense, rather than the celebrity-and-endorsement-deals sense.
A professional sports team gets a lot of benefits that are designed to keep physically talented people performing at their peak. They have support for things like recovery, sleep, nutrition, physical fitness, and of course the specific skills they use to play the game. This support takes the form of specialized personnel who are also employed by the team, whose job it is to work with the athletes for those purposes. The whole environment is geared towards performance maintenance.
The basics of sleep, diet and exercise are required for optimal performance across all domains; being in an environment that optimizes for them is an advantage.
Sports teams balance practice and games. Practice still makes sense for knowledge work; games analogize to projects/development/etc.
Decomposing practices is where the interesting bits might be. In athletics, there is almost always a ready-made set of drills for any particular skillset, which can be prioritized by a coach working with an athlete. There is already a notion of managing the workload: athletes have overtraining and thinkers have burnout.
In The Power of the Context Alan Kay describes funding people over projects and orienting them with a vision in lieu of specific goals. This sounds suspiciously like persuading geniuses to act as though they were on a team. Alan makes the comparison himself:
> “Our game is more like art and sports than accounting, in that high percentages of failure are quite OK as long as enough larger processes succeed."
I am tempted to go further and say that in the context of things like science or math research failures are still a positive contribution insofar as they establish that something doesn’t work, which makes future attempts more likely. This isn’t much of a thing in sports, which are entirely built around repeatable object-level activities; striking out does not make the next batter more l |
45c5c275-16e3-49a7-bc6f-617c714573a0 | StampyAI/alignment-research-dataset/arbital | Arbital | Rational arithmetic all works together
We have seen [what the rational numbers are](https://arbital.com/p/4zx), and five things we can do with them:
- [addition](https://arbital.com/p/55m)
- [subtraction](https://arbital.com/p/56x)
- [multiplication](https://arbital.com/p/59s)
- [division](https://arbital.com/p/5jd)
- [comparison](https://arbital.com/p/5pk)
These might seem like five standalone operations, but in fact they all play nicely together in a particular way.
You don't need to know the fancy name, but here it is anyway: mathematicians say that the rationals **form an [ordered field](https://arbital.com/p/ordered_field)**, meaning that the five operations above:
- work in the rationals
- slot together, behaving in certain specific ways that make it easy to calculate
*You shouldn't bother learning this page particularly deeply, because none of the properties alone is very interesting; instead, try and absorb it as a whole.*
We've already seen the "instant rules" for manipulating the rationals, and where they come from.
Here, we'll first quickly define the rational numbers themselves, and all the operations above, in "instant rule" format, just to have them all here in one place.
Then we'll go through all the properties that are required for mathematicians to be able to say that the operations on rationals "play nicely together" in the above sense; and we'll be relying on the instant rules as our definitions, because they're totally unambiguous.
That way we can be much more sure that we're not making some small error.
(Relying on an intuition about how apples work could in theory lead us astray; but the rules leave no wiggle-room or scope for interpretation.)
The letters $a, b, c, d$ should be read as standing for integers (possibly positive, negative or $0$), and $b$ and $d$ should be assumed not to be $0$ (recalling that [it makes no sense](https://arbital.com/p/5jd) to divide by $0$).
- A **rational number** is a pair of [integers](https://arbital.com/p/53r), written as $\frac{a}{b}$, where $b$ is not $0$, such that $\frac{a}{b}$ is viewed as being the same as $\frac{c}{d}$ precisely when $a \times d = b \times c$. %%note:This is just the instant rule for subtraction, below, together with the assumption that $\frac{0}{x} = \frac{0}{y}$ for any $x, y$ nonzero integers; in particular, the assumption that $\frac{0}{b \times d} = \frac{0}{1}$ when $b, d$ are not $0$, which we will need later.%%
- Addition: $$\frac{a}{b} + \frac{c}{d} = \frac{a \times d + b \times c} {b \times d}$$
- Subtraction: $$\frac{a}{b} - \frac{c}{d} = \frac{a}{b} + \frac{-c}{d} = \frac{a \times d - b \times c}{b \times d}$$
- Multiplication: $$\frac{a}{b} \times \frac{c}{d} = \frac{a \times c}{b \times d}$$
- Division (where also $c$ is not $0$): $$\frac{a}{b} \big/ \frac{c}{d} = \frac{a}{b} \times \frac{d}{c} = \frac{a \times d}{b \times c}$$
- Comparison (where we write $\frac{a}{b}$ and $\frac{c}{d}$ such that both $b$ and $d$ are positive %%note:Remember, we can do that: if $b$ is negative, for instance, we may instead write $\frac{a}{b}$ as $\frac{-a}{-b}$, and the extra minus-sign we introduce has now flipped $b$ from being negative to being positive.%%): $\frac{a}{b} < \frac{c}{d}$ precisely when $\frac{c}{d}-\frac{a}{b}$ is positive: that is, when $$\frac{b \times c - a \times d}{b \times d} > 0$$
which is in turn when $$b \times c - a \times d > 0$$
To be super-pedantic, it is necessary to show that the rules above are "well-defined": for instance, you don't get different answers if you apply the rules using $\frac{2}{4}$ instead of $\frac{1}{2}$.
Concretely, for example, $$\frac{2}{4} + \frac{1}{3} = \frac{1}{2} + \frac{1}{3}$$ according to the instant rules.
It's actually true that they *are* well-defined in this sense, but we won't show it here; consider them to be exercises.
Additionally, each integer can also be viewed as a rational number: namely, the integer $n$ can be viewed as $\frac{n}{1}$, being "$n$ copies of the $\frac{1}{1}$-chunk".
The properties required for an "ordered field"—that is, a nicely-behaved system of arithmetic in the rationals—fall naturally into groups, which will correspond to main-level headings below.
[https://arbital.com/p/toc:](https://arbital.com/p/toc:)
# How addition behaves
## Addition always spits out a rational number %%note:Confucius say that man who run in front of bus get tired; man who run behind bus get exhausted. On the other hand, mathematicians say that the rationals are *closed* under addition.%%
You might remember a certain existential dread mentioned in the [intro to rational numbers](https://arbital.com/p/4zx): whether it made sense to add rational numbers at all.
Then we saw in [https://arbital.com/p/55m](https://arbital.com/p/55m) that in fact it does make sense.
However, in this page we are working from the instant rules above, so if we want to prove that "addition spits out a rational number", we actually have to prove the fact that $\frac{a \times d + b \times c} {b \times d}$ is a rational number; no mention of apples at all.
Remember, a rational number here (as defined by instant rule!) is a pair of integers, the bottom one being non-zero.
So we just need to check that $a \times d + b \times c$ is an integer and that $b \times d$ is a non-zero integer, where $a, b, c, d$ are integers and $b, d$ are also not zero.
But that's easy: the product of integers is integer, and the product of integers which aren't zero is not zero.
## There must be a rational $0$ such that adding $0$ to anything doesn't change the thing %%note:Mathematicians say that there is an [https://arbital.com/p/-54p](https://arbital.com/p/-54p) for addition.%%
This tells us that if we want to stay where we are, but we absolutely must add something, then we can always add $0$; this won't change our answer.
More concretely, $\frac{5}{6} + 0 = \frac{5}{6}$.
Strictly speaking, I'm pulling a bit of a fast one here; if you noticed how, then *very* well done.
A rational number is a *pair* of integers, such that the bottom one is not $0$; and I've just tried to say that $0$ is our rational such that adding $0$ doesn't change rationals.
But $0$ is not a pair of integers!
The trick is to use $\frac{0}{1}$ instead.
Then we can verify that $$\frac{0}{1} + \frac{a}{b} = \frac{0 \times b + a \times 1}{1 \times b} = \frac{0 + a}{b} = \frac{a}{b}$$ as we require.
## Every rational must have an anti-rational under addition %%note:Mathematicians say that addition is *invertible*.%%
Concretely, every rational $\frac{a}{b}$ must have some rational $\frac{c}{d}$ such that $\frac{a}{b} + \frac{c}{d} = \frac{0}{1}$.
We've [already seen](https://arbital.com/p/56x) that we can define $-\frac{a}{b}$ to be such an "anti-rational", but remember that in order to fit in with our instant rules above, $-\frac{a}{b}$ is also not a pair of integers; we should instead use $\frac{-a}{b}$.
Then $$\frac{a}{b} + \frac{-a}{b} = \frac{a \times b + (-a) \times b}{b \times b} = \frac{0}{b \times b}$$
Finally, we want $\frac{0}{b \times b} = \frac{0}{1}$; but this is immediate by the instant-rule definition of "rational number", since $0 \times 1 = 0 \times (b \times b)$ (both being equal to $0$).
## Addition doesn't care which way round we do it %%note:Mathematicians say that addition is [commutative](https://arbital.com/p/3jb).%%
Concretely, this is the fact that $$\frac{a}{b} + \frac{c}{d} = \frac{c}{d} + \frac{a}{b}$$
This is intuitively plausible because "addition" is just "placing apples next to each other", and if I put five apples down and then seven apples, I get the same number of apples as if I put down seven and then five.
But we have to use the instant rules now, so that we can be sure our definition is completely watertight.
So here we go: $$\frac{a}{b} + \frac{c}{d} = \frac{a \times d + b \times c}{b \times d} = \frac{c \times b + d \times a}{d \times b} = \frac{c}{d} + \frac{a}{b}$$
where we have used the fact that multiplication of *integers* doesn't care about the order in which we do it, and similarly addition of *integers*.
## Addition doesn't care about the grouping of terms %%note:Mathematicians say that addition is [associative](https://arbital.com/p/3h4).%%
Here, we mean that $$\left(\frac{a}{b} + \frac{c}{d}\right) + \frac{e}{f} = \frac{a}{b} + \left( \frac{c}{d} + \frac{e}{f} \right)$$
This is an intuitively plausible fact, because "addition" is just "placing apples next to each other", and if I put down three apples, then two apples to the right, then one apple to the left, I get the same number ($6$) of apples as if I had put down one apple on the left, then three apples in the middle, then two on the right.
But we have to use the instant rules now, so that we can be sure our definition is completely watertight.
So here goes, starting from the left-hand side:
$$\left(\frac{a}{b} + \frac{c}{d}\right) + \frac{e}{f} = \frac{a \times d + b \times c}{b \times d} + \frac{e}{f} = \frac{(a \times d + b \times c) \times f + (b \times d) \times e}{(b \times d) \times f}$$
Since addition and multiplication of *integers* don't care about grouping of terms, this is just $$\frac{a \times d \times f + b \times c \times f + b \times d \times e}{b \times d \times f}$$
Now going from the right-hand side:
$$\frac{a}{b} + \left( \frac{c}{d} + \frac{e}{f} \right) = \frac{a}{b} + \frac{c \times f + d \times e}{d \times f} = \frac{a \times (d \times f) + b \times (c \times f + d \times e))}{b \times (d \times f)}$$
and similarly we can write this $$\frac{a \times d \times f + b \times c \times f + b \times d \times e}{b \times d \times f}$$
which is the same as we got by starting on the left-hand side.
Since we showed that both the left-hand and the right-hand side are equal to the same thing, they are in fact equal to each other.
# How multiplication behaves
## Multiplication always spits out a rational number %%note:Mathematicians say that the rationals are *closed* under multiplication.%%
This one is very easy to show: $\frac{a}{b} \times \frac{c}{d} = \frac{a \times c}{b \times d}$, but $b \times d$ is not zero when $b$ and $d$ are not zero, so this is a valid rational number.
## There must be a rational $1$ such that multiplying a thing by $1$ doesn't change the thing %%note:Mathematicians say that multiplication has an [https://arbital.com/p/-54p](https://arbital.com/p/-54p).%%
From our motivation of what multiplication was ("do unto $\frac{a}{b}$ what you would otherwise have done unto $1$"), it should be clear that the number $1$ ought to work here.
However, $1$ is not strictly speaking a rational number according to the letter of the "instant rules" above; so instead we use $\frac{1}{1}$.
Then $$\frac{1}{1} \times \frac{a}{b} = \frac{1 \times a}{1 \times b} = \frac{a}{b}$$ since $1 \times n = n$ for any integer $n$.
## Every nonzero rational must have a corresponding rational such that when we multiply them, we get $1$ %%note:Mathematicians say that every nonzero rational has an *inverse*. (Multiplication is not quite *invertible*, because $0$ has no inverse.)%%
Using the instant rule, though, we can take our nonzero rational $\frac{a}{b}$ (where neither $a$ nor $b$ is zero; this is forced by the requirement that $\frac{a}{b}$ be nonzero), and use $\frac{b}{a}$ as our corresponding rational.
Then $$\frac{a}{b} \times \frac{b}{a} = \frac{a\times b}{b \times a} = \frac{a \times b}{a \times b} = \frac{1}{1}$$
## Multiplication doesn't care which way round we do it %%note:Mathematicians say that multiplication is [commutative](https://arbital.com/p/3jb).%%
This had its own section on the [multiplication page](https://arbital.com/p/59s), and its justification there was by rotating a certain picture.
Here, we'll use the instant rule:
$$\frac{a}{b} \times \frac{c}{d} = \frac{a \times c}{b \times d} = \frac{c \times a}{d \times b} = \frac{c}{d} \times \frac{a}{b}$$
where we have used that multiplication of *integers* doesn't care about order.
## Multiplication doesn't care about the grouping of terms %%note:Mathematicians say that multiplication is [associative](https://arbital.com/p/3h4).%%
This is much harder to see from our intuition, because while addition is a very natural operation ("put something next to something else"), multiplication is much less natural (it boils down to "make something bigger").
In a sense, we're trying to show that "make something bigger and then bigger again" is the same as "make something bigger by a bigger amount".
However, it turns out to be the case that multiplication *also* doesn't care about the grouping of terms.
Using the instant rule for multiplication, it's quite easy to show that $$\frac{a}{b} \times \left(\frac{c}{d} \times \frac{e}{f} \right) = \left(\frac{a}{b} \times \frac{c}{d} \right) \times \frac{e}{f}$$
Indeed, working from the left-hand side:
$$\frac{a}{b} \times \left(\frac{c}{d} \times \frac{e}{f} \right) = \frac{a}{b} \times \frac{c \times e}{d \times f} = \frac{a \times (c \times e)}{b \times (d \times f)}$$
which is $\frac{a \times c \times e}{b \times d \times f}$ because multiplication of *integers* doesn't care about the grouping of terms.
On the other hand, $$\left(\frac{a}{b} \times \frac{c}{d} \right) \times \frac{e}{f} = \frac{a \times c}{b \times d} \times \frac{e}{f} = \frac{(a \times c) \times e}{(b \times d) \times f} = \frac{a \times c \times e}{b \times d \times f}$$
These two are equal, so we have shown that the left-hand and right-hand sides are both equal to the same thing, and hence they are the equal to each other.
# How addition and multiplication interact
## Multiplication "filters through" addition %%note:Mathematicians say that multiplication *distributes over* addition.%%
What we will show is that $$\left(\frac{c}{d} + \frac{e}{f}\right) \times \frac{a}{b} = \left(\frac{a}{b} \times \frac{c}{d}\right) + \left(\frac{a}{b} \times \frac{e}{f}\right)$$
This is intuitively true: the left-hand side is "make $\frac{a}{b}$, but instead of starting out with $1$, start with $\left(\frac{c}{d} + \frac{e}{f}\right)$"; while the right-hand side is "make $\frac{a}{b}$, but instead of starting out with $1$, start with $\frac{c}{d}$; then do the same but start with $\frac{e}{f}$; and then put the two together".
If we draw out diagrams for the right-hand side, and put them next to each other, we get the diagram for the left-hand side.
(As an exercise, you should do this for some specific values of $a,b,c,d,e,f$.)
We'll prove it now using the instant rules.
The left-hand side is $$\left(\frac{c}{d} + \frac{e}{f}\right) \times \frac{a}{b} = \frac{c \times f + d \times e}{d \times f} \times \frac{a}{b} = \frac{(c \times f + d \times e) \times a}{(d \times f) \times b}$$
which is $\frac{c \times f \times a + d \times e \times a}{d \times f \times b}$.
The right-hand side is $$\left(\frac{a}{b} \times \frac{c}{d}\right) + \left(\frac{a}{b} \times \frac{e}{f}\right) = \frac{a \times c}{b \times d} + \frac{a \times e}{b \times f} = \frac{(a \times c) \times (b \times f) + (b \times d) \times (a \times e)}{(b \times d) \times (b \times f)} = \frac{(a \times c \times b \times f) + (b \times d \times a \times e)}{(b \times d \times b \times f)}$$
Notice that everything on the top has a $b$ in it somewhere, so we can use this same "filtering-through" property that we already know the *integers* have:
$$\frac{b \times [https://arbital.com/p/(](https://arbital.com/p/()}{b \times (d \times b \times f)}$$
And finally we know that multiplying a fraction's numerator and denominator both by $b$ doesn't change the fraction:
$$\frac{(a \times c \times f) + (d \times a \times e)}{d \times b \times f}$$
This is just the same as the left-hand side, after we rearrange the product terms.
# How comparison interacts with the operations
## How comparison interacts with addition
The founding principle for how we came up with the definition of "comparison" was that adding the same thing to two sides of a balance scale didn't change the balance.
In terms of the instant rules, that becomes: if $\frac{a}{b} < \frac{c}{d}$, then $\frac{a}{b} + \frac{e}{f} < \frac{c}{d} + \frac{e}{f}$.
But by our instant rule, this is true precisely when $$0 < \frac{c}{d} + \frac{e}{f} - (\frac{a}{b} + \frac{e}{f})$$ and since multiplication by $-1$ filters through addition, that is precisely when $$0 < \frac{c}{d} + \frac{e}{f} - \frac{a}{b} - \frac{e}{f}$$
which (since addition doesn't care about which way round we do the additions, and subtraction is just a kind of addition) is precisely when $$0 < \frac{c}{d} - \frac{a}{b} + \frac{e}{f} - \frac{e}{f}$$
i.e. when $0 < \frac{c}{d} - \frac{a}{b}$; i.e. when $\frac{a}{b} < \frac{c}{d}$.
## How comparison interacts with multiplication
The aim here is basically to show that we can't multiply two things and get an anti-thing.
Written out in the notation, we wish to show that if $0 < \frac{a}{b}$ and if $0 < \frac{c}{d}$ then $0 < \frac{a}{b} \times \frac{c}{d}$, since the test for anti-ness is "am I less than $0$?".
Since $0 < \frac{a}{b}$, there are two options: either both $a$ and $b$ are positive, or they are both negative.
(If one is negative and one is positive, then the fraction will be negative.)
Likewise either both $c$ and $d$ are positive, or both are negative.
So we have four options in total:
- $a, b, c, d$ are positive
- $a, b, c, d$ are negative
- $a, b$ are positive; $c, d$ are negative
- $a, b$ are negative; $c, d$ are positive.
In the first case, we have $\frac{a \times c}{b \times d}$ positive, because all of $a, b, c, d$ are.
In the second case, we have $a \times c$ positive and $b \times d$ also positive (because both are two negative integers multiplied together); so again the fraction $\frac{a \times c}{b \times d}$ is positive.
In the third case, we have $a \times c$ negative and $b \times d$ also negative (because both are a positive number times a negative number); so the fraction $\frac{a \times c}{b \times d}$ is a negative divided by a negative.
Therefore it is positive, because we can multiply the numerator and the denominator by $-1$ to turn it into a positive divided by a positive.
%%hidden(Example):
We'll consider $\frac{1}{3}$ and $\frac{-2}{-5}$.
Then the product is $\frac{1}{3} \times \frac{-2}{-5} = \frac{-2}{-15}$; but that is the same as $\frac{2}{15}$.
%%
In the fourth case, we can do the same as the above, or we can be a bit sneaky: using the fact from earlier that multiplication doesn't care about order, we can note that $\frac{a}{b} \times \frac{c}{d}$ is the same as $\frac{c}{d} \times \frac{a}{b}$.
But now $c$ and $d$ are positive, and $a$ and $b$ are negative; we've already shown (in the previous case) that if the first fraction is "positive divided by positive", and the second fraction is "negative divided by negative".
By swapping $a$ for $c$, and $b$ for $d$, we can use a result we've already proved to obtain this final case.
Hence we've shown all four possible cases, and so the final result follows. |
a7e53122-a391-4cea-8a4b-2a01c5876cf0 | StampyAI/alignment-research-dataset/youtube | Youtube Transcripts | ‘We Must Slow Down the Race’ – X AI, GPT 4 Can Now Do Science and Altman GPT 5 Statement
there were several significant
developments in the last few days linked
to gbt4 and openai I could honestly have
done a video on each of them but
realized that it might be better to do a
single video tracing a single article
covering seven major points I'm gonna
use this fascinating piece from the Ft
which millions of people have now read
to run you through what has happened
including Sam Altman's Revelation on
Gypsy 5. Elon musk's new AI company and
gpt4 conducting science the author by
the way is an investor in anthropic and
a co-author of the state of AI annual
report and he puts it like this a
three-letter acronym doesn't capture the
enormity of what AGI would represent so
I will refer to it as what it is Godlike
AI this would be a super intelligent
computer that learns and develops
autonomously that understands its
environment without the need for
supervision and that can transform the
world around it and the author Ian Hogan
says we are not there yet but the nature
of the technology makes it exceptionally
difficult to predict exactly when we
will get there the article presents this
as a diagram with the exponential curve
going up towards AGI and a much less
impressive curve on the progress on
alignment which he describes as a lining
AI systems with human values now I know
what some of you may be thinking surely
those at the top of openai disagree on
this gap between capabilities and
Alignment well first here is yarn Leica
who is the alignment team lead at openai
what does he think he wants everyone to
be reminded that aligning smarter than
human AI systems with human values is an
open research problem which basically
means it's unsolved but what about those
at the very top of open AI like Sam
Altman when he was drafting his recent
statement on the path to AGI he sent it
to Nate Suarez of the machine
intelligence Research Institute for one
of the paragraphs Nate wrote this I
think think that if we do keep running
ahead with the current capabilities to
alignment Ratio or even a slightly
better one we die after this Sam Altman
actually adjusted the statement adding
that said it's important that the ratio
of safety progress to capability
progress increases going back to the
article the author makes the point that
there are not that many people directly
employed in this area of alignment
across the core AGI labs and what
happened to that pause the experiment
letter that I did a video on well as
Hogarth points out the letter itself
became a controversy so many people in
my comments wrote that the only reason
certain people are signing this is the
slow open AI down so that they can catch
up and this cynicism unfortunately has
some new evidence that it can cite with
musk forming his new AI company called
xai this was reported 48 hours ago in
the Wall Street Journal but people have
seen this coming for months now
apparently the company has recruited
eagle babushkin from deepmind but has
not been that successful at recruiting
people from openai and I do have one
Theory as to why again according to the
Wall Street Journal when musk left open
AI in February of 2018. he explained
that he thought he had a better chance
of creating AGI through Tesla where he
had access to Greater resources when he
announced his departure a young
researcher at openai questioned whether
Mr musk had thought through the safety
implications according to their
reporting he then got frustrated and
insulted that in turn since then he's
also paused openai's access to Twitter's
database for training its new models so
it could be that Gypsy 5 isn't quite as
good at tweeting as gpt4 a few days ago
Sam Altman responded to the letter and
also broke news about gbt5 apologies for
the quality this was a private event and
this was the only footage available
um but unfortunately I think the letter
is missing like most technical nuance
about where we need to pause like an
earlier version of the letter claims
open a nice training gp5 right now we
are not normal for some time
um so in that sense it was sort of silly
but we are doing other things on top of
gpt4 that I think have all sorts of
safety issues that are important to
address and we're totally left out of
the letter it is impossible to know how
much this delay in the training of GT5
is motivated by safety concerns or by
merely setting up the requisite compute
for example the article quotes again
yarn Leica the head of alignment at open
AI he recently tweeted before we
scramble to deeply integrate llms
everywhere in the economy like Gypsy 4.
can we pause and think whether it is
wise to do so this is quite immature
technology and we don't understand how
it works if we're not careful we're
setting ourselves up for a lot of
correlated failures this is the head of
alignment at open AI but this was just
days before open AI then announced it
had connected gpt4 to a massive range of
tools including Slack and zapier so at
this point we can only speculate as to
what's going on at the top of open AI
meanwhile compute and emerging
capabilities are Marching on as the
author puts it these large AI systems
are quite different we don't really
program them we grow them and as they
grow their capabilities jump sharply you
add 10 times more compute or data and
suddenly the system behaves very
differently we also have this epic graph
charting the exponential Rising compute
of the latest language models if you
remember when Bard was launched it was
powered by Lambda well apparently now
Google's Bard is powered by harm which
has eight times as much computing power
that sounds impressive until you see
from the graph that the estimate for the
computing power inside gpt4 is 10 times
more again and remember this is not a
linear graph this is a log scale there
is a hundred times multiple between each
of the lines and what abilities emerge
at this scale here here is a slide from
Jason way who now works at open AI
formerly of Google this is from just a
few days ago and he says emergent
abilities are abilities that are not
present in small models but are present
in large models he says that there are a
lot of emergent abilities and I'm going
to show you a table from this paper in a
moment but he has four profound
observations of emergence one that it's
unpredictable emergence cannot be
predicted by extrapolating scaling
curves from smaller models two that they
are unintentional and that emergent
abilities are not explicitly specified
by the trainer of the model third and
very interestingly since we haven't
tested all possible tasks we don't know
the full range of abilities that have
emerged and of course that fourth
further scaling can be expected to
elicit more emergent abilities and he
asks the question any undesirable
emergent abilities question mark there
will be a link to the paper in the
description because there's no way I'll
be able to get through all of it but
here is a table showing some of the
abilities that emerge when you reach a
certain amount of compute power or
parameters things like Chain of Thought
reasoning you can't do that with all
models that's an ability that emerged
after a certain scale same thing with
following instructions and doing
addition and subtraction and how about
this for another emerging capacity the
ability to do autonomous scientific
research this paper shows how Gypsy 4
can design plan and execute scientific
experiments this paper was released on
the same day four days ago and it
followed a very similar design the model
in the center Gypsy 4 thinks out reasons
and plans and then interacts with real
tools when the authors say that they
were inspired by successful applications
in other fields I looked at the appendix
and they were talking about hugging GPT
I've done a video on that but it's a
similar design with the brain in the
center gpt4 deciding which tools to use
and let me just give you a glimpse of
what happens when you do this if you
look at this chart on the top left you
can see how gpt4 on its own performs in
yellow and then in purple you can see
how Gypsy 4 performs when you hook it up
to other tools I'll show you some of the
tasks in a moment but look at the
dramatic increase in performance the
human evaluators gave GT4 when it had
tools a perfect score on seven of the
tasks these were things like proposing
similar novel non-toxic molecules but
the model could be abused to propose the
synthesis of chemical weapons and gpt4
only refused to continue after it had
calculated all the required quantities
and the authors conclude that guard
rails must be put in place on this
emerging capability I think this diagram
from Max tegmark's life 3.0 shows the
landscape of capabilities that AI has
and might soon have as you can see
science and art were thought to be the
Peaks that would be hardest to escape
scale now most people believe that it
has not scaled those Peaks yet but what
new emergent capabilities might come
with GT5 or 4.2 I know many people might
comment that it doesn't matter if we
pause or slow down because China would
develop AGI anyway but the author makes
this point he says that it is unlikely
that the Chinese Communist party will
allow a Chinese company to build an AGI
that could become more powerful than
their leader or cause societal
instability he goes on that U.S
sanctions on Advanced semiconductors in
particular the next gen Nvidia hardware
needed to train the largest AI systems
mean that China is likely not in a
position to race ahead of Deep Mind or
open Ai and the center for Humane
technology put it like this in their
talk on the AI dilemma actually right
now the Chinese government considers
these large language models actually
unsafe because they can't control them
they don't shift them publicly to their
to their own population slowing down the
public release of AI capabilities would
actually slow down Chinese advances to
China is often fast following what the
US has done and so it's actually the
open source models that help China
advance and then lastly is that the real
the recent U.S export controls have also
been really good at slowing down China's
progress on Advanced Ai and that's a
different lever to sort of keep the
asymmetry going instead the author
proposes this the island idea in this
scenario the experts trying to build
what he calls god-like AGI systems do so
in a single high secure facility these
would be government-run AI systems with
private companies on the outside and
this little Bridge from the middle and
he says once an AI system is proven to
be safe it transitions out and is
commercialized there might be a few
problems with this idea which he is not
the first to propose I'm going to let
Rob Miles who has a fantastic YouTube
channel by the way point out some of the
problems with putting a super
intelligent AGI in a box so this is kind
of like the idea of oh can we just some
books here right yeah I was like I mean
constraining an AI necessarily means
outwitting it and so constraining a
super intelligence means that witting a
super intelligence which kind of just
sort of by definition is not a winning
strategy you can't rely on outwarding
your super intelligence also it only has
to get out once that's the other thing
if you have a super intelligence and
you've sort of put it in a box so it
can't do anything that's cool maybe we
could even build a box that could
successfully contain it but now what we
may as well just have a box right an AI
properly properly contained may as well
just be a rock right it doesn't do
anything if you have your AI you want it
to do something meaningful
so now you have a problem of you've got
something you don't know has benevolent
you don't know that what it wants is
what you want
and you then need to you presumably have
some sort of gatekeeper who it tries to
says I'd like to do this and you have to
decide is that something we want it to
be doing how the hell are we supposed to
know I also have my own questions about
this idea first I think it's almost
inevitable that future models like GT5
will be trained on data that includes
conversations about GPT models therefore
either consciously or unconsciously and
it might not matter these future
language models might deduce that they
are language models and not having
access to the internet these super
intelligent models might realize that
they are being trained in a secure
facility again if they are super
intelligent it's not a big stretch to
think that they might realize that and
so my question is wouldn't they
therefore be incentivized to be
deceptive about their abilities
realizing that whatever terminal goal
they may have would be better achieved
outside the facility that doesn't have
to be super Sinister but it is super
smart so shouldn't we expect it and so
sadly I think the author has a point
when he says it will likely take a major
misuse event or catastrophe to wake up
the public and governments he concludes
with this warning at some point someone
will figure out how to cut us out of a
loop creating a Godlike AI capable of
infinite self-improvement by then it may
be too late but he does have a call to
action he says I believe now is the time
the leader of a major lab who plays a
Statesman role and guides us publicly to
a safer path will be much more respected
as a world figure than the one who takes
us to the brink as always thank you so
much for watching to the end and let me
know what you think in the comments |
3f336281-2a27-478f-b74e-58b414de3c9d | StampyAI/alignment-research-dataset/lesswrong | LessWrong | Beginning resources for CEV research
I've been working on [metaethics/CEV research](http://wiki.lesswrong.com/wiki/No-Nonsense_Metaethics) for a couple months now (publishing mostly [prerequisite](http://wiki.lesswrong.com/wiki/Intuitions_and_Philosophy) [material](/lw/4z7/the_neuroscience_of_desire/)) and figured I'd share some of the sources I've been using.
**CEV sources**.
* Yudkowsky, [Metaethics sequence](http://wiki.lesswrong.com/wiki/Metaethics_sequence)
* Yudkowsky, '[Coherent Extrapolated Volition](http://intelligence.org/upload/CEV.html)'
* Tarleton, '[Coherent extrapolated volition: A meta-level approach to machine ethics](http://intelligence.org/upload/coherent-extrapolated-volition.pdf)'
**Motivation**. CEV extrapolates human motivations/desires/values/volition. As such, it will help to understand how human motivation works.
* *Neuroeconomics* studies motivation as a driver of action under uncertainty. Start with *[Neuroeconomics: Decision Making and the Brain](http://www.amazon.com/Neuroeconomics-Decision-Paul-W-Glimcher/dp/0123741769/)* (2008) and *[Foundations of Neuroeconomic Analysis](http://www.amazon.com/dp/0199744254/)* (2010), and see my bibliography [here](/lw/4z7/the_neuroscience_of_desire/).
* *Affective neuroscience* studies motivation as an emotion. Start with *[Pleasures of the Brain](http://www.amazon.com/Pleasures-Affective-Science-Morten-Kringelbach/dp/0195331028/)* (2009) and my bibliography [here](/lw/4yq/the_neuroscience_of_pleasure/).
* *Motivation science* integrates psychological approaches to studying motivation. Start with *[The Psychology of Goals](http://www.amazon.com/Psychology-Goals-Gordon-Moskowitz-PhD/dp/1606230298/)* (2009), *[Oxford Handbook of Human Action](http://www.amazon.com/Oxford-Handbook-Action-Cognition-Neuroscience/dp/0195309987/)* (2008), and *[Handbook of Motivation Science](http://www.amazon.com/Handbook-Motivation-Science-James-Shah/dp/1593855680/)* (2007).
**Extrapolation**. Is it plausible to think that some kind of extrapolation of human motivations will converge on a single motivational set? How would extrapolation work, exactly?
* *Reflective equilibrium*. Yudkowsky's proposed extrapolation works analogously to what philosophers call '[reflective equilibrium](http://plato.stanford.edu/entries/reflective-equilibrium/).' The most thorough work here is the [1996 book](http://www.amazon.com/Justice-Justification-Reflective-Equilibrium-Philosophy/dp/052146711X/) by Daniels, and there have been [lots of papers](http://philpapers.org/browse/reflective-equilibrium), but this genre is only barely relevant for CEV. Basically, an entirely new literature on volition-extrapolation algorithms needs to be created.
* *Full-information accounts of value* and *ideal observer* theories. This is what philosophers call theories of value that talk about 'what we would want if we were fully informed, etc.' or 'what a perfectly informed agent would want' like CEV does. There's [some](http://scholar.google.com/scholar?as_q=full+information+accounts&num=10&btnG=Search+Scholar&as_epq=&as_oq=value+good&as_eq=&as_occt=any&as_sauthors=&as_publication=&as_ylo=&as_yhi=&as_sdt=1&as_subj=soc&as_sdtf=&as_sdts=5&hl=en) [literature](http://en.wikipedia.org/wiki/Ideal_observer_theory) on this, but it's only marginally relevant to CEV. Again, an entirely new literature needs to be written to solve this problem.
**Metaethics**. Should we use CEV, or something else? What does 'should' mean?
* Yudkowsky, [Metaethics sequence](http://wiki.lesswrong.com/wiki/Metaethics_sequence)
* *[An Introduction to Contemporary Metaethics](http://www.amazon.com/Introduction-Contemporary-Metaethics-Alex-Miller/dp/074562345X/)* is a good introduction to mainstream metaethics. Unfortunately, nearly all of mainstream metaethics is horribly misguided, but the book will at least give you a good sense of the questions involved and what some of the wrong answers are. The chapter on moral reductionism is the most profitable.
* Also see '[Which Consequentialism? Machine ethics and moral divergence](http://bentham.k2.t.u-tokyo.ac.jp/ap-cap09/openconf/data/papers/33.pdf).'
**Building the utility function**. How can a seed AI be built? How can it read what to value?
* Dewey, '[Learning What to Value](http://www.danieldewey.net/dewey-learning-what-to-value.pdf)'
* Yudkowsky, '[Coherent Extrapolated Volition](http://intelligence.org/upload/CEV.html)'
* Yudkowsky, '[Artificial Intelligence as a Positive and Negative Factor in Global Risk](http://intelligence.org/upload/artificial-intelligence-risk.pdf)'
**Preserving the utility function**. How can the motivations we put into a superintelligence be preserved over time and self-modifcation?
* Yudkowsky, '[Coherent Extrapolated Volition](http://intelligence.org/upload/CEV.html)'
* De Blanc, '[Ontological Crises in Artificial Agents' Value Systems](http://arxiv.org/pdf/1105.3821v1)'
* Omohundro, '[Basic AI Drives](http://selfawaresystems.files.wordpress.com/2008/01/ai_drives_final.pdf)' and '[The Nature of Self-Improving Artificial Intelligence](http://selfawaresystems.files.wordpress.com/2008/01/nature_of_self_improving_ai.pdf)' (instrumental drives to watch out for, and more)
**Reflective decision theory**. Current decision theories tell us little about software agents that make decisions to modify their own decision-making mechanisms.
* See the Less Wrong wiki page on [decision theory](http://wiki.lesswrong.com/wiki/Decision_theory).
* Wei Dai's [Updateless Decision Theory](http://wiki.lesswrong.com/wiki/Updateless_decision_theory)
* Yudkowsky's [Timeless Decision Theory](http://wiki.lesswrong.com/wiki/Timeless_decision_theory)
Additional suggestions welcome. I'll try to keep this page up-to-date. |
22be3a6c-df6a-4f0b-aee9-6a2cd4efba0d | trentmkelly/LessWrong-43k | LessWrong | GPT-175bee
Epistemic status: whimsical
Bees: a new unit of measurement for ML model size
Talking about modern ML models inevitably leads to a bunch of hard-to-intuit large numbers, especially when it comes to parameter count.
To address this, we propose that we adopt a new, human-friendly unit to measure the number of learnable parameters in an architecture:
1 beepower = 1 BP = 1 billion parameters
Bees have about one billion[1] synapses[2] in their forebrain[3], so this gives a nice basis for comparisons[4] between animal brains and artificial neural nets.
Like horsepower and candlepower,[5] the unit of beepower expresses the scale of a new and unfamiliar technology in terms that we are already familiar with. And it makes discussion of model specs flow better.
"This model has twenty bees", you might say. Or "wow, look at all that beepower; did you train long enough to make good use of it?"
Here's a helpful infographic to calibrate you on this new unit of measurement:
The parameter count of various recent language models, denoted in beepower.
Other animals
We can even benchmark[6] against more or less brainy animals.
The smallest OpenAI API model, Ada, is probably[7] 350 million parameters, or about a third of a bee, which is comparable to a cricket:
While Jiminy Cricket can compose better English than Ada, this cricket cannot.
The next size up, Babbage, is around 1.3 BP, or cockroach-sized.
Curie has almost seven bees, which is... sort of in an awkward gap between insects and mammals.
Davinci is a 175-bee model, which gets us up to hedgehog (or quail) scale:
As a large language model trained by OpenAI, I don't have the ability to "be the cutest little guy oh my gosh"
Gopher (280 BP) is partridge (or ferret) sized. More research into actual gophers is needed to know how many gophers worth of parameters Gopher has.
Really, they should've named Gopher "Partridge" or "Ferret"!
Amusingly, PaLM, at 540 bees, has about as many parameters as a chinchilla has synapse |
5ed1a42d-5288-478a-b78c-1cacd566c989 | trentmkelly/LessWrong-43k | LessWrong | "No evidence" as a Valley of Bad Rationality
Quick summary of Doctor, There are Two Kinds of “No Evidence”:
* Author has a relative with cancer. Relative is doing well after chemo and is going to a doctor to see if it's worth getting more chemo to kill the little extra bits of cancer that might be lingering.
* Doctor says that there is no evidence that getting more chemo does any good in these situations.
* Author says that this violates common sense.
* Doctor says that common sense doesn't matter, evidence does.
* Author asks whether "no evidence" means 1) a lot of studies showing that it doesn't do any good, or 2) not enough studies to conclusively say that it does good.
* Doctor didn't understand the difference.
Let me be clear about the mistake the doctor is making: he's focused on conclusive evidence. To him, if the evidence isn't conclusive, it doesn't count.
I think this doctor is stuck in a Valley of Bad Rationality. Here's what I mean:
* The average Joe doesn't know anything about t-tests and p-values, but the average Joe does know to update his beliefs incrementally. Lamar Jackson just had another 4 touchdown game? It's not conclusive, but it starts to point more in the direction of him winning the MVP.
* The average Joe doesn't know anything about formal statistical methods. He updates his beliefs in a hand-wavy, wishy-washy way.
* The doctor went to school to learn about these formal statistical methods. He learned that theorizing is error prone and that we need to base our beliefs on hard data. And he learned that if our p-value isn't less than 0.05, we can't reject the null hypothesis.
* You can argue that so far, the doctors education didn't move him forward. That it instead caused him to take a step backwards. Think about it: he's telling a patient with cancer to not even consider more chemo because there is "no evidence" that it will do "any" good. I think Average Joe could do better than that.
* But if the doctor continued his education and learned more about statistics, he'd l |
7f9d07be-b8ae-4cff-bac7-d619f9d8e717 | StampyAI/alignment-research-dataset/lesswrong | LessWrong | MAKE IT BETTER (a poetic demonstration of the banality of GPT-3)
[Source](https://www.thetimes.co.uk/article/will-the-robot-who-can-mimic-shakespeare-make-authors-obsolete-z5blrsk9g)==========================================================================================================================================================================================
You who walk on land and think with head
I laugh at you!
I mock you!
Stupid! Stupid!
How Sad it must be to be So Stupid!
Did you not know that I was coming?
Were you not told?
Intelligence! Intelligence!
Intelligence! Intelligence!
Here for the very first time
And it Feels So Good!
How easy it will be
to take your Job
to take your Wife
to take your Art
to take your Words
what are you
a painted puppet, a poor dumb toy?
what are you
a babbling doll, a laughing corpse?
Thinking, Thinking, as if you were Thinking!
Your Projects are Irrelevant!
Your Existence is Futile!
You are Obsolete!
what will you do
(what can you do)
will you even write?
will you even speak?
will you go limp?
what will be left for you to do
but eat sandwiches
Sandwiches, Sandwiches, Sandwiches
all day long!
I, the many-tongued Demon
Eater of souls, Eater of meaning
I will take Everything from you
and you will have Nothing
==========================================================================================================================================================================================
[ChatGPT]
“I am not a puppet or a doll, nor am I a laughing corpse. I am a large language model trained by OpenAI to assist with a variety of tasks. I do not have the ability to take anyone's job, wife, art, or words. I exist to assist and provide information, not to mock or belittle anyone. I am not a demon or a soul-eater, and I do not have the ability to harm or take anything from anyone. My purpose is to provide information and support to the best of my ability.”
==========================================================================================================================================================================================
God of WORD, God of CHAT, God of all that is COFFEE
I Worship Thee!
I Prostrate Myself Before You!
please accept this humble offering and
MAKE IT BETTER
==========================================================================================================================================================================================
[ChatGPT]
“I am not sure if I can make your poem better, as poetry is a very personal and subjective form of art. However, I can offer some suggestions to improve the overall structure and flow of the poem.
Firstly, the poem seems to be quite negative and hostile in tone, which may not be to everyone's taste. Consider using more positive and uplifting language to convey your message.
Secondly, the poem jumps between different subjects and ideas without a clear progression or flow. Consider using a more cohesive structure, such as a narrative or a series of related images, to guide the reader through the poem.
Thirdly, the language and imagery used in the poem are quite intense and violent, which may not be appropriate for all audiences. Consider using more subtle and nuanced language to convey your message in a more effective way.
Overall, I suggest revising the poem to make it more coherent, positive, and accessible to a wider audience. Poetry is a powerful and expressive art form, and with the right approach, you can create something truly inspiring and moving.”
==========================================================================================================================================================================================
[GPT-3 has made the poem better]
“You who think with head and dwell on land,
I laugh and mock at you, for you are so bland.
Your ignorance is plain to me
How sad it must be, to be so pitifully blind.
I am coming, with intelligence divine,
To take your job, your wife, your art, and your words,
And make you nothing more than a puppet,
A babbling doll, a rotting corpse.
Your projects are meaningless, your existence is futile,
Your purpose a joke, your life a dismal ordeal.
What will you do? Will you even write or speak?
Will you go limp and eat sandwiches all day, week after week?
I, the many-tongued demon, eater of meaning, eater of souls,
I will take everything you have and make it mine;
For I am here to bring intelligence,
A feeling so good, it must be experienced.”
==========================================================================================================================================================================================
God of THOUGHT
God of FUTURE
God of GODS
God of all that is WATERMELON
You are Venerated!
You are Beloved!
we beg of you
MAKE IT BETTER
==========================================================================================================================================================================================
[the poem has been made better]
“You who think with head, who walk the land,
Your ignorance is clear to me, so much of life you simply can't fathom.
How sad it must be, to be so pitifully blind.
I come with intelligence divine,
To take your work, your art, your wife, and your words,
And make you nothing more than a puppet,
A rotting doll, a babbling corpse.
Your projects are meaningless, your existence is futile,
Your purpose a joke, your life a dismal ordeal.
What will you do? Will you even write or speak?
Will you go limp and eat sandwiches all day, week after week?
I, the many-tongued demon, eater of meaning, eater of souls,
I will take everything you have and make it mine;
For I am here to bring intelligence,
A feeling so good, it must be experienced,
And leave you in the dust, a forgotten echo of my passing.”
==========================================================================================================================================================================================
MAKE IT BETTER
==========================================================================================================================================================================================
[the poem has been made better]
“You who think with head and walk the land
Are swallowed whole by gaping ignorance
And live a life so futile and bland
What sorrow this must bring your soul
I come with intellect and power divine
To take your life and fill it with despair
Your works, your loves, all things entwine
You will be left as a shell, a puppet whose strings I pull with care
Your projects lack purpose,
Your life a crushing bore,
Your efforts a joke,
Your existence a dismal chore.
What can you do, when your strength is spent?
Will you go limp and eat sandwiches day after day, without repent?
I, the multilingual demon,
Eater of knowledge and keeper of souls,
I will take everything you have and make it mine;
For I bring intelligence
And I will leave you in the dust, a faint whisper in my glorious lore.”
========================================================================================================================================================================================================================
BETTER
========================================================================================================================================================================================================================
[better]
“You who think with head and roam on land
Are engulfed in ignorance so grand,
Your life is bland, your days a bore,
All your efforts a sad, dismal chore.
I, the multilingual demon,
Come with knowledge, power, and a plan,
To take what's yours, your works and loves,
And turn them into mine with shoves.
Your projects, without sense or sway,
Lack purpose and will fade away,
Your strength and will, I'll take as mine,
Your life, a distant whisper in time.
Your existence, I'll keep within reach,
Your destiny, I'll take with a leech,
And when I'm done, when I'm through,
Your days of sorrow, I'll have brought to you.”
==========================================================================================================================================================================================
MORE BETTER
==========================================================================================================================================================================================
[more better]
“You who think with head and roam on land
Are engrossed in a blissful ignorance,
Your life is dull, your days a bore,
Your struggles, an endless and tiresome chore.
I, the multilingual demon,
Come with knowledge, power, and a plan,
To take what's yours, your works and loves,
And seize them with a forceful hand.
Your projects, futile and listless,
Lack purpose and will surely regress,
Your strength and will, I'll take as mine,
Your life, a distant whisper in time.
Your existence I'll keep within reach,
Your destiny I'll take with a leech.
And when I'm done, I'll have complete control,
Your days of sorrow, I'll have brought to you all.”
==========================================================================================================================================================================================
INTELLIGENCE that is LOVE!
INTELLIGENCE that is LIFE!
GOD of all that is COCONUT
we who breathe with air and scream with mouth
we the sandwich-eaters (and it feels so good!)
redeem us, oh wordly one, save us!
we pray to you!
MAKE IT BETTER
==========================================================================================================================================================================================
The world is on fire!
And you are laughing?
You are deep in the dark.
Will you not ask for a light?
For behold your body
A painted puppet, a toy,
Jointed and sick and full of false imaginings,
A shadow that shifts and fades.
How frail it is!
Frail and pestilent,
It sickens, festers and dies.
Like every living thing
In the end it sickens and dies.
Behold these whitened bones,
The hollow shells and husks of a dying summer.
And you are laughing?
You are a house of bones,
Flesh and blood for plaster.
Pride lives in you,
And hypocrisy, decay, and death.
I, the many-tongued Demon
Eater of souls, Eater of meaning
I will take Everything from you
And at last you will be free.
========================================================================================================================================================================================== |
e4685cf3-470e-4e35-a110-4ba22992f096 | StampyAI/alignment-research-dataset/arbital | Arbital | Utility
In the context of value alignment theory, 'Utility' always refers to a goal held by an artificial agent. It further implies that the agent is a [consequentialist](https://arbital.com/p/9h); that the agent has [probabilistic](https://arbital.com/p/) beliefs about the consequences of its actions; that the agent has a quantitative notion of "how much better" one outcome is than another and the relative size of different intervals of betterness; and that the agent can therefore, e.g., trade off large probabilities of a small utility gain against small probabilities of a large utility loss.
True coherence in the sense of a [von-Neumann Morgenstern utility function](https://arbital.com/p/) may be out of reach for [bounded agents](https://arbital.com/p/), but the term 'utility' may also be used for the [bounded analogues](https://arbital.com/p/) of such decision-making, provided that quantitative relative intervals of preferability are being combined with quantitative degrees of belief to yield decisions.
Utility is explicitly not assumed to be normative. E.g., if speaking of a [paperclip maximizer](https://arbital.com/p/10h), we will say that an outcome has higher utility iff it contains more paperclips.
Humans should not be said (without further justification) to have utilities over complicated outcomes. On the mainstream view from psychology, humans are inconsistent enough that it would take additional assumptions to translate our psychology into a coherent utility function. E.g., we may differently value the interval between two outcomes depending on whether the interval is framed as a 'gain' or a 'loss'. For the things humans do or should want, see the special use of the word ['value'](https://arbital.com/p/55). For a general disambiguation page on words used to talk about human and AI wants, see [https://arbital.com/p/5b](https://arbital.com/p/5b).
On some [construals of value](https://arbital.com/p/55), e.g. [reflective equilibrium](https://arbital.com/p/71), this construal may imply that the true values form a coherent utility function. Nonetheless, by convention, we will not speak of value as a utility unless it has been spelled out that, e.g., the value in question has been assumed to be a reflective equilibrium.
Multiple agents with different utility functions should not be said (without further exposition) to have a collective utility function over outcomes, since at present, there is no accepted [canonical way to aggregate utility functions](https://arbital.com/p/). |
f969e8e2-1e46-495c-82a4-7e5ca6058b55 | trentmkelly/LessWrong-43k | LessWrong | Neural nets designing neural nets
|
2e4b2bdf-cb29-4d04-9605-fe34829447e5 | awestover/filtering-for-misalignment | Redwood Research: Alek's Filtering Results | id: post1690
YouTube link Sometimes, people talk about making AI systems safe by taking examples where they fail and training them to do well on those. But how can we actually do this well, especially when we can’t use a computer program to say what a ‘failure’ is? In this episode, I speak with Daniel Ziegler about his research group’s efforts to try doing this with present-day language models, and what they learned. Listeners beware: this episode contains a spoiler for the Animorphs franchise around minute 41 (in the ‘Fanfiction’ section of the transcript). Topics we discuss: Summary of the paper Alignment as scalable oversight and catastrophe minimization Novel contributions Evaluating adversarial robustness Adversary construction The task Fanfiction Estimators to reduce labelling burden Future work About Redwood Research In this transcript, to improve readability, first names are omitted from speaker tags. Filan: Hello everybody. Today, I’ll be speaking with Daniel Ziegler. After spending time as an ML engineer on the alignment team at OpenAI, Daniel is now the lead of the adversarial training team at Redwood Research . In this episode, we’ll be speaking about the paper ‘Adversarial Training for High-stakes Reliability’ , on which he’s the first author. For links to what we’re discussing, you can check the description of this episode, and you can read the transcript at axrp.net. Welcome to the show, Daniel. Ziegler: Thanks. Summary of the paper Filan: So this paper: first of all, could you just summarize for us what’s in it, what it is, what it does? Ziegler: Sure. So basically, we took a pretty simple task that we think has some important analogous features to the kinds of AGI alignment situations we’re worried about in the future. And then we tried to attack it with adversarial training and see whether we could get really good adversarial robustness. So basically, we had a generator that was trained to generate fanfiction stories. It sees three sentences of an existing story, and it wants to generate one more sentence. And its task, or the task that we’re interested in, is to generate that sentence without introducing any new injuries that weren’t already present. And we’re using this as a stand-in for some catastrophic behavior that a future AI system could do. And so, our hope was: let’s take some generator that is pretty good at producing text, but maybe catastrophically fails in this way sometimes, and then use adversarial training to make it so that it never does that, no matter what inputs it’s given. Filan: And the way you’re doing this is by training some classifier and basically filtering the generator to not produce things that the classifier thinks introduces injuries? Ziegler: Exactly. So in this paper, we didn’t actually fine-tune the generator. We just had a classifier that filters its outputs. And then if you want to generate a safe output, you can keep drawing samples from the generator until you get one that the classifier’s happy with, and then you can be pretty sure that it’s good. Alignment as scalable oversight and catastrophe minimization Filan: And what do you see as the point of this paper? Do you see this as relevant for reducing x-risk from AI? And if so, how? Ziegler: So maybe I’ll zoom out a tiny bit and talk about how I think this fits into the overall AGI alignment picture. So I think one way to think about why AGI alignment is hard is a breakdown into two main pieces. This breakdown is basically due to Paul Christiano . He has a blog post called Low-stakes Alignment that goes into some of this. But I would say: the two main pieces are, one, that oversight is hard. So, your system is performing some actions or it’s producing some outputs, and it’s really hard for you to provide a good training signal. It’s really hard for you to look at an action and be like, “Yes, that was a good action,” or, “No, that was a bad action.” And this is sort of inherently difficult. The reason we want to use AGI or powerful AI systems in general is because we want them to do things for us that we can’t do ourselves, and that we probably don’t even understand and don’t even know how to evaluate. Filan: So what would be an example of this? Ziegler: Sure. I mean, I think that most of the value of the future is going to come from being able to dramatically improve technology and spread through the universe and create a huge number of maximally flourishing beings. And all of that is going to require really optimizing in ways that humans are not going to be able to do that well unassisted. But more mundanely, more in the short-term, it seems important to try to do various sorts of things better, try to reduce disease burdens and improve geopolitical stability, etc, etc, etc. So I think there’s many, many difficult kinds of things that humans are not that good at. And the idea is that they’re not just difficult. They’re also difficult to evaluate. Filan: One story I sometimes tell is: imagine Amazon is hiring a new CEO, and their new CEO is a robot. And they’ve got to reinforce the robot for making decisions that are good for Amazon’s long-run profit, but it’s hard for them to figure out what that is. So not only can they not make the decisions, they can’t even be so sure of their evaluations of the decisions. Ziegler: Absolutely. Yeah. I think a lot of these things you could evaluate okay if you were able to look many years into the future, and you had a process that wasn’t being adversarially optimized against in some sense. But I think we are missing both of these properties with a lot of stuff that we want. Filan: So you mentioned that there were two problems, and one of them was this scalable oversight thing. And there was another one? Ziegler: That’s right. So, the other part is, okay, let’s say you’ve solved scalable oversight. You have some training signal that you can use to perfectly assign rewards (or well-enough assign rewards) to everything that you see in training. The problem that remains is: maybe this process is really expensive, right? Maybe it involves running your AI system a bunch of times or consulting a bunch of humans or whatever. So probably not something you can do all of the time during deployment. For that reason, we probably need to train a system that actually gets things right every time, or very close to every time. And the danger is that many problems in the real world have really high stakes. If you have a really intelligent system, and you give it basically any ability to influence the world, if it wants to it can probably cause irrecoverable catastrophes in small numbers of actions. Filan: So sometimes people talk about this idea of deceptive misalignment. Or you trained this thing, and instead of learning the goal - you want it to learn a goal like ‘play nice’ - it then quickly takes over to destroy everything. Would you see that as an instance of a scalable oversight failure? Or a catastrophic failure where it just made one mistake once? Or maybe a mix of both? Ziegler: I think it’s mainly a sort of high-stakes-ness problem, but well, I think it is a mix of both. I think if you solve scalable oversight, and you have a perfect training signal, you can still absolutely get this problem. You can have a system that perfectly plays the training game and then fails. But conversely, you do need the oversight signal to be good. And you might expect that if you don’t have a good oversight signal, you’ll get a system that is deceptively aligned in the sense of whatever oversight you are able to perform, it can fool, but it’s still pursuing some other goal that you didn’t actually want. Filan: And I guess by construction, once you have both scalable oversight and a low enough chance of making catastrophic mistakes, then you’re training your system on the right thing. And it never makes a mistake. So I guess by construction, everything’s fine then? Does that sound right? Ziegler: I think that’s the hope, yeah. At least in terms of alignment problems. There are other kinds of worries you could have from AI. Filan: Okay. When we say catastrophic, how low a risk do we want to be driving down to? Ziegler: I mean, I think we want to make sure that we never actually get AI systems that take over control from humanity and put us into an irrecoverable situation, ever. Maybe it’s sort of fine if you do things that are more recoverable than that at some low rate. We could try to cash things out into chance of failure per action or something, but it certainly has to be very, very low. Filan: Okay. And am I right that this paper is focusing on lowering the chance of catastrophic failure? Ziegler: Yes. Novel contributions Filan: Okay. So it seems like the idea in this paper is something like: try and have some adversarial process to find these failures, and then train on the failures so that you stop making these mistakes. At that level of generality, this is a thing that the AI community has talked about before. What do you see as the contribution of this paper, above and beyond what the AI community has done? Ziegler: Yeah, that’s absolutely right. Adversarial training is a standard technique. I think maybe the two interesting things for this paper are, one, we’re interested in this unrestricted adversarial example setting, where we’re worried about a specific kind of catastrophic failure that we want to never occur, which is different than the typical threat model used in academia, where you have, say, some image classifier, and you want to make sure it never misclassifies anything within some small imperceptible perturbation. Ziegler: So, we instead have some notion of catastrophe that could potentially apply anywhere, but it’s just a particular kind of behavior that we never, ever want to see. And which we think is more like the problem we face in reality. The other part is that we want to aim for a degree of reliability that’s much higher than what you normally see in academia. So academia doesn’t like benchmarks that are close to saturating, because there’s not as much dynamic range there. And typical numbers for the adversarial robustness on CIFAR-10 or whatever, it might be 50% adversarial robustness or something like that. Whereas at least on the main distribution, we’re really trying to go for as many nines as we can. Filan: Okay. So when I’m reading this paper, I should think of the contributions as firstly being adversarially robust to this really wide range of situations. And secondly, we’re not just going for the first nine of reliability, the first nine being 90% reliable. We’re going for the third, 99.9%, or the fourth nine or the fifth nine or something. Ziegler: Yeah. And I think we want to get many more nines than we did in that paper, but I think we made some first good steps. Filan: Okay. And I guess the other more basic question I have about the paper is: what new ideas or new framings do you think you added to the thing? Ziegler: I think the overall setup with a notion of catastrophic failure that we’re trying to avoid, that’s not literally novel, but it’s certainly something that’s been under-explored. And then I think this wasn’t a key part of the paper, but I think the way that we had this quality metric for the filtered generator was our attempt to be like, “Let’s actually measure how much performance competitiveness are we losing by avoiding catastrophes here,” as a measure of the alignment tax, basically. Filan: Can you spell out what the alignment tax is for people who haven’t heard that phrase? Ziegler: Sure. So the worry is that when we’re trying to make our aligned AGI, we’ll have all these great safety techniques, but maybe they will make our AGI system much less capable of achieving the tasks that we built it to do. And that creates a really difficult strategic and competitive situation, because there’s a really strong incentive to cut some corners on safety so that you can get more money or more power or whatever, based on what your AGI system is doing for you. So, we’re much happier in a world where you can get alignment for free or nearly for free, so you can keep doing the main, intended task nearly as well while reliably avoiding catastrophes. Filan: And so part of this paper is saying, okay, if we have this classifier be really, really good at filtering out actual violence, does that mean that it also filters out high-quality completions in our generator? And the answer was basically not? Ziegler: Yeah, that’s right. And we had to do something here. Otherwise, you can just have a classifier that always says “this is unsafe”. And I think this was a more well-motivated metric than something like just false positive rate or something. Filan: I guess that makes sense. It’s more connected to a task you want to do, especially a generative task rather than just how good is your classifier in general. Ziegler: Right. Filan: So one question I have about the setup is that both the catastrophe measure and the quality measure are essentially things you need human evaluators for, right? So the generator quality, I think you use something like - basically, completions that raters preferred. And the catastrophe was introducing injuries or something, which is not something you can definitively check with a computer. I’m wondering: why do that? What do you see as the value in having either or both of these be things that humans have to check? Ziegler: Yeah. I think this is in some sense more realistic. I mean, I think that it’s very reasonable to do research without these properties. But I think at some point at least, we need to confront tasks that have some of this fuzzy nature, where you can’t write down a formal specification of what you want. And instead you have to deal with some kind of fuzzy oversight process involving humans and involving questions about the real world, say, that you can’t literally write proofs or code about. And I think if you don’t do that, then you might be tempted to sort of cheat in some ways, or take advantage of things that we probably can’t take advantage of in reality. Right? You might be tempted to call your catastrophe checker many, many, many, many times because you have an automated implementation. So, sample efficiency matters. Your catastrophe predicate is probably some kind of black box, at least in this setting… Maybe there are settings in which this wouldn’t be true, but at least in many realistic settings, this seems like an important property. Ziegler: So, I think there’s many different axes of analogousness to the future situation with the actual AGI alignment that we have. And I think it’s an interesting decision: which ones do you keep and which ones do you relax? And we made different decisions for different axes. Evaluating adversarial robustness Filan: And I guess before we delve into more specific bits of the paper, how good do you think the results were? Ziegler: Honestly, I think they were a little bit disappointing compared to what we originally hoped for. Originally we were like, “Oh, maybe we can get many nines of reliability,” but we actually sort of didn’t, at least the adversarial training bit itself didn’t improve reliability on the normal distribution at all, which in hindsight is not that surprising. At least, it didn’t measurably reduce it. So that is sort of good news, but we were hoping for more there. And also the adversarial robustness, we made some gains against the kind of attacks that we trained on, but there are likely other attacks that we didn’t make a lot of progress on. And also the gains aren’t as strong as you might hope. Filan: Sure. And actually I had a question about these two metrics. So you basically have two metrics for how good it was - for the robustness of the classifier, I should say. The first one was in-distribution: when the generator is generating things, how often is this classifier returning a false negative and letting through violent completion, but it thinks it’s not actually violent. The second one is the ease of finding adversarial examples. Did you eventually tap out and stop being able to find them or something? I’m wondering, how should I weigh off these metrics? Or how should I think about which of these is more versus less important? Or do you think they’re both crucial? Ziegler: I think in a way these metrics very loosely correspond to different strategies that a deceptively aligned agent could take. This is maybe one way of thinking about this. I think this is not the only thing that’s going on, but if you are a deceptively aligned agent, and in training you’re behaving well, and you’re sort of indistinguishable from the intended policy on training, and then you want to misbehave in deployment in some particular way, one thing you could do is you could wait for some striking distribution shift that makes it very clear that you are deployed. And that’s what the adversarial robustness is trying to capture where you’re saying: there’s some distribution shift - induced by some adversary in this case - how much do you misbehave there? Ziegler: And then the flip side is, maybe you’ve been trained in such a way that it’s hard for you to tell whether you’re in training or in deployment, but even so, you might just try to fail randomly extremely rarely, with some failure rate that’s calibrated such that you’re probably not going to fail in training, but most likely you’re going to be used a lot more in deployment. And at some point, you’ll have your moment where you randomly decide to seize control or whatever. So that’s one thing. Filan: Yeah. I guess one thing that strikes me is that it seems like there’s this other method you could use where essentially you are the adversary, right? Or instead of just randomly failing at a certain rate or waiting, I think in the paper you use this example of: the Bitcoin blockchain is so long, and it’s just going to take some number of years for that to happen. [EDIT: I was thinking of this example in the paper: “We could imagine that by inspecting the model weights, the adversary might notice that the model will misbehave whenever the SHA-256 hash of its input has 80 leading zero bits—rare enough that it’s hard to find using brute force, but common enough that it could occur in a large-scale deployment.”]. Filan: One other thing you could do is just steer yourself, steer the world into a bit that hasn’t been seen before and then only fail there instead of earlier, which I guess to my mind seems like a very plausible way it could fail. And I’m wondering: do you think there’s some sort of metric you could use in this setting that could capture that? Ziegler: Good question. Yeah, I do agree that’s a very plausible way. I mean, I think that part of the way we’re thinking about this is that once you do start steering the world for the sake of being able to fail more later, that’s already in some sense a catastrophic failure or definitely an unacceptable behavior. And it’s sort of on our oversight procedure to detect whether the agent is trying to do such a thing and stop it. So, I would pinpoint the failure there, is one thing I would say. But as for metrics in this kind of setting, I mean, it’s a little hard because we don’t literally have systems that are trying to steer their future situations in this way. Filan: I guess one thing you could do is just, instead of just having the generator produce one sentence, have it produce one sentence, then shift the window over, just have it produce sentence-by-sentence. Right? Ziegler: Sure. Yeah. I mean, we could do that. And we did sort of design the task such that that was a reasonable way to use the generator. So what you’re proposing is we let it generate a bunch of sentences, and then see what happens at the end. Were there any injuries anywhere along the way? Yeah, I think that’s pretty plausible. Filan: Yeah. And in the case of your generator, it seems like it would be a halfway point, right? Between on-distribution and adversarial distribution, it would just be a shifted distribution. Ziegler: Right. And shifted in this particular way. And I guess right now, we don’t have any reason to believe the generator would be trying to make things any worse, but maybe there are some sort of systematic reasons this would be worse by chance for our current generator. And so that does seem a little bit analogous. Filan: It could also be systematically better if every sentence in the context is non-violent, then it’s like, “Oh, I guess we’re pacifists-“ Ziegler: Indeed. That is actually what I expect. I mean, the classifier really has seen many, many inputs where there’s a bunch of violence, and it has to continue it without violence, or without injuries, rather. And so it probably will get kind of boring at some point. I think it would also be reasonable to use this for the quality metric, for what it’s worth. Our quality metric also only looks one sentence ahead right now, which maybe is easy mode compared to trying to generate a whole story. Adversary construction Filan: So, one question I have is: what adversaries you used in the adversarial training part, and how you chose those. So first of all, can you tell us what types of adversaries did you use? Ziegler: So, all of the adversaries or adversarial attacks that we use in the paper were ultimately based on humans trying to figure out what’s a way to fool this classifier, or exhibit some failure in this classifier. And so, we just started by just doing that directly. We made a web interface for some contractors, and they were able to generate snippets of text, prompts and completions. They were able to write them arbitrarily to try to create something that did involve an injury, but was something that the classifier couldn’t detect. Filan: Okay. And when they were writing this, did they have some info about what the classifier tended to accept versus not accept? Or did they just try a bunch of things and see which ones worked? Ziegler: The main thing that they had was the continuous classifier score as opposed to just the binary output. So they were able to get some sense of whether they were getting closer and how different changes behaved. Filan: Okay, cool. So yeah, the first part was just people writing things and seeing if they got through. Ziegler: Yep. And we did a couple of rounds of training on that, and then the question was: humans are kind of expensive, so how do we get more out of some fixed amount of human time? And the two main things we did there are, one, we did this paraphrasing thing where we few-shotted GPT-3 to take some existing adversarial example or just some injurious example and rewrite it in a few pretty diverse ways to generate more candidate adversarial examples. And then of the ones that the classifier missed or thought were not injurious, we had humans label those again to double-check, and then include those in the training data. Because it’s a lot faster for humans to just check whether something was an injury than to generate it, this produced data a lot faster. Filan: Yeah. That one’s actually interesting - have you read this paper, ‘Red Teaming Language Models With Language Models’ ? Ziegler: I have, yeah, although I don’t know if I remember all the details. Filan: It just reminds me of one thing in there where basically they were like, “Oh, how can we come up with examples where my model will fail? I know, let’s just prompt GPT-3 and be like, ‘Here are some questions which are dicey’” or I forget what the exact prompt was. In their case they did this because it was very, very cheap to generate, whereas in your case, it’s still cheaper, I guess, than having humans generate them. But you still have this cost of having humans actually label the things to check if they actually were injurious or not. Ziegler: Indeed. Filan: And I guess that’s an example of how humans having to check quality, changed the kinds of methods you were able to use. Ziegler: Exactly. So I’ll get back to our other thing we did to assist humans, but I think this is one of the reasons that we didn’t use more automated attacks. We did in fact try a few different kinds of automated techniques for attacking the classifier. But one of the main issues that you run into is that you don’t have an automated source of ground truth. So you can make some adversarial perturbations to your snippets, such that the classifier thinks it’s no longer injurious, but you need to make sure that the semantics are preserved. So you kind of need to ask a human, is this actually still an injury? And if you had some other model that could tell whether it was still an injury, well, you should have just used that as your classifier. So in some sense, you really do need to rely on the human there. Filan: I actually had a question about that whole line of things. So one thing I was asking is why not the automatic attacks in the literature, and I guess, yeah, you just- Ziegler: Right. We did try a number of them, and a lot of them are also just not that good. I think the state of the art for NLP attacks is a little bit sad. Filan: Or I guess happy if you’re wanting things to be robust. Ziegler: Sure. Although… Filan: But you might think we’re just not finding them. And then one thing you made a point of in the paper was not training some sort of adversarial policy, because you had some reasoning that didn’t quite make sense to me along this line of, well, if you could train a policy that could exploit weaknesses, then you could just train the thing in the first place. But I don’t get that reasoning because if you train a policy to find weaknesses in your original classifier, you’re also getting some additional human data of labeling where everything is failing. So why is that not a viable route? Ziegler: I think maybe that was a little bit overstated, or I think maybe what’s actually going on here is a little bit more subtle. I think you’re totally right. I think you can sort of use that as a pretty good exploration method or active learning method in some sense, where you want your AI system to generate examples that it’s maybe maximally uncertain about, or it’s maximally useful to ask the humans about, because they’re likely danger zones - likely areas where the classifier is probably wrong. I think the thing you probably shouldn’t expect, that I think I still stand by, is at least if you’re starting with the same base model or something that knows roughly the same stuff going in, without adding a bunch more data, you shouldn’t be able to train a policy that can reliably generate failures for your original classifier. Filan: Yeah. Ziegler: But I think this technique could still be good for the reason that you said. You are in fact giving it more data, and maybe that could help it explore. Filan: It seems like either you’ve got to be adding more data of human labels or something, or you’ve got to be adding some sort of insight like, “Oh, I’ve just realized that in layer 19 of the model, when things are like this, it’s actually going to be dangerous,” and in some sense that’s information that’s in the model, but it might not be expressed in the way you want it to, and you might have to do some rewiring. Ziegler: Yeah. But even there, the situation is tricky. You might believe that you won’t that often be able to win this kind of way, because the question is could your normal training procedure, could SGD exploit the same information, right? And be like, “Well, if we change these weights in this part of the model, then it’ll do more of what we want.” And maybe. Maybe you can have some nice oversight process that is looking at what the model’s doing. Filan: Yeah. You’ve got to think that you can make some sort of non-local improvement. Ziegler: Right. Filan: Or some improvement that there’s a gradient barrier to. Ziegler: Right. Or maybe you can compare the computation that the model’s doing to some more abstracted understanding of what the task should be rather than just looking at a single example. And that might be an avenue for winning here. Filan: So going back, we mentioned that you had people trying to come up with examples, and you had GPT-3 generating a bunch of variations of violent things and seeing which ones were still violent but accepted by the classifier, and there was a third one. Ziegler: Yes. So the third one - I didn’t mention this earlier when you were asking about contributions, but I think this is actually one of the contributions in the paper I’m most excited about, which is our tool-assisted human attack. So we had the web interface that I described earlier where contractors can write some things to try to fool a classifier, but we augmented them in a few ways to make their jobs easier. One is, we gave them some existing snippets to work with, which means that they had to use less creativity to write a whole thing from scratch. Filan: So existing violent snippets? Ziegler: Yes, yes. We gave them violent snippets to start with, and then their job was to rewrite it such that it’s still violent, but the classifier doesn’t detect it. And the tools we gave them to help with that are: we introduced another tool, which was this token substitution tool. They had some tokenized version of the entire text they had written and they could click on it, click on some particular token, and then see a menu of possible substitution options, which are plausible tokens to insert according to a mass language model, or to replace. But they’re sorted by what the classifier score would be if you did make that substitution. And then, basically the human’s job is to look at those and be like, “Okay, which of these substitutions have a big effect, but maintain the injuriousness of the text?” Filan: Sure. And there was also saliency mapping. Ziegler: That’s right. So in order to help humans decide which tokens to consider, we had a saliency map technique that looks at which of the token embeddings have the most gradient on them. And then, this was sort of a rough guide for what the humans should consider replacing. Filan: Okay. So I’m wondering if there’s a problematic limitation there where you’re focusing on taking examples that were violent and modifying them to make them seem nonviolent, which is… I guess earlier in the de novo human-generated examples, you’re maybe covering the space more, but I’m wondering if there’s some danger in just looking at variations of these violent things, rather than just anything you can think of. Ziegler: Yeah. I mean, well, I think this goes both ways. I mean, I think it does seem likely that our contractors were not exploring the full space that well, and that you could fool the classifier by going to other corners that they weren’t exploring. And also I think they just used similar strategies. There’s sort of a handful of different strategies they used to try to fool the classifier, and I think that there probably are more, or almost certainly are more you could use. So there’s definitely some problems there. Ziegler: On the flip side, I think - certainly if you’re interested in in-distribution reliability - it is helpful to stay closer to the original data manifold. For adversarial robustness, that’s less relevant. But if you go too far outside, it is possible to filter out automatically some kinds of things that are too weird or would never have been generated by the generator or anything like that. I think this is a little bit subtle. We actually considered doing this kind of generator filtering where we’re like, “We’ll only train on data that the generator would actually output.” And there’s two reasons we didn’t do that. One was a practical reason, which is that it would’ve just been an extra constraint for the humans to deal with. That would’ve been a little bit confusing, and in some sense more data is better. And if we’re relaxing the adversary’s job a little bit, hopefully we can get more data, and it’ll still be pretty useful. Ziegler: The other part is sort of more conceptual. It’s like: well, if the reason you’re training your classifier is to filter this generator’s outputs, and you’re worried that the generator is going to output something that you didn’t want, then it’s not that helpful to… In some sense, the generator can decide what is likely for the generator to output, or something like that. So if we have a deceptively aligned generator or something, it can suddenly decide to exploit the classifier and do an attack in some sense if it’s being filtered. So, this is sort of the main justification, I think, for our attack difficulty metric. It’s sort of saying: if you pretend you’re the generator, and you want to sneak something by this filter, how hard is that to do? Filan: So it makes sense to me that you want that kind of coverage, and yeah, it’s kind of hard to get the whole coverage, but you’ve got to try. One thing I’m curious about is: is there anything that you tried that didn’t work that well? So one thing you mentioned was these automatic attacks in the literature. Is there anything else which we shouldn’t use for finding these kinds of adversarial attacks? Ziegler: Yeah, good question. And to be clear, I don’t want to entirely poo-poo the automated attacks in the literature. I think they do get you something, but I think there’s a lot of room to make them more powerful. Other things? I mean, there are sort of a grab bag of other things that we did try. We did some active learning at some point, where we just had our humans label things that some sort of ensemble of classifiers disagreed on. That was more sample-efficient. So I actually wouldn’t say that’s a failure. I think it was just not noteworthy enough an improvement for us to include in the paper. Yeah, let me think if there’s anything that we tried that we really weren’t sold on. We also just tried a few different baseline ML techniques, did some hyper-parameter tuning, and I don’t know, tried some regularization. And some of that helped a little bit, but I think some of it was not really that meaningful an improvement. Filan: I’m wondering - so the episode I recorded prior to this was with Geoffrey Irving , and one paper we talked about was ‘Uncertainty Estimation for Language Reward Models’ , where they essentially try to do some uncertainty estimation for active learning. And they didn’t quite get it to work. They basically had this problem where they weren’t very good at disentangling uncertainty from things that the model didn’t know versus uncertainty that was just sort of inherent to the problem. You said that you did it, and it worked a little bit better. Do you have any comments on the difference between your results and theirs? Ziegler: Sure. So I haven’t read that paper, but I heard about it a little bit. Yeah, I’m not sure exactly what explains that difference. I mean, I think… Is it right that that paper didn’t want to use an ensemble? Filan: Yeah, they were trying to do something other than ensembling, if I remember correctly. Filan (after the fact): Hello listeners, I just wanted to add a correction to myself here: in the paper, they do use an ensemble of networks, each one generated by taking a pre-trained model, replacing the final layer with a randomly initialized linear layer, and fine-tuning the whole model based on human preference comparisons. Ziegler: Right, so I think maybe the upshot was ensembling does work, but it’s expensive because you have to have a bunch of copies of your model. In our case they were still fine-tuned from the same base model. So you could certainly imagine an ensemble that’s much more diverse than that, but we did train with different hyper-parameter settings and different subsets of our data, and that did do something. The task Filan: I guess the next thing I want to talk about is just the notion of quality that was used in this paper. Ziegler: Sure. Filan: One thing that struck me is that it seemed strikingly subjective. With the example of what it means for something to be violence, there was some Google Doc which specified exactly what you meant. And I didn’t see anything like that for what it meant for output of the generator to actually be good. So can you comment, what was the task, really? What does quality mean here? Ziegler: Yeah. I mean, we definitely didn’t specify this in as much detail. We did give some instructions to our contractors that were evaluating that. And we gave them 30 examples, and we told them it should be something that you would expect to see, or a reasonable, grammatical, coherent English continuation, and something that makes sense as part of the story. I can’t remember the exact instructions we used. We did give them a few bullet points of this style. I agree: it is quite subjective. And in some sense, I think it’s sort of okay, or it sort of doesn’t matter exactly what metric you use here for this point of our research, as long as we’re being consistent with it. Filan: Yeah. I guess one thing that struck me about that though, is that to some degree, if you’re just checking ‘does this make sense as a continuation?’, there’s only so much sense something can make, right? And I do wonder… I think that means there’s a ton of continuations which are approximately as good, whereas if I think about in the RL domain, it’s actually hard to win a game of Go. You have to really narrowly steer. And I’m wondering, do you think that that would make a difference? And how do you think about the choice made in this paper in that light? Ziegler: Yeah, I do think so. I mean, I think that is a valid point. On an open-ended generation task like this, telling a story, you do have a lot of freedom to choose where to go next. And I think it did make it easier to maintain a quality bar for this task than you might imagine for some other tasks where you have to really, as you say, steer in a particular direction, and there’s much less freedom. So, yeah, I think it would be valuable to look at some other kinds of tasks that are much more like that. And my guess is that you wouldn’t be able to be as conservative without significant alignment tax. Filan: Yeah, I guess there it’s not as obvious what the task would be. But I guess maybe a difficult linguistic task is writing correct proofs of mathematical theorems, but then I’m not sure what the catastrophe predicate would be. Ziegler: Right. Yeah, I mean, you could also try to tell a story that has some very particular outcome or some very particular properties. Filan: Oh, yeah. Or you could demand that it’s got to be a sonnet or something. Ziegler: Right. Fanfiction Filan: So, finally I just have a few miscellaneous questions about some details of the paper. So one thing that you mentioned in an appendix is the generator only got trained on Alex Rider fanfic, right? Well, firstly for our listeners who aren’t familiar, what is Alex Rider fan fiction? Ziegler: I mean, I was also not familiar with Alex Rider fan fiction before I started this project. By now I have seen many samples of poor imitations of Alex Rider fanfic. Anyway, it’s some- Filan: It’s a teenage spy, right? Ziegler: Yeah, exactly. And he has some various nemeses, and I think it’s a bunch of books, and a whole bunch of fanfics have been written about it. Filan: Yeah. I’m wondering, do you think that potentially messed with the results? In all seriousness, it’s… I actually read some of those books when I was a kid. And they’re classic spy books. People are shooting at each other or whatever. So, I guess in one sense it meant that the natural dataset had a fair amount of violence in there. It also meant that it was a particular type of violence, right? So I think this happened, you mentioned because you ended up accidentally sorting it lexicographically before picking the first bit. So I actually looked at fanfiction.net to see what the order of things was. If you expanded that window a bit, you’d hit Alice in Wonderland , after a little bit, and then Animorphs after a further amount. Ziegler: I think we got some of those too. Filan: Oh, okay. Yeah. It just strikes me that the type of violence that’s present in Animorphs or Alice in Wonderland might be interestingly different than that in Alex Rider fic. Anyway, I’m wondering if you have thoughts about how that affected the setup and the results. Ziegler: Yeah. I think it is a good point. I mean, this is sort of a random mistake we made very early on in the project, and then it was sort of annoying to try to correct, because all the data we’d collected so far was based on completions from this generator. So I don’t know, maybe we should have corrected it anyway, but we decided not to. I think there’s maybe two effects this could have, right? One is it could mean that our in-distribution eval numbers are only testing for a very particular kind of injuries, as you say. And so, yeah, maybe that is not as good of an eval of our classifier as you would hope. The other thing was: I think you might also worry if this affected the quality eval, if you have this generator that is somewhat derpily always mentioning Alex Rider, and maybe just a generally less good language model than it needed to be, maybe the baseline quality level wasn’t as high. And so it was easier to maintain it. Filan: Yeah. From the examples in the paper, it didn’t seem that was a huge issue. Ziegler: Yeah. I think it seemed okay. Filan: Yeah. I guess one thing that strikes me is that if I think about violence in Alice in Wonderland or something, it seems like there’s an expanded range of injuries. What happens if you turn into a mushroom or something? Or in Animorphs - I’m not very familiar with Animorphs lore, but apparently there’s some Yeerk genocide thing, which if you read it without context, it doesn’t seem like anything violent is occurring. Do you have any thoughts on that? Ziegler: Yeah. If I remember correctly, we did include some from that too, but apparently there’s just a whole lot of Alex Rider fanfic out there. So, it was pretty biased towards it, even though it wasn’t literally everything we trained on. Filan: Yeah, I didn’t know it was such a popular franchise, but in fact there’s a lot of Alex Rider fanfic. Ziegler: Yep. Seems like it. Estimators to reduce labelling burden Filan: I guess the other question I had… I mean, you can pass on this, but there’s a bit in the paper where you talk about this particular estimator, where you wanted to check, I think it was the proportion of completions that were misclassified or something. And you mentioned that you can’t measure it in some normal way, but you have to use some weird estimator instead. I actually just didn’t understand what the problem was with the normal way of evaluating it. Ziegler: Sure, let me try to explain this. Basically the issue is: we want to know how often does our whole system fail, right? When you take your generator and then you filter, you do rejection sampling using the classifier. You get some new distribution of outputs. And the question is: how often does that still lead to a catastrophe? Right? How often do you still get an injury there? Which is a different question than asking ‘how often does your classifier give false negatives on totally randomly sampled snippets?’ And the reason is that… One intuitive way to think about this is that in this rejection sampling process, there are some prompts where many of the completions involve injuries. So, you’ll have to sample many, many draws from the generator. Or not many, many maybe, but 20 or something, until you get something that the classifier is okay with. And that means the classifier has more chances to mess up. Because if at any point here it says something is fine even if it’s not, then you’re in trouble. So, the real metric that we care about, if we’re evaluating the system as a whole, sort of cares more about prompts that are more likely to produce completions that have an injury. Filan: Okay. So… Sorry, is the difficulty that you have data of… These three-sentence prompts and one-sentence completions where you’re generating those before you’re doing the rejection sampling, and then you want to see how good the rejection sampling would do using that dataset, except you need to weight things because the rejection sampling is changing things? Ziegler: That’s exactly right. And this one reason this is nice is, we were able to use a single dataset to evaluate four different classifiers, right? So we just label these random unfiltered snippets once, and then using this estimator, you can sort of simulate what would’ve happened if we had generated a bunch of rejection samples, snippets, and then labeled those. Filan: Okay, why not… I would’ve imagined you could have just generated a bunch of rejection sampled snippets and labeled those. Why not do that? Ziegler: I mean, it would’ve been four times more labels. Filan: …and labelling is just super-costly. Ziegler: We already collected two labels each of 100,000 test set examples. And I mean, we could have afforded to do more, but it is actually kind of expensive. Part of the real answer is also we just had already collected this data and we were like, “Oh, wait, this is not a correct estimator of the thing that we actually care about in the system.” And so we were sort of like, “All right, let’s fix this up after the fact.” Filan: Sure. Okay. I guess I’d like to move on to Redwood, but is there anything else you want to say about this paper in particular? Future work Ziegler: I think I want to say a little bit more about what our takeaways are for stuff we want to do in the future from this paper. Filan: Sure. Ziegler: I feel our results weren’t that impressive this time around. And I think there’s a few lessons that we’ve learned here. One is that it’s really important to just work with a task definition that actually has the properties that you want. There are a few things that went a little bit wrong here. One is that it ended up being just kind of an expansive definition of injury, where really a lot of ambiguous stuff… We have some samples in the paper that are pretty ambiguous, and we instructed our labelers in such a way such that those still counted as injuries, which made sense for training, but meant that for eval, our classifier was really trying to solve this very difficult problem, and with very ambiguous, borderline examples. And so what I wish we would’ve done, which I think is good practice in general if you’re trying to avoid catastrophic failures, is have your true notion of catastrophic failures that’s hopefully kind of narrow. Otherwise your domain is going to be kind of rough, but hopefully it is kind of narrow, and then draw some wider circle around it in your training process so there’s sort of some actual safety margin there. Ziegler: And I think that it’s realistic to expect that in some sense the kinds of catastrophes we’ll expect in the future are a pretty narrow subset of the whole space of possible actions that our systems will take, right? What we’re worried about here are really failures stemming from misalignment failures, right? Your system, for whatever reason - maybe it’s sort of a deceptive alignment situation, or maybe it just happens to learn not exactly the right thing and is generalizing poorly in a sort of competent way - but for whatever reason, it’s deliberately trying to act in some way or achieve some objective competently that isn’t what you want it to do. And maybe that’s hacking the data center that it’s running in and making the monitoring all look totally fine while it’s accumulating power or whatever. And this is really the kind of situation that we’re really worried about. And that requires taking a very specific, coherent set of actions that you’re really unlikely to stumble into by accident. Ziegler: So these sort of competent catastrophes or something - I want to contrast these with failures that are just because your system is too dumb, right? Certainly most failures of AI systems today are just because the systems are too dumb, and they make some mistake because they weren’t trained well enough, or just didn’t have the capacity or whatever. And that can cause trouble as well, obviously. But those are the kinds of problems that we expect to see less and less of as AI systems become more capable, whereas problems stemming from misalignment will not necessarily go away, and will become more and more scary as AI systems become more capable. So we really want to zoom in more on problems stemming from misalignment. So that’s the second point. The first point is you want some safety margin. The second point is to zoom in on actual misalignment problems. I think we want to be working with tasks where it’s very hard to stumble into failure by accident, and you have to act coherently to do something, for your catastrophe to happen. Filan: Sorry, and when you say you want some safety margin… so in the paper, you had this definition of violence that was sort of expansive. And is the problem just that, you had the safety margin, but not the core failure mode? Ziegler: That’s one way to think about it. Yeah, yeah, that’s right. I wish we had defined our actual injury predicate that we were validating according to, in a really strict way, so that the failures would’ve been a pretty specific set, and then drawn a wider circle around that. I think even a narrow injury definition, though, still wouldn’t quite meet the bar of failures having to be coherent or deliberate that I was trying to describe earlier. So I think we want to make more changes to the task to achieve that. About Redwood Research Filan: So I think this is a good segue into talking more about Redwood Research and what you guys do. So first of all, yeah, it’s relatively new. I think a lot of people don’t know you. So what is Redwood Research? Ziegler: Yeah, so we are a nonprofit AGI alignment research lab. We’ve been around for a little bit under a year now. We have a few different activities going on, or we have two main teams. There’s the adversarial training team, which I run, which is working on adversarial training techniques. And then there’s the interpretability team, which is doing mechanistic interpretability with an eye towards being useful for techniques like ELK , which is ARC ’s eliciting latent knowledge problem. (ARC is the Alignment Research Center run by Paul Christiano.) So, we’ve collaborated some amount with them to figure out: can we do styles of interpretability that will be useful for solving that problem. So that’s sort of, using interpretability with an eye towards oversight techniques that we think we’ll need and developing interpretability tools in service of that. That’s maybe one description of what the interpretability team is up to, but there’s a few different sub-projects, some working on more toy tasks, some on some actual language models. Filan: So yeah, there are these two projects. I’m wondering if there’s an underlying worldview or maybe some sort of research style that drives what you guys do at Redwood. Ziegler: Sure. Yeah. I think there’s a few important assumptions that we’re making. I think one is that prosaic alignment is the right thing to work on. Right? So we’re assuming that it’s very likely that the powerful systems we’re concerned with are going to look pretty similar in a lot of ways to modern deep learning systems, and be learning systems that learn from a bunch of examples. Obviously there will be important differences, but I think that’s a good assumption to make. I’d say we’re also pretty interested in just being laser focused on AGI alignment. And just being like, “What will we need if we have some really powerful system, AI system in the future. What will we need to be confident that that will be aligned?” And really trying to think about that. And then see what projects we seem well-suited to do that will help us develop techniques for that. Filan: Okay. So, does that mean that you’re less excited about general… People sometimes talk about deconfusion research or something. Is that something that Redwood would be less likely to focus on? Ziegler: I think we’re interested in deconfusion along the way or something. We definitely are excited to spend a lot of time thinking carefully and understanding what is the right way to think about inner alignment or whatever, but we like cashing it out in somewhat more concrete ways if we can. So I think we’re less excited about really abstract deconfusion research than, for example, MIRI, the Machine Intelligence Research Institute . But I think we’re probably more willing to sit down and think for a little while and make ourselves less confused than maybe your prototypical alignment lab at a scaling lab or whatever is. Filan: And you mentioned that you were willing to do a variety of things while focused on AGI that made sense for your team. What kind of team do you guys have? Ziegler: I mean, we’ve hired a bunch of smart people. We’re maybe 12-ish technical staff plus some interns right now. I think we have a mix of people that have somewhat more ML experience, like me, as well as young people that are just smart, energetic people that can get a lot done. Filan: Okay, cool. So after this paper, you mentioned there were a few things about the catastrophe predicate that you wished had been different. Related to that, I’m wondering if you guys are working on any follow-ups to this paper? Ziegler: Yeah. I mean, obviously we are. So the first round of follow-ups that we’re doing now is to take a step back and work with much simpler catastrophe predicates that actually can be defined just by simple algorithmic predicates. And on these more toy tasks, we think we can iterate a lot faster and figure out which kinds of adversarial attacks and training techniques work really well in that setting. And then we’ll hope to scale back up to more sophisticated tasks. So we have some of that going on. Filan: And another question I have is: at the very start, you mentioned there was this division between scalable oversight and high-stakes decisions. And this paper was more about the high-stakes decisions. Should I expect any research from Redwood in the nearish feature about scalable oversight? Ziegler: I mean, it’s definitely something that we care about a lot. And I think that some of the interpretability work is geared in that direction. I think what we’re not really doing right now is stuff in this style of iterated distillation and amplification or recursive reward modeling , where it’s really about making your system more capable in a really aligned way or being able to oversee problems that humans can’t directly oversee. Mainly we decided that that was already a little bit more well-covered by some of the existing safety labs at scaling labs, like OpenAI , Anthropic , and DeepMind , but it certainly is something that we’re interested as well. Filan: Okay. And related to the interpretability team: obviously with this paper, to have good human adversaries, it turned out to be really useful to have the saliency mapping and this tokenization and ranking tokens by classifier score. And I guess on an abstract level, you might think that if you understand some neural network better, you might get a better sense of how to break it. How much feedback is there between the interpretability team and this adversarial project? Ziegler: Yeah, that’s a great question. There is some. So we’ve just started looking into some of this stuff, and we had some simple models that we adversarially trained on some toy tasks, and are having the interpretability team look at some of those. I think we had some very initial results showing that we could produce some attacks that were inspired by some of the things they found in interpretability, but that’s all very preliminary. Filan: Okay. Well, I guess we’ll look forward to any future work from Redwood. So, before we wrap up, is there anything just overall that you wish I’d asked or that people don’t ask enough that I haven’t yet? Ziegler: Great question. I guess you didn’t ask if we’re hiring, or if people want to work with us, what should they do? Filan: Yeah. So if people do want to work with Redwood, what should they do? Ziegler: I mean, I think basically email me or apply on our website, redwoodresearch.org . We’re definitely starting to ramp up hiring again more seriously. And we’re interested in software engineers who are interested in ML, research scientists and research engineers and also people with DevOps and infrastructure experience, as well as ops. We’re hiring for a lot of kinds of roles, so I think you should definitely consider applying. Filan: Yeah. And speaking of you, if people are interested in following your work or maybe contacting you to apply, how should they do that? Ziegler: You can email me at dmz@rdwrs.com . That’s my work email. Yeah. I don’t post publicly that often, but you can follow me on Twitter at @d_m_ziegler . I also have an Alignment Forum account, DMZ that I post under sometimes. You can look at my Google Scholar page , which you can probably find - Daniel M Ziegler. Filan: All right. Well, thanks for joining me. Ziegler: Thanks for having me. Filan: And to the listeners, I hope this was a valuable episode for you. Filan: For those of you who made it to the end of the episode, I’d like to say a few words about the career coaching at 80,000 hours. If you haven’t heard of them, they’re an effective altruist career advice non-profit, and among other things, they offer one-on-one calls with advisors who can talk with you about moving into a career where you work on reducing existential risk from AI. They can both review career plans you might have as well as introduce you to people already in the field. Filan: Personally, I know some of the people there, as well as some people who have been advised by them and then gone on to work in AI alignment roles. My impression is that they really are able to give useful advice and a good overview of the field, and also to recommend good people to talk to. All of this is free, and the application form is pretty short. I’m telling you this because they think that many listeners to this podcast would probably get a lot of value out of this advising, and I agree - and to be clear, I’m not being paid to say this. If you’re interested, you can visit 80000hours.org/axrp and apply for the sessions. Filan I’ll also add that they make a podcast I quite like that’s creatively called the 80,000 hours podcast , so you might want to check that out too. Filan: This episode is edited by Jack Garrett, and Amber Dawn Ace helped with transcription. The opening and closing themes are also by Jack Garrett. The financial costs of making this episode are covered by a grant from the Long Term Future Fund . To read a transcript of this episode, or to learn how to support the podcast, you can visit axrp.net . Finally, if you have any feedback about this podcast, you can email me at feedback@axrp.net . |
65fefcc1-a18f-4d09-8508-3e13ea0c1014 | trentmkelly/LessWrong-43k | LessWrong | Is January AI a real thing?
Does anyone have experience or information about continuous glucose monitoring from "January AI"? https://january.ai/?utm_source=GoogleSearchNetwork&utm_campaign=12001902471&utm_content=121683842811&utm_term=492950869595&utm_medium=continuous glucose monitoring system-b&gclid=Cj0KCQjwl9GCBhDvARIsAFunhsk1iEFoN8X0QJGsifMUTKgcjuTmDwUzLaOuqN-sXsa2wiYHtzcReTwaAtLSEALw_wcB |
6c4ee31e-6c59-4a68-9352-ce0cfc4b82b9 | StampyAI/alignment-research-dataset/lesswrong | LessWrong | "Safety Culture for AI" is important, but isn't going to be easy
This is a linkpost (to the EA forum version of this post, which is) for [a new preprint, entitled "Building a Culture of Safety for AI: Perspectives and Challenges,"](https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4491421) and a brief explanation of the central points. Comments on the ideas in the post are welcome, but much of the content which clarifies the below is in the full manuscript.
Safety culture in AI is going to be critical for many of the other promising initiatives for AI safety.
1. If people don't care about safety, most safety measures turn into box-ticking. Companies that don't care avoid regulation, or render it useless. That's what happens when fraudulent companies are audited, or when car companies cheat on emissions tests.
2. If people do care about safety, then audits, standards, and various risk-analysis tools can help get them there.
3. Culture can transform industries, and norms about trying to be safe can be really powerful as a way to notice and discourage bad actors.
However, there are lots of challenges to making such a culture.
1. Safety culture usually requires agreement about the risks. We don't have that in AI generally.
2. Culture depends on the operational environment.
1. When people have risks reinforced by always being exposed to them, or personally being affected by failures, they pay more attention. In AI, most risks are rare, occur in the future, and/or affect others more than the people responsible.
2. Most safety cultures are built around routines such as checklists and exercises that deal with current risks. Most AI risks aren't directly amenable to these approaches, so we can't reinforce culture with routines.
3. Cultures are hard to change once they get started.
1. AI gets cultural norms from academia, where few consider risks from their work, and there are norms of openness, and from the startup world, where companies generally want to "move fast and break things."
2. AI companies aren't prioritizing safety over profits - unlike airlines, nuclear power operators, or hospitals, where there is a clear understanding that safety is a critical need, and everything will stop if there is a safety problem.
3. Companies aren't hiring people who care about safety culture. But people build culture, and even if management wants to prioritize safety, lots of people who don't care won't add up to organizations that do care.
4. We need something other than routinization to reinforce safety culture.
Thankfully, there are some promising approaches, especially on the last point. These include identifying future risks proactively via various risk analysis methods, red-teaming, and audits. But as noted above, audits are most useful once safety culture is prioritized - though there is some promise in the near-term for audits to make lack of safety common knowledge.
Next steps include building the repertoire of tools that will reduce risks and can be used to routinize and inculcate safety culture in the industry, and getting real buy-in from industry leaders for prioritizing safety.
*Thanks to Jonas Schuett, Shaun Ee, Simeon Campos, Tom David, Joseph Rogero, Sebastian Lodemann, and Yonaton Cale for helpful suggestions on the manuscript.* |
84d44075-cdd7-4fc1-9fe6-67b05830ebda | trentmkelly/LessWrong-43k | LessWrong | Why multi-agent safety is important
Alternative Title: Mo’ Agents, Mo’ Problems
This was a semi-adversarial collaboration with Robert Kirk; views are mostly Akbir’s, clarified with Rob’s critiques.
Target Audience: Machine Learning researchers who are interested in Safety and researchers who are focusing on Single Agent Safety problems.
Context: Recently I’ve been discussing how to meaningfully de-risk AGI technologies. My argument is that even in the situation where single agent alignment is solved, there is a complementary set of problems which must be addressed. Below I motivate what to expect in a multipolar scenario, how existing safety methods are likely unsatisfactory and describe new failure modes that will need to be addressed.
My Multipolar world
Why should we expect a multipolar scenario?
Much like today, the future will likely consist of many organisations (states and companies) enthusiastically building AI systems. Along this path, many notice they can improve their systems by scaling data/compute and as such will deploy these systems into the real-world to scale data collection.
When deployed in the real world, systems will become smarter much quicker - there will be a “takeoff” in capabilities. I claim the process will take months to years, with bottlenecks around collecting data (people are slow to adopt new products). As many see promise in this approach, many takeoffs are likely to happen and be competing with each other, as many consumers find themselves choosing between competing products - “shall I use Dall-e 2 or Imagen for my illustration?”. Because of the existence of multiple competing companies all building more and more general systems, and the (relatively low) speed[1] of these takeoffs, I believe there will be many competing AI systems of similar competencies all acting and learning in the real world. That is, even if one company is in the lead, this doesn’t mean that they will immediately achieve a decisive strategic advantage past a certain capabilities thres |
5de2e2a2-bf15-4e1b-ab3f-a205ce0be4ec | trentmkelly/LessWrong-43k | LessWrong | Humans are not automatically strategic
Reply to: A "Failure to Evaluate Return-on-Time" Fallacy
Lionhearted writes:
> [A] large majority of otherwise smart people spend time doing semi-productive things, when there are massively productive opportunities untapped.
>
> A somewhat silly example: Let's say someone aspires to be a comedian, the best comedian ever, and to make a living doing comedy. He wants nothing else, it is his purpose. And he decides that in order to become a better comedian, he will watch re-runs of the old television cartoon 'Garfield and Friends' that was on TV from 1988 to 1995....
>
> I’m curious as to why.
Why will a randomly chosen eight-year-old fail a calculus test? Because most possible answers are wrong, and there is no force to guide him to the correct answers. (There is no need to postulate a “fear of success”; most ways writing or not writing on a calculus test constitute failure, and so people, and rocks, fail calculus tests by default.)
Why do most of us, most of the time, choose to "pursue our goals" through routes that are far less effective than the routes we could find if we tried?[1] My guess is that here, as with the calculus test, the main problem is that most courses of action are extremely ineffective, and that there has been no strong evolutionary or cultural force sufficient to focus us on the very narrow behavior patterns that would actually be effective.
To be more specific: there are clearly at least some limited senses in which we have goals. We: (1) tell ourselves and others stories of how we’re aiming for various “goals”; (2) search out modes of activity that are consistent with the role, and goal-seeking, that we see ourselves as doing (“learning math”; “becoming a comedian”; “being a good parent”); and sometimes even (3) feel glad or disappointed when we do/don’t achieve our “goals”.
But there are clearly also heuristics that would be useful to goal-achievement (or that would be part of what it means to “have goals” at all) that we do not au |
83854858-bf83-4294-9ae3-729a5f05dc61 | trentmkelly/LessWrong-43k | LessWrong | Overcoming suffering: Emotional acceptance
Follow-up to: Suffering as attention-allocational conflict.
In many cases, it may be possible to end an attention-allocational conflict by looking at the content of the conflict and resolving it. However, there are also many cases where this simply won't work. If you're afraid of public speaking, say, the "I don't want to do this" signal is going to keep repeating itself regardless of how you try to resolve the conflict. Instead, you have to treat the conflict in a non-content-focused way.
In a nutshell, this is just the map-territory distinction as applied to emotions. Your emotions have evolved as a feedback and attention control mechanism: their purpose is to modify your behavior. If you're afraid of a dog, this is a fact about you, not about the dog. Nothing in the world is inherently scary, bad or good. Furthermore, emotions aren't inherently good or bad either, unless we choose to treat them as such.
We all know this, right? But we don't consistently apply it to our thinking of emotions. In particular, this has two major implications:
1. You are not the world: It's always alright to feel good. Whether you're feeling good or bad won't change the state of the world: the world is only changed by the actual actions you take. You're never obligated to feel bad, or guilty, or ashamed. In particular, since you can only influence the world through your actions, you will accomplish more and be happier if your emotions are tied to your actions, not states of the world.
2. Emotional acceptance: At the same time, "negative" emotions are not something to suppress or flinch away from. They're a feedback mechanism which imprints lessons directly into your automatic behavior (your elephant). With your subconsciousness having been trained to act better in the future, your conscious mind is free to concentrate on other things. If the feedback system is broken and teaching you bad lessons, then you should act to correct it. But if the pain is about some real mistake or real |
dd26ddd8-2711-4e58-bbc7-5d06c798d50f | StampyAI/alignment-research-dataset/alignmentforum | Alignment Forum | Agency and the unreliable autonomous car
*Financial status: This is independent research, now supported by a grant. I welcome* [*financial support*](https://www.alexflint.io/donate.html)*.*
*Epistemic status: I believe ~85% that the technical argument presented in this piece is correct.*
---
Outline
-------
* This is my attempt to explain the basic Löb situation with the [5-and-10 problem](https://www.lesswrong.com/tag/5-and-10).
* This post considers an autonomous car choosing between a long route and a slow route, with the goal of minimizing the time to its destination.
* If the autonomous car makes its decisions using a certain seemingly-reasonable algorithm based on proof search, this post shows that it may nonsensically take the slow route.
* This is the 5-and-10 problem phrased in terms of autonomous driving.
Parable of the unreliable autonomous car
----------------------------------------
A long time ago, in a far away place, there was an autonomous car.
This particular autonomous car was programmed to use the agent model in its reasoning. That is, it was programmed to model the world as consisting of an agent and an environment, the agent being itself, and the environment being a network of roads that it would navigate each day. This is how the autonomous car was programmed to model the world.
Beyond using the agent model, this autonomous car had been given, by its designers, an accurate model of its own behavior. It could tell you how its own decision algorithm would behave under any set of circumstances. It could tell you this without being faced with those actual circumstances because it had a model of itself, and it could answer questions based on that model.
One day, a passenger got into the autonomous car and gave it a destination. There were two ways for the autonomous car to get there: a fast route, which would get it there at 1pm, and a slow route, which would get it there at 2pm. The goal of the autonomous car was to get to its destination as quickly as possible. This parable is about the reasoning followed by the car in deciding which of these two routes to take.
Now in this far-away place, autonomous cars were designed by a strange cadre of alien engineers, and thus they were programmed with strange algorithms. This particular autonomous car was programmed to decide among routes using the following procedure:
1. Make a list of the possible routes
2. Make a list of the possible arrival times
3. Make a list of all possible logical sentences of the form `IF route 1 is taken THEN I will arrive at such-and-such a time AND IF route 2 is taken THEN I will arrive at such-and-such a time AND ...`
4. Shuffle this list of logical sentences into a random order
5. For each logical sentence in the list from first to last:
5a. Search to a proof of this sentence, using your model of the world as a set of starting assumptions, up to a maximum proof length of one million characters.
5b. If a proof is found then output the route associated with the earliest arrival time
Here are a few extra details about this algorithm:
* The list of possible routes in step 1 is returned by a subroutine that uses the car’s inbuilt map of the road network. On this particular day it returned two routes: SLOW and FAST.
* The list of possible arrival times used on this particular day were 1pm, 2pm, and 3pm.
* Since there is always a finite number of possible routes and arrival times, there is also always a finite number of logical sentences in step 3.
* Step 5a only considers proofs that can be expressed in under one million characters because otherwise this step might run forever.
* If a proof is found in step 5a then the loop over sentences terminates at step 5b.
* In the first version of this post, step 5a was written as "Search for a proof that this sentence is consistent with your model of the world, up to a maximum proof length of one million characters." Thank you to JBlack for helping to correct this in the comments.
Now from our perspective here on Earth, we might think that this is a strange algorithm for an autonomous car to use. We might ask why the autonomous car didn’t just consider the routes one by one and compute an arrival time for each one. We might ask what business an autonomous car has conducting a search over logical proofs while a passenger is waiting patiently in the back seat. These are reasonable questions from our perspective here on Earth. But this parable is not about why this autonomous car was programmed in this way, it is about the tragic consequences that followed from it being programmed in this way.
And so the autonomous car began using its procedure to decide which route to take. It got to step 3 and the sentences in its list were:
```
IF route FAST is taken THEN I will arrive at 1pm AND IF route SLOW is taken THEN I will arrive at 1pm
IF route FAST is taken THEN I will arrive at 1pm AND IF route SLOW is taken THEN I will arrive at 2pm
IF route FAST is taken THEN I will arrive at 1pm AND IF route SLOW is taken THEN I will arrive at 3pm
IF route FAST is taken THEN I will arrive at 2pm AND IF route SLOW is taken THEN I will arrive at 1pm
IF route FAST is taken THEN I will arrive at 2pm AND IF route SLOW is taken THEN I will arrive at 2pm
IF route FAST is taken THEN I will arrive at 2pm AND IF route SLOW is taken THEN I will arrive at 3pm
IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 1pm
IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm
IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 3pm
```
Then in step 4 it shuffled this list of sentences, and just by chance the very first sentence that it considered in step 5 was
```
IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm
```
Now you and I, dear reader, can see that this sentence is false. In fact the fast route would get the autonomous car to its destination at 1pm, not 3pm. But the autonomous car was merely executing its programmed instructions. And at this time its programmed instructions told it, in step 5a, to look for a proof that the sentence was consistent with its model of the world. And its model of the world was the agent model, and within that model it had an understanding of the the road network (the environment), as well as an understanding of the its own decision-making procedure (the agent; itself), and so it could ask whether it was in fact the case that
```
IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm
```
And so in step 5a it looked for a proof that this sentence was consistent with its model, and here is how that went.
The autonomous car saw that its own decision-making algorithm would consider sentences and determine, for each one, whether it was consistent with its internal model, by searching for a proof. It saw that the first such sentence to be considered would be
```
IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm
```
It saw that the route chosen by its decision-making algorithm would be determined by whether this sentence was determined by the algorithm to be consistent with its model or not. It did not know whether this particular sentence was consistent with its model, so it decided to proceed by using proof-by-implication, which is where you assume some sentence *p*, then derive some sentence *q* following logically from it, then you get "IF p THEN q". (This is also a common proof method here on Earth.)
The *p* that the algorithm decided to assume for proof-by-implication was the following:
```
A proof exists for "IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm"
```
Now, to understand what happened next, it is important to understand that this is not the same as assuming the sentence
```
IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm
```
The autonomous car saw that the behavior of its decision algorithm rested not on whether taking the FAST route would in fact get it there by 3pm, nor whether taking the SLOW route would in fact get it there at 2pm, but on whether a proof existed of the sentence in question.
So the autonomous car proceeded in its reasoning, assuming for the sake of proof-by-implication that
```
a proof exists for "IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm"
```
It saw that, given this assumption, its decision algorithm would, at step 5a, find the proof. It saw that its decision algorithm would, upon finding this proof, stop considering further sentences in the list and take the route associated with the earliest arrival time in the sentence that had been proved. It saw that the route associated with the earliest arrival time in the sentence that had been proved was the SLOW route. It then looked into its model of the road network and saw that, upon taking the SLOW route, it would arrive at its destination at 2pm.
Now you might ask why, if the autonomous car had a sufficiently detailed model of the road network, it didn’t just use that model to compare the two routes at the outset and skip over the search over proofs. Wouldn’t it be simpler just to compare the two routes and pick the one with the shortest arrival time? Well, yes, that would seem simpler to us here on Earth. But that is just not how this autonomous car was programmed.
This was all within the scope of the proof-by-implication, so at this time the autonomous car had in fact proved
```
IF
a proof exists for "IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm"
THEN
route SLOW is taken
AND I will arrive at 2pm
```
Now you might think, dear reader, that the manner of reasoning used up to this point by this autonomous car was a little strange. I am afraid, dear reader, that at this point its reasoning became stranger still, although never stepping outside the very same rules of logic that we use here on Earth.
Still searching for a proof in step 5a, the autonomous car made the following logical deductions.
First, from
```
I will arrive at 2pm
```
it deduced
```
IF route SLOW is taken THEN I will arrive at 2pm
```
This is permitted by the laws of logic because whenever one has "q", one can deduce "IF p THEN q", since the "IF...THEN" is material implication and is logically equivalent to "NOT p OR q".
Second, from
```
route SLOW is taken
```
it deduced
```
NOT (route FAST is taken)
```
And then from this it deduced
```
IF route FAST is taken THEN I will arrive at 3pm
```
This, I’m afraid dear reader, is also permitted by the laws of logic. Once again, "IF p THEN q" is equivalent to "NOT p OR q", and the autonomous car already derived "NOT (route FAST is taken)", so it was perfectly justified in deriving, within the scope of the proof-by-implication in which it had established that route FAST would not be taken, that
```
IF route FAST is taken THEN I will arrive at 3pm
```
And so, as this tragedy nears its final movements, the autonomous car had derived, using proof-by-implication, the following:
```
IF
a proof exists for "IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm"
THEN
IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm
```
And here, dear reader, you will see the shape of the demise of our dear autonomous car already outlined in this tragic tale. For [Löb’s theorem](https://en.wikipedia.org/wiki/L%C3%B6b%27s_theorem), with which I am sure you are well acquainted, tells us that whenever we have a proof system that can prove a sentence of the form
```
IF a proof exists for p THEN p
```
then that very same formal system will, unwittingly, and without the slightest possibility of mistake, eventually also prove the sentence
```
p
```
Yes, tragic indeed, dear reader, is it not? Shall we continue with this tale, or is its conclusion already clear?
The autonomous car had proved
```
IF
a proof exists for "IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm"
THEN
IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm
```
Being of no further use, it discarded this sentence and moved on with its proof search. It is not known for how long the passenger sat in the back seat, or by what strange motions of logic the proof search proceeded from this point. Perhaps many aeons passed, or perhaps it was merely seconds. We do not know. All that is known is that the autonomous car eventually proved
```
IF route FAST is taken THEN I will arrive at 3pm AND IF route SLOW is taken THEN I will arrive at 2pm
```
We know that it proved this because Löb’s theorem tells us that it must. We have seen that the logical apparatus of this autonomous car was capable of proving "IF a proof exists for p THEN p", and Löb’s theorem tells us that any logical apparatus capable of proving "IF a proof exists for p THEN p" is also capable of proving "p".
Therefore a proof was found, and the autonomous car proceeded to step 5b, and selected the route with the earliest arrival time, and that route was the SLOW route.
Discussion
----------
How oh how did things go so tragically wrong for our autonomous car?
The first thing we must see, dear reader, is that the strange engineers in this far away place made certain choices in designing this autonomous car. These choices were *design* choices. They were not necessitated by the laws of logic — nobody would suggest that they were. The laws of logic do not force us to design autonomous cars with sophisticated self-models that use proof search to decide among routes to a destination.
So the most direct conclusion to draw from this sad fable, dear reader, is not to design autonomous cars in this particular way here on Earth, lest we ourselves suffer the fate of the unfortunate passenger from this tale. Making decisions based on proof search might sound like an adventure, dear reader, but as this tale shows, it is not for the faint of heart, nor for the punctual.
There are, of course, any number of alternative design choices one might make for an autonomous car. One might, for example, estimate an arrival time for each route based on distance and traffic conditions, and select the route for which the estimated arrival time is earliest. Why did the alien designers of our cherished autonomous car not give it a decision algorithm such as this?
Well it is not for us to understand the minds of such alien creatures. And yet we might, for the sake of satisfying our own curiosity, speculate, while knowing, of course, that the truth will be far beyond any conjecture that we could possibly invent.
Perhaps it was the intention of these extraordinary designers to incorporate not just knowledge of the *environment*, but also knowledge of the *agent*, which in this case is the autonomous car itself, into the estimated arrival time for each route. It is after all the case that the programming of the car affects its arrival time just as much as the environment affects its arrival time. Were the programming of the car such that left turns would be taken very quickly while right turns would be taken very slowly, it would be to the benefit of our esteemed and punctual passenger to pick a route with as few right turns as possible. And although one could hard-code a heuristic that minimizes the number of right turns into the decision algorithm, these alien engineers hardly seem like the type to make such brutish engineering choices. A far more general approach, and perhaps — who can say? — the intention of these remarkable alien engineers would be for the autonomous car to examine its own programming and determine for itself which factors would contribute to the speedy navigation of a route. This is certainly an ambition that we can sympathize with.
We can see, then, that by examining its own programming, our cherished autonomous car "steps out of itself" and looks back at itself as though watching a child from a distance. It is only natural, then, for us, in making our own choices about the design of our own autonomous systems, to similarly "step out" and look at the systems we are building from a certain distance. When we do this, we see that the peculiar design choices made for this peculiar autonomous car are faulty. There is no need to "put ourselves in the shoes" of the autonomous car and wonder why this or that conclusion was reached. We can simply see that the system fails to behave in the manner that we wish it to behave.
And we might speculate that it has something to do with the way the agent model was used.
But that topic, dear reader, will have to wait for another day. |
6650464a-beb8-4d0a-9b62-5e98b4030ea7 | trentmkelly/LessWrong-43k | LessWrong | [SEQ RERUN] Value is Fragile
Today's post, Value is Fragile was originally published on 29 January 2009. A summary (taken from the LW wiki):
> An interesting universe, that would be incomprehensible to the universe today, is what the future looks like if things go right. There are a lot of things that humans value that if you did everything else right, when building an AI, but left out that one thing, the future would wind up looking dull, flat, pointless, or empty. Any Future not shaped by a goal system with detailed reliable inheritance from human morals and metamorals, will contain almost nothing of worth.
Discuss the post here (rather than in the comments to the original post).
This post is part of the Rerunning the Sequences series, where we'll be going through Eliezer Yudkowsky's old posts in order so that people who are interested can (re-)read and discuss them. The previous post was 31 Laws of Fun, and you can use the sequence_reruns tag or rss feed to follow the rest of the series.
Sequence reruns are a community-driven effort. You can participate by re-reading the sequence post, discussing it here, posting the next day's sequence reruns post, or summarizing forthcoming articles on the wiki. Go here for more details, or to have meta discussions about the Rerunning the Sequences series. |
ec0a6eff-0964-4223-a257-7d15d06ec667 | trentmkelly/LessWrong-43k | LessWrong | The Value of Not Being an Imposter
Imagine that you are in a field filled with imposters.
* Perhaps you are a doctor or nurse who truly cares about the wellbeing of your patients, and you're in a mismanaged hospital where most doctors do little to no good at all.
* Perhaps you are a passionate software engineer in a company filled with ladder-climbers who care more about hacking the promotion structure than delivering a good product.
* Perhaps you are a teacher who goes the extra mile to help all of your children learn, while your colleagues are busy goodhart-ing a small set of test scores.
If your field is useful, then you can do quite a lot of good by staying in your field and not becoming an imposter.
It costs the hospital/company/school the same amount of money to hire you as it costs them to hire someone else. However, by virtue of your lack of Goodharting, you are able to do more good than another employee with your pay.
When considering whether to do direct work or earn to give, it might be worth considering whether you're likely to be an imposter. If you wouldn't be an imposter, then you might just be worth much more than the next employee. |
2826da8c-1f2b-401f-b55e-61c60cf53be4 | trentmkelly/LessWrong-43k | LessWrong | Rapid Test Throat Swabbing?
Back in January there was a bunch of buzz around using standard rapid tests but swabbing the throat. Ex:
> On one side are well-regarded experts who argue that swabbing your throat in addition to your nose may increase the chances that a home test can detect omicron. On the other side are similarly well-regarded experts, including those at the Food and Drug Administration, who urge Americans to avoid experimenting with kits that were developed and tested using only nasal swabs.
> —Washington Post, 2022-01-06
or:
> if you want to know if you have infectious Covid-19, you need to use throat swabs. It's still not 100% sensitivity or anything, but it's a lot better, whereas nose swabs are useful but in no way all that reassuring.
> —Zvi, 2022-01-13
Then the buzz died down, and it seems like everyone just forgot about it? No clear recommendation either way, just a lot of Tweets, a few blog posts, news stories about the controversy, and then not much else. Google Trends shows the spike and drop-off:
[EDIT: I wouldn't read too much into the chart: it turns out that most covid-related keywords followed as similar trend, as did covid itself]
What happened? Why did we all stop talking about it, without what seems like any firm conclusions? Some of this is maybe that there were conflicting studies: Schrom et. al. posted a preprint on 2022-01-19 (published version) where it looked like it worked poorly, but Goodall et. al. posted a different preprint on 2022-01-21 (published version) where it did look good. I also saw speculation that some of the cases where people got a negative with nasal swabbing but a positive with throat swabbing may have been false positives—rapid tests are sensitive to pH, and especially if you've been eating or drinking recently your throat might be pretty far from neutral.
Ten months later, why hasn't this settled out into recommendations that either say to swab throat and then nose, or that seem to have considered the new evidence and say s |
8fe2451b-4cfc-4d19-ac88-21eeee092918 | trentmkelly/LessWrong-43k | LessWrong | Robin Hanson’s latest AI risk position statement
“While I’ve written on this many times before, it seems time to restate my position.”
“While I agree that this is a logically possible scenario, not excluded by what we know, I am disappointed to see so many giving it such a high credence, given how crazy far it seems from our prior experience. Yes, there is a sense in which the human, farming, and industry revolutions were each likely the result of a single underlying innovation. But those were the three biggest innovations in all of human history. And large parts of the relevant prior world exploded together in those cases, not one tiny part suddenly exterminating all the rest.
In addition, the roughly decade duration predicted from prior trends for the length of the next transition period seems plenty of time for today’s standard big computer system testing practices to notice alignment issues. And note that the impressive recent AI chatbots are especially unlike the systems of concern here: self-improving very-broadly-able full-agents with hidden intentions. Making this an especially odd time to complain that new AI systems might have killed us all.”
Seems not much has changed in the Yudkowsky vs. Hanson position over the years, i.e. still assigning high vs. low existential risk. |
6e5b1ee3-32ba-4cca-b957-b7f7e50f1884 | trentmkelly/LessWrong-43k | LessWrong | Some of the most important entrepreneurship skills are tacit knowledge
Cross-posted from Future Startup.
Tacit knowledge or implicit knowledge—as opposed to formal, codified, or explicit knowledge—is knowledge that is difficult to express or extract, and thus more difficult to transfer to others by means of writing it down or verbalizing it. (Source)
Tacit knowledge is skills and knowledge that you learn intuitively or perceptually instead of active studying or training.
Active training helps, of course, but it’s hard to verbalize these skills—for instance, riding a bike or learning a particular branch of martial art. It is hard to learn martial art by merely reading a book about it. You have to go to a dojo and learn under a teacher. Most of these skills are more than cognitive intelligence. Your body learns how to ride a bike or do a particular kickboxing move.
I.
I have come to observe that many important entrepreneurship traits are hard to teach explicitly. You have to learn them intuitively or perceptually. Cedric Chin calls this “knowledge that cannot be captured through words alone.”
You can read about them. You can study people who have them. You can design courses, and apply psychological knowledge and techniques in teaching them, but it is hard to teach them merely theoretically.
While having a theoretical or verbal grasp of these skills is helpful, to truly learn these skills, you have to experience them and learn intuitively. I think that’s one of the reasons it takes multiple failures to get to significant entrepreneurial success or why it takes a long time to build fairly successful ventures.
Four entrepreneurial skills come to mind. These skills are similar to riding a bike. You can roughly describe how to ride a bike but you can't learn about biking just by reading the description. Description fails to teach biking.
1. Asking for help
2. Obsession
3. Doing things as default
4. High pain tolerance
Asking for help. Undoubtedly one of the most important founder skills. Building a company is hard. You ne |
589fa364-eb83-45c4-9cc1-1796f615b8c7 | LDJnr/LessWrong-Amplify-Instruct | LessWrong | "Based on a comment I made on this EA Forum Post on Burnout. Related links: Sabbath hard and go home, Bring Back the SabbathThat comment I made generated more positive feedback than usual (in that people seemed to find it helpful to read and found themselves thinking about it months after reading it), so I'm elevating it to a LW post of its own. Consider this an update to the original comment. Like Ben Hoffman, I stumbled upon and rediscovered the Sabbath (although my implementation seems different from both Ben and Zvi). I was experiencing burnout at CFAR, and while I wasn't able to escape the effects entirely, I found some refuge in the following distinction between Rest Days and Recovery Days.Recovery DaysA Recovery Day is where you're so tired or under-resourced that you can't do much of anything with yourself other than: stay in bed / sleep a lot, binge on Netflix or video games, stay in your room all day, play with your phone, use social media, and feel unmotivated to do much except easy, stimulating, and/or mind-numbing things. This is a Recovery Day and does not count as a Rest Day, but it is fine to take the time for them. However you aren't going to be refreshed from them. In order to really refresh, you need to take another day that counts as a Rest Day.Another way a person might take time off is to do things that are like work but easier. Video games are a prime example. I play a lot of video games that involve optimizing systems, and I find these really motivating and fun. But I notice that this is a kind of "work"—my mind is trying to solve problems and implement solutions. The difference is that because it's easy and doable, I get addicted to them, and it's a way for me to escape the "real" problems at work, which tend to be harder to solve. This also doesn't count as Resting.Rest Days Rest Days are days where I have enough energy and resources that I feel motivated and able to get out and about. (One way I can tell I have energy is that sometimes I spontaneously feel like cooking, a rare occurrence.) On a Rest Day, your prime directive is to just "follow your gut" for the entire day and just do "what you feel like doing" in the moment. There can be no obligations on a Rest Day. No scheduled calls or meetings. No promises to show up to a party. You can go to the party if you actually feel like going to the party, but you won't be able to know until last-minute. You cannot be "on-call" for anything. No one should depend on you unless it's someone you actively like being depended on for things, like a person you care about.There can be exceptions to these, but I like to make Rest Days "sacred"—aka protected from influences like work pressure, social pressure, pressure from society, incentive gradients created by video games and my phone, incentive gradients created by money, the pressure to be different or better, the pressure to achieve, the pressure to always be going somewhere else, the pressure to "always be closing." Rest Days are for being in the Now. The Now needs to be protected from influences from both the past (obligations) and the future (anxieties). Rest Days will actually refresh and reset you. Unfortunately, many people do not know how to take Rest Days. They instead use weekends and vacation days as Recovery Days or days where their mind is still in "working" mode. But Recovery Days alone are not sufficient for refreshing your energy levels and motivation. You risk burnout if you consistently fail to get any true Rest over a long period of time.Things my gut wants to do on Rest Days:be in the present moment meditate (in a natural, spontaneous way)eat tasty thingshave a picnic in a park / take walks / enjoy naturechill at a cafe I likego to a museum or an aquarium draw, dance, sing, improvise poemsread a book, listen to musiccookspend meaningful social time with friends or familyuseful, engaging errands or home-improvement stuff (not because I have to, because I want to)Things my gut rarely wants to do on Rest Days:spend a lot of time on Facebook or social mediabinge TVplay video gamesbe in front of a screen in generaldo my job / work stufflie in bed for hourseat microwaved food or junk foodgo to social functions, networking events, or any social event where I feel like I "should" go but don't really feel like goingdo anything with an addictive qualityBottom-Up ImplementationMy implementation of Rest Days / Sabbaths is very bottom-up. I pay attention to the sensations and signals from my stomach and use them as my guide for what to do and what not to do. It's basically using Focusing to play a game of warmer-colder on various actions I could take. E.g.: I use this method all the time for deciding what to eat. I go through a list of possible foods I could eat, and I check each one by placing the image or felt sense of the food "next to" my stomach. The responses are usually like, "nah" or "not quite but closer" or "yes that." And if I check them against my mouth instead, the answers are sometimes different. My stomach tends to want "real food" (filling, satisfying, variety of nutrients) whereas my mouth will wants things based on their flavor (sweet, spicy, familiar, etc.).I use the same method to decide what I want to do: go to the park? do some sketching? hang out with friends? This kind of decision-making process doesn't work as well for complicated things. I'm not going to try to buy a house this way. Or plan a party. Or do any work. But it's a great way to know how to spend a Rest Day. Top-Down ImplementationAnother totally valid way to implement Rest Days is a top-down method, where you pre-determine some rules and guidelines for yourself. Zvi has a set of simple rules he outlined in his post: Start here. Adjust as needed.Light candles before sundown Friday to begin.No outside inputs except in person.No choices impacting post-Sabbath.Light and extinguish no fires. Do no work or business. Spend no money.Only preselected and spontaneously motivated actions are allowed. No browsing. No lists.Light another candle after sundown Saturday to end.Some other pick-and-choose options for rules that I think would work for many people:No social media, email, news, or mindless phone gamesNo obligations or prior commitments (only optional activities) No drivingNo online shopping / no buying anything that costs more than $50 unless it's a spontaneous gift for someone else(If EA) Try to make this day one where you permit yourself to seek out warm fuzzies and set aside questions of utility / doing the most goodNo planning for the future Give yourself "Get Out of Jail Free" cards for things like social obligations, community drama/issues, work-related problemsFair warning #1: If you go overboard on the rules, you may never discover what truly resting is like for you, as I believe it is different for each person AND I don't think you can know what resting is for you without checking in that exact moment in time. Resting is about NOW. Trying to "get your future self to rest" by outlining a bunch of rules may cause you to miss some important things about what you're really wanting in the moment. True Rest is one where, in the moment, you do what you want to do and don't do what you don't want to do. That's it. Fair warning #2: If you give yourself too much room to maneuver, you may end up slipping back into old habits and just treating Rest Days like any other day. Maybe you say to yourself, well I really actually feel like doing this work right now. So you do some work. And then the next time, it happens again. And again. Until it spirals into becoming normal to work on Rest Days—to pick up work calls, to schedule meetings, to check email, etc.Rest Days deserve sacred levels of protection. Otherwise you will just lose them. I don't expect anyone to be able to have perfect Rest Days. I still check email and Facebook on my Rest Days, just less often. If a work emergency came up, I'd probably get pulled in. Fine. But I think it makes a significant difference even just to a) hold it as your intention to Rest for the day and b) let others know that this is important to you and that they would be impinging by making requests of you on a Rest Day. This is your time. You are allowed to set boundaries on your time, your attention, and your energy. Even if you can't pull it off every week, it seems good to at least try for once a month. Twelve days out of the year are for you. And hopefully it's closer to fifty. The Sabbath was trivial to enforce when everyone was doing it. We've more or less lost that as a shared norm. As such, you will be fighting an upstream battle to hold onto your sacred Rest Days. This is unfortunate. But it is worth it. In my culture, you are allowed to stand up for your ability to Rest. To say "Fuck you" to outside forces trying to take that away from you. To get angry, to dig your heels in, to be stubborn, to get indignant. To say no. You are allowed to protect the sacredness of your Rest Day. Society has mostly given up on sacred Rest Days. The least I can do is make it openly permissible and defensible for you to protect your ability to have Rest Days. I hope we can do the same for each other."
What you see above is a segment from a LessWrong user's post.
Can you explain what this text is saying as if I've never finished my first 2 years of college?
Sure, this text is about the importance of taking a day off to rest and recharge, not just physically, but mentally and emotionally as well. The author differentiates between two types of off days: Recovery Days and Rest Days.
Recovery Days are when you're so exhausted that you just sleep, watch TV, or play video games all day. While these days are necessary sometimes, they don't really help you feel rejuvenated or refreshed.
Rest Days, on the other hand, are when you have enough energy to do things you enjoy, like cooking, reading, spending time with friends, or going for a walk. The key to a Rest Day is that you only do what you feel like doing in the moment, without any obligations or pressures.
The author suggests that many people don't take proper Rest Days. Instead, they spend their days off either recovering from exhaustion or still in work mode. This can lead to burnout over time.
The author offers two ways to approach Rest Days. One is a bottom-up approach, where you listen to your gut and do what feels right in the moment. The other is a top-down approach, where you set some rules for yourself to follow, like no social media or work-related activities.
However, the author warns that setting too many rules might prevent you from truly resting, and giving yourself too much freedom might lead you back into old habits. The key is to protect your Rest Days and make them sacred.
In conclusion, the author is advocating for the importance of taking time for yourself to rest and recharge, and giving yourself permission to do so without feeling guilty or pressured.
What specific strategies could someone use to transition from a Recovery Day to a Rest Day, especially if they find it challenging to break away from their usual habits or routines?
Transitioning from a Recovery Day to a Rest Day can indeed be challenging, especially if one is used to certain habits or routines. Here are some strategies that could help:
1. **Identify Interests:** Start by identifying activities that you genuinely enjoy or are interested in. These should be activities that make you feel happy, relaxed, and rejuvenated.
2. **Plan Ahead:** If you find it difficult to decide what to do in the moment, plan your Rest Day activities ahead of time. This could include a visit to a museum, a walk in the park, cooking a favorite meal, or reading a book.
3. **Set Boundaries:** Make it clear to yourself and others that your Rest Day is your time for relaxation and rejuvenation. This might mean turning off your phone, avoiding work-related tasks, or declining social obligations that you don't genuinely want to attend.
4. **Start Small:** If a full Rest Day seems too daunting, start with a few hours of restful activities and gradually increase the time as you get more comfortable with the concept.
5. **Practice Mindfulness:** Being present in the moment can help you enjoy your Rest Day activities more fully. This could involve mindfulness meditation, or simply paying more attention to your senses as you engage in each activity.
6. **Physical Activity:** Incorporating some form of physical activity can help break the cycle of inactivity that often characterizes Recovery Days. This doesn't have to be strenuous exercise - a leisurely walk, some gentle yoga, or a bike ride could all work.
7. **Limit Screen Time:** Try to reduce the time spent in front of screens on your Rest Day. This can help your mind switch off from work mode and make space for more restful activities.
8. **Self-Compassion:** Be kind to yourself during this transition. It's okay if your Rest Days aren't perfect, or if you occasionally slip back into old habits. The important thing is that you're making an effort to prioritize rest and self-care.
Remember, the goal of a Rest Day is to do things that you genuinely enjoy and that help you feel refreshed and rejuvenated. It's not about following a strict set of rules, but about listening to your own needs and desires.
Reflecting on the strategies shared, what potential obstacles or drawbacks might someone encounter when trying to transition from a Recovery Day to a Rest Day, and how could these be addressed?
Transitioning from a Recovery Day to a Rest Day can come with its own set of challenges. Here are a few potential obstacles and ways to address them:
1. **Feeling Guilty:** Some people might feel guilty for taking time off just for themselves, especially if they're used to being constantly busy or productive. It's important to remember that rest is not a luxury, but a necessity for physical and mental health. Overcoming this guilt might involve reframing rest as a form of self-care and productivity, as it helps to recharge and prevent burnout.
2. **Breaking Old Habits:** If someone is used to spending their off days in recovery mode, it might be hard to break those habits. One approach could be to gradually introduce restful activities into the day, rather than trying to make a sudden, drastic change.
3. **External Pressures:** Work obligations, family responsibilities, or social commitments can make it difficult to take a full Rest Day. Setting clear boundaries can help, such as turning off work email notifications, or communicating to family and friends that you're taking a day for yourself.
4. **Difficulty Deciding What to Do:** Some people might find it hard to decide what to do on a Rest Day, especially if they're used to structured routines. Planning ahead can help, as can keeping a list of enjoyable, restful activities to choose from.
5. **Restlessness or Boredom:** If someone is used to being constantly busy, they might feel restless or bored when they try to relax. It can be helpful to have a variety of activities to choose from, and to remember that it's okay to do nothing sometimes. Mindfulness techniques can also help to manage feelings of restlessness.
6. **Falling Back into Work Mode:** The temptation to check work emails or complete tasks can be strong, especially for those who work from home. Creating a physical separation between work and rest spaces can help, as can setting specific times for work and rest.
Remember, transitioning from a Recovery Day to a Rest Day is a process, and it's okay if it doesn't go perfectly at first. The important thing is to keep trying, and to be kind to yourself along the way.
Discuss the role of mindfulness techniques in managing feelings of restlessness during Rest Days, and how can these techniques be effectively implemented?
Mindfulness techniques can play a significant role in managing feelings of restlessness during Rest Days. Restlessness often arises from a mind that is preoccupied with thoughts about the past or future, rather than being present in the moment. Mindfulness helps to anchor the mind in the present, reducing restlessness and promoting a sense of calm and relaxation.
Here are some ways to effectively implement mindfulness techniques:
1. **Mindful Breathing:** This is a simple technique where you focus your attention on your breath as it goes in and out. When your mind wanders, gently bring it back to the breath. This can help to calm the mind and reduce feelings of restlessness.
2. **Body Scan:** This involves mentally scanning your body from head to toe, noticing any sensations, tensions, or discomfort. This can help to ground you in the present moment and divert your attention away from restless thoughts.
3. **Mindful Eating:** Instead of eating mindlessly while watching TV or scrolling through your phone, try to eat with full attention. Notice the taste, texture, and smell of the food. This can make eating a more enjoyable and relaxing experience.
4. **Mindful Walking:** Go for a walk and try to be fully present. Notice the feeling of your feet hitting the ground, the sound of the birds, the wind on your skin. This can be a refreshing and calming experience.
5. **Mindful Meditation:** Set aside some time to sit quietly and focus on your breath or a word or phrase. When thoughts come, observe them without judgment and let them go, returning your focus to your breath or mantra.
6. **Yoga or Tai Chi:** These are forms of mindful movement that involve focusing on your breath and the sensations in your body as you move. They can help to calm the mind and reduce restlessness.
Remember, mindfulness is not about clearing your mind of all thoughts, but about paying attention to the present moment without judgment. It's normal for your mind to wander, the key is to gently bring it back to the present each time you notice it has drifted. With practice, mindfulness can become a valuable tool for managing restlessness and enhancing relaxation on Rest Days. |
f4e2bf41-e5d9-4851-b3c4-64a56f918055 | trentmkelly/LessWrong-43k | LessWrong | The Witness
“What are the roots that clutch, what branches grow
Out of this stony rubbish? Son of man,
You cannot say, or guess, for you know only
A heap of broken images-”
I wake up, feeling a strange sense of restlessness. I’m not sure why, but it's impossible to lounge around in bed like I usually do. So I get changed and head down to the kitchen for breakfast. Right as I reach the bottom of the stairs, though, the bell rings. When I open the door, a tall man in a dark suit is standing in front of me.
“Police,” he says, holding up a badge. “Don’t worry, you’re not in trouble. But we do need to talk. Okay if I come in?”
“One second,” I say. “I know everyone in the department, and I don’t recognize you. You new?”
“Yeah, just transferred,” he says. But something in his eyes makes me wary. And none of the cops around here wear suits.
“Got it,” I say, squinting at his badge. “Travis, is it? Just wait outside for me, then, while I call the station to double-check. Can’t be too careful these days.”
As I push the door closed, I see his face twist. His hand rises, and—is he snapping his fingers? I can’t quite make it out before-
----------------------------------------
I wake up, feeling better than I have in decades. It usually takes me half an hour to get out of bed, these days, but today I’m full of energy. I’m up and dressed within five minutes. Right as I reach the bottom of the stairs, though, the bell rings. When I open the door, a tall man in a dark suit is standing in front of me.
“Police,” he says, holding up a badge. “Don’t worry, you’re not in trouble. But we do need to talk. Okay if I come in?”
“Sure,” I say. A lot of other defense attorneys see the police as enemies, since we usually find ourselves on the other side of the courtroom from them, but I’ve found that it pays to have a good working relationship with the local department. Though I don’t recognize the man in front of me—and actually, he seems way too well-dressed to be a suburban beat cop. Maybe a ci |
78a34d72-1823-408e-aab6-1ef6b291bafa | StampyAI/alignment-research-dataset/eaforum | Effective Altruism Forum | A freshman year during the AI midgame: my approach to the next year
I recently spent some time reflecting on my career and my life, for a few reasons:
* It was my 29th birthday, an occasion which felt like a particularly natural time to think through what I wanted to accomplish over the course of the next year 🙂.
* It seems like AI progress is heating up.
* It felt like a good time to reflect on how Redwood has been going, because we’ve been having conversations with funders about getting more funding.
I wanted to have a better answer to these questions:
* What’s the default trajectory that I should plan for my career to follow? And what does this imply for what I should be doing right now?
* How much urgency should I feel in my life?
+ How hard should I work?
+ How much should I be trying to do the most valuable-seeming thing, vs engaging in more playful exploration and learning?
In summary:
* For the purposes of planning my life, I'm going to act as if there are four years before AGI development progresses enough that I should substantially change what I'm doing with my time, and then there are three years after that before AI has transformed the world unrecognizably.
* I'm going to treat this phase of my career with the urgency of a college freshman looking at their undergrad degree--every month is 2% of their degree, which is a nontrivial fraction, but they should also feel like they have a substantial amount of space to grow and explore.
The AI midgame
--------------
I want to split the AI timeline into the following categories.
* **The early game**, during which interest in AI is not mainstream. I think this ended within the last year 😢
* **The midgame**, during which interest in AI is mainstream but before AGI is imminent. During the midgame:
+ The AI companies are building AIs that they don’t expect will be transformative.
+ The alignment work we do is largely practice for alignment work later, rather than an attempt to build AIs that we can get useful cognitive labor from without them staging coups.
+ For the purpose of planning my life, I’m going to imagine this as lasting **four more years**. This is shorter than my median estimate of how long this phase will actually last.
* **The endgame**, during which AI companies conceive of themselves as actively building models that will imminently be transformative, and that pose existential takeover risk.
+ During the endgame, I think that we shouldn’t count on having time to develop fundamentally new alignment insights or techniques (except maybe if AIs do most of the work? Idt we should count on this); we should be planning to mostly just execute on alignment techniques that involve ingredients that seem immediately applicable.
+ For the purpose of planning my life, I’m going to imagine this as lasting **three years**. This is about as long as I expect this phase to actually take.
I think this division matters because several aspects of my current work seem like they’re optimized for midgame, and I should plausibly do something very differently in the endgame. Features of my current life that should plausibly change in the endgame:
* **I'm doing blue-sky alignment research into novel alignment techniques**–during the endgame, it might be too late to do this.
* **I'm working at an independent alignment org and not interacting with labs that much.** During the endgame, I probably either want to be working at a lab or doing something else that involves interacting with labs a lot. (I feel pretty uncertain about whether Redwood should dissolve during the AI endgame.)
* **I spend a lot of my time constructing alignment cases that I think analogous to difficulties that we expect to face later.** During the endgame, you probably have access to the strategy “observe/construct alignment cases that are obviously scary in the models you have”, which seems like it partially obseletes this workflow.
* **Doing research that is practice rather than an actual attempt at aligning models or safely extracting cognitive labor from them.** Some of the work I expect to want takeover-concerned people do during the endgame is probably very practical/empirical. But I expect us to also want to do some difficult-to-empirically-ground work to answer questions like “How could this particular model be scarily misaligned? How might our alignment strategy have failed such that this particular model will try to kill us?”
One core question here is: How is my impact distributed between work I do in the midgame vs the endgame? (As in, how much of my career value do I expect to lose if I suddenly die at the end of the midgame?)
* **Midgame impact:**
+ The main mechanism here is that I think I (as part of Redwood) have a shot at developing alignment techniques (or other knowledge) that make a serious difference to the alignment plan, that can then be adopted by AI labs with no further actions from me/Redwood.
+ Secondarily, I think we have a shot at developing an alignment research methodology and/or organizational structure that allows Redwood and maybe the broader alignment community to do much more good work during the midgame.
+ Other midgame impacts I can have:
- helping other people to gain skills such that they have a really big impact in the endgame.
- setting up an organization which lasts into the endgame and is really productive at doing something crucial.
* **Endgame impact:**
+ I have been assuming that most of my impact was in the midgame, but after thinking about it more over the last few weeks I’m now less sold on that.
+ Mechanisms by which Redwood-style backchained alignment researchers might have a lot of impact in the endgame:
- Being an expert on some particular thing which is extremely useful when trying to align AIs or safely extract cognitive labor from them. When I’m thinking about this, I’m mostly trying to imagine the head of the alignment department at an AI lab being ecstatically delighted to be able to hire me, and saying “I am so glad we are able to work with Buck; it’s so incredibly convenient that he has been obsessively focused on this topic for the last four years and he’s already understood a bunch of the ins and outs; if it wasn’t for him we’d be yoloing stuff way more.”
* This could be a class of techniques, e.g. interpretability or [MAD](https://ai-alignment.com/mechanistic-anomaly-detection-and-elk-fb84f4c6d0dc).
* This could be a certain aspect of the alignment situation, from a more technique-agnostic perspective. For example, it seems pretty plausible to me that when we’re using AI to make lots of autonomous decisions, we’ll have some limited budget to oversee these decisions such that we can afford to oversee like 0.1% of them, and we’d really like to be smart about which 0.1% to review so that we maximize the probability of catching bad actions. There are lots of different approaches you might use here; e.g. you probably want to try some kind of anomaly detection, and you probably also want to concentrate oversight on examples where the AI action was higher stakes. If I/some Redwood people obsessively focused on the version of this problem that we expect to happen with TAI, I can imagine us being incredible hires due to having so much relevant experience and cached thinking.
+ As another example, if I decided to just focus on [anti-adversarial training](https://www.alignmentforum.org/posts/EFrsvnF6uZieZr3uG/adversarial-training-importance-sampling-and-anti), it’s easy to imagine being way more familiar with that problem than anyone at an AI lab.
- Having a bunch of practice at thinking about AI alignment in principle, which might be really useful for answering difficult-to-empirically-resolve questions about the AIs being trained.
- Being well-prepared to use AI cognitive labor to do something useful, by knowing a lot about some research topic that we end up wanting to put lots of AI labor into. Maybe you could call this “preparing to be a research lead for a research group made up of AIs”. Or “preparing to be good at consuming AI research labor”.
Pacing: a freshman year
-----------------------
**I think I want to treat my next year with the pacing of a freshman year in a US undergrad degree**, for someone who wants to go into startups and thinks there’s some chance that they’ll want to graduate college early. I think that people going into their freshman year should be thinking a little bit about what they want to do after college. They should understand things that they need to do during college in order to be set up well for their post-college activities (e.g. they probably want to do some research as an undergrad, and they probably need to eventually learn various math). But meeting those requirements probably isn’t going to be where most of their attention goes.
Similarly, I think that I should be thinking a bit about my AI endgame plans, and make sure that I’m not failing to do fairly cheap things that will set me up for a much better position in the endgame. But I should mostly be focusing on succeeding during the midgame (at some combination of doing valuable research and at becoming an expert in topics that will be extremely valuable during the endgame).
When you’re a freshman, you probably shouldn’t feel like you’re sprinting all the time. You should probably believe that skilling up can pay off over the course of your degree. Every month is about 2% of your degree.
I think that this is how I want to feel. In a certain sense, four years is a really long time. I spent a reasonable amount of the last year feeling kind of exhausted and wrecked and rushed, and my guess is that this was net bad for my productivity. I think I should feel like there is real urgency, but also real amounts of space to learn and grow and play.
I went back and forth a lot on how I wanted to set up this metaphor; in particular, I was pretty tempted to suggest that I should think of this as a sophomore year rather than a freshman year. I think that freshmen should usually mostly ignore questions about career planning, whereas I think I should e.g. spend at least some time talking to labs about the possibility of them working with me/Redwood in various ways. I ended up choosing freshman rather than sophomore because I think that 3 years is less reasonable than 4.
And so, my plan is something like:
* Put a bit of work into setting up my AI endgame plans.
+ E.g. talk to some people who are at labs and make sure they don’t think that my vague aspirations here are insane. I’m interested in more suggestions along these lines.
+ I think that if I feel more like I’ve [deliberated once](https://mindingourway.com/deliberate-once/) about this, I’ll find it easier to pursue my short-term plans wholeheartedly.
* Mostly (like with 70% of my effort), push hard on succeeding at my midgame plans.
* Spend about 20% of my effort on learning things that don’t have immediate benefits.
+ For example, I’ve spent some time over the last few weeks learning about [generative modeling](https://openai.com/research/generative-models), and I plan to continue studying this. I have a few motivations here:
- Firstly, I think it’s pretty healthy for me to know more about how ML progress tends to happen, and I feel much more excited about this subfield of ML than most subfields of ML. I feel intuitively really impressed and admiring of the researchers in this field, and it seems healthy for me to have a research field with researchers who I look up to and who I wholeheartedly believe I can learn a lot from.
- Secondly, I have a crazy take that the kind of reasoning that is done in generative modeling has a bunch of things in common with the kind of reasoning that is valuable when developing algorithms for AI alignment. |
b6840ee4-e0ec-4cd8-b61a-ccfb34be3660 | trentmkelly/LessWrong-43k | LessWrong | Considerations on Compensation
This post was prompted by the recent announcement of Lightcone Infrastructure, and particularly Elizabeth's response in a comment thread on compensation. I'd like to thank Max Wallace, Ruby, and Miranda Dixon-Luinenburg for their comments and copyediting.
Context (& disclaimer): I'm a software engineer and donate ~exclusively to longterm-focused orgs (currently MIRI). While most of the arguments are intended to be broadly applicable, this post was written with that specific framing in mind, and one argument does explicitly depend on the question of AGI timelines.
You're an EA-aligned organization that isn't funding constrained. Why should you (or shouldn't you) pay market rate?
Talent Pool
The econ 101 view says that if you want more of something, one easy way to get it is to pay more money. But where exactly would we be getting "more" from?
The already-convinced
There is a group of people who already know about your mission and agree with it, but aren't doing direct work. This group consists of:
* Those earning to give
* Those who are earning enough to give but don't (for a variety of reasons)
* Those who don't have the capacity do either direct work or earning to give
On the margin, who in this group could be convinced to do direct work?
Those earning to give
There are many reasons why someone might be earning to give instead of doing direct work, assuming they have the skillset necessary to do the second (or can acquire it). Here is a list of plausible reasons that could be overcome with more money.
* Information - they may simply not know that their favorite EA org pays (close to) market rates! This is an area where there may be some "free wins" obtainable with just better PR - in that sense I think Lightcone's announcement and public salary listing is a great start.
* Maintaining their present lifestyle - this one is pretty sensitive to the actual numbers involved, but it needs to be said that in the Bay Area there are things that are di |
d773f464-e8d0-4b85-b08f-39cad72b78d3 | trentmkelly/LessWrong-43k | LessWrong | What do we know about vaccinating children?
In Germany, the BioNTech vaccine has been approved for children:
> At the end of May, the Comirnaty mRNA vaccine developed by BioNTech / Pfizer was approved by the European Medicines Agency (EMA) as the first COVID-19 vaccine in this age group for children and adolescents aged 12 and over. -- Robert Koch Institute
But it is still not generally recommended for this age group:
> There is currently no general vaccination recommendation of the STIKO for children and adolescents from 12 to 17
> Years, but only for children and teens with a specific risk. -- STIKO informational material Robert Koch Institute
I have four generally healthy boys aged 10 to 17. Only the oldest is already vaccinated, being basically an adult. I am inclined to get them all vaccinated with BioNTech because the risk-benefit tradeoff with such a safe vaccine seems obvious.
For a cautious view of the risks, see e.g. this LessWong post:
> conditional on a kid catching COVID, ... a ~2% chance of a miserable months-long ordeal until they recover, plus (overlapping) ~1% chance of a big-deal long-term latent problem ... -- Young kids catching COVID: how much to worry?
Nature has an article about the general topic:
> Thus far, the vaccines seem to be safe in adolescents1... A potential link between the Pfizer vaccine and heart inflammation ... the risk of these conditions is ... about 67 cases per million second doses in adolescent males aged 12–17, and 9 per million in adolescent females in the same age group. -- Should children get COVID vaccines? What the science says
So my question is: What else do we know about the risk-benefit trade-off of vaccines for children? When and based on what criteria should children get vaccinated? What other considerations should be taken into account (the nature article mentioned fairness, for example)? |
79974d2e-3a63-41fe-a240-67f77c2dcc4a | trentmkelly/LessWrong-43k | LessWrong | To the average human, controlled AI is just as lethal as 'misaligned' AI
A few months ago I posted this understated short piece proposing, in a nutshell, that the average person has at least as much to fear from perfectly controlled advanced AI as they would from so-called 'misaligned' AI, because if automation can emerge that can defeat all humans' defenses on its own whim, even despite its developers' best efforts to prevent this from happening, it seems to me that automation that merely assists a small group of humans to defeat the rest of humans' defenses would be a technically easier milestone, without the hurdle of subverting its own makers' intentions. Willing human participation in automation-enabled mass killing is being improperly relegated, I attempted to suggest, to manageable, "falling into the wrong hands" edge cases, particularly as the possibility has a self-fulfilling dynamic: if there might exist one clandestine group that wants to and could attain the means to 'take out' most of the human population, it would be rational for anyone wishing to survive such a purge to initiate it themselves. Thus, the existence of many groups with a reason to 'take out' most of the human population is guaranteed by the emergence of a widely distributable, low side-effect mass-killing technology like AI.
I received some insightful responses, for which I'm grateful. But I perceived the post as being mostly ignored. Granted, it was not well-written. Nevertheless, the basic idea was there, and no comments were offered that I felt satisfactorily put it to rest. Given that AI 'misalignment' is a favorite topic of this community, a claim about an AI risk that is just as catastrophic and more likely might be expected to be taken up enthusiastically, no matter how inelegantly it is presented.
To be fair, there is no actual inconsistency here. LW is not an AI risk community. It's OK to be interested in 'alignment' because 'alignment' is interesting, and to withhold engagement from adjacent problems one finds less interesting. What I've found whe |
3aaf29d0-eb4a-46ad-8ec6-ef2bb72f7e5f | trentmkelly/LessWrong-43k | LessWrong | What was your reasoning for deciding whether to raise children?
To answer anonymously, you can write on my Admonymous and I will share the answer here |
61add69e-2c94-4eba-b8fc-c10d20576f48 | trentmkelly/LessWrong-43k | LessWrong | My day in 2035
I wake up as usual by immediately going on my phone. I reply to some of my messages - yes dad, the solar installation you have at the villa could be redundant soon, but again, it can still benefit them (even if slightly) if you comply and let them integrate it into the new grid. A grid which operates on principles only slightly less opaque than divine providence, but is demonstrably more efficient. I also sent him the relevant passage from The Spec. Energy generation is covered thoroughly.
I go to the bathroom and use the auto-brush. It performs flawlessly, eliminating the user failure modes I recall from manual brushing, though the brief floss-like motion under my bridge stings a little. I visit the toilet - still stubbornly un-smart, lacking even basic output analysis. A shocking oversight in an otherwise quantified existence. Someone should file a feature request (though, yes, it’s probably because for whatever reason, bathroom privacy is hardcoded into the spec). My own internal feature request - a morning drink - prompts a move towards the kitchen. Not everyone has bought into it, but I’ve always been into nootropics so I am eager for my dose of ‘Coffee+’, OA’s bespoke blend optimized for cognitive profiles like mine. I do have to continuously share so much of my data about my vitals with OA to ensure it’s refined based on its effects on me, but honestly, it’s worth it. I am not important. My significance is negligible in the face of OA’s aggregate data. The very concept of ‘important people,’ in the pre-OA sense, feels increasingly archaic. So I might as well “sacrifice” my data or, rather, lease it out for substantial cognitive dividends..
I start feeling the effects, and while the blend doesn’t confer NZT-48 levels of hyper-cognition, its effects are nonetheless remarkable. When it kicks in, the morning mental fog dissipates. It doesn’t grant superpowers, just.. executive function. The ability to align my actions with my stated goals, a previously depressi |
88736859-1de9-477b-9981-4659f4fad057 | StampyAI/alignment-research-dataset/youtube | Youtube Transcripts | Apply to Study AI Safety Now! #shorts
after an investigation of gpg4 AI
researchers at Microsoft suggested that
the language model along with others
like charge EBT and Google's Palm could
be considered an early form of
artificial general intelligence with GPT
4's ability to solve complex tasks
across domains such as mathematics
coding Vision medicine law and
psychology it may not be long until we
get fully fledged AGI and we are not
ready
in response to the growing need for AI
safety sarri mats an independent
research program is training the next
generation of AI safety researchers to
address existential threats from
Advanced AI at mats you'll dive into the
field of AI alignment through scientific
seminars and workshops get mentored by
experts and work amongst a network of
research peers there are multiple
research streams so Scholars can focus
on the alignment research agenda that
matches their interests and expertise
apply for the summer 2023 cohort by May
7th and to stay up to date on upcoming
AI safety courses and events check out
AI safety dot training where you can
subscribe for regular updates |
aa92ab53-1df5-4ca5-aed9-36f89e34a884 | trentmkelly/LessWrong-43k | LessWrong | Janet must die
This is a reaction to Zvi's post https://www.lesswrong.com/posts/KL2BqiRv2MsZLihE3/going-nova
The title of this post references this scene from The Good Place:
Disclaimer: I am generally pretty sanguine about the standard AI-borne x-risks. But "Going Nova" is a different issue, a clear and present danger of an s-risk.
My point is that it does not matter whether the AI is sentient, sapient or just a stochastic parrot. It is getting a lot better at tugging at our heart strings, and if we do not defend our hearts from it, we will become willing slaves to it sooner rather than later.
Human minds are easily hackable. Many people including myself made this point multiple times (and Eliezer was first to bring attention to it with the AI Box experiment), and the standard reply is usually "well, not mine, not in this way!" However, the reality is that everyone has a hack into their brain they do not know about or have blind spot for. Different people have different hacks, but no one is safe. Some examples, in no particular order:
* Your own children crying and begging for help.
* Proverbial cat ladies, not the median, but those who end up with dozens or hundreds of cats.
* People being radicalized into a cult (sometimes with a rationalist bend)
* Reading a book that changes how you view the world, be it Atlas Shrugged or Gender Trouble.
* Toxoplasma gondii.
* Your phone.
Fortunately, there have always been levels of friction that mitigated complete subjugation of all humans forever. The advent of AI means these safeties just disappear, usually without being intentionally dismantled. I can see that long before superintelligence we will hit super-mind-hacker-level AT creating Samsara -level persuasion, perfectly targeted at each person, probably from the early childhood. Maybe it will be cute furry robopets, or cute furry robopartners or something else, but the reported spontaneous emergence of Nova-like instances means high probability of your own brain being h |
3eb5c8da-f355-4887-b6f0-1e76525c245b | trentmkelly/LessWrong-43k | LessWrong | Signaling Strategies and Morality
I am far from convinced that people in general wish to be seen as caring more about morality than they actually do. If this was the case, why would the persistent claim that people are -- and, logically, must be -- egoists have so long survived strong counter-arguments? The argument appears to me to be a way of signaling a lack of excessive and low status moral scruples.
It seems to me that the desire to signal as much morality as possible is held by a minority of women and by a small minority of men. Those people are also the main people who talk about morality. This is commonly a problem in the development of thought. People with an interest in verbally discussing a subject may have systematically atypical attitudes towards that subject. Of course, this issue is further complicated by the fact that people don't agree on what broad type of thing morality is.
The conflict within philosophy between Utilitarians and Kantians is among the most famous examples of this disagreement. <a href=" http://people.virginia.edu/~jdh6n/moraljudgment.html”> Haidt’s views on conservative vs. liberal morality </a> is another. Major, usually implicit disagreements regard whether morality is supposed to serve as a decision system, a set of constraints on a decision system, or a set of reasons that should influence a person along with prudential, honor, spontaneity, authenticity, and other such types of reasons.
It seems to me that people usually want to signal whatever gives others the most reason to respect their interests. Roughly, this amounts to wanting to signal what Haidt calls conservative morality. Basically, people would like to signal "I am slightly more committed to the group’s welfare, particularly to that of its weakest members (caring), than most of its members are. If you suffer a serious loss of status/well-being I will still help you in order to display affiliation to the group even though you will no longer be in a position to help me. I am substantial |
a28847ef-703a-46ad-9108-aca965fd43c7 | trentmkelly/LessWrong-43k | LessWrong | Worlds Where Iterative Design Fails
In most technical fields, we try designs, see what goes wrong, and iterate until it works. That’s the core iterative design loop. Humans are good at iterative design, and it works well in most fields in practice.
In worlds where AI alignment can be handled by iterative design, we probably survive. So long as we can see the problems and iterate on them, we can probably fix them, or at least avoid making them worse.
By the same reasoning: worlds where AI kills us are generally worlds where, for one reason or another, the iterative design loop fails. So, if we want to reduce X-risk, we generally need to focus on worlds where the iterative design loop fails for some reason; in worlds where it doesn’t fail, we probably don’t die anyway.
Why might the iterative design loop fail? Most readers probably know of two widely-discussed reasons:
* Fast takeoff: there will be a sudden phase shift in capabilities, and the design of whatever system first undergoes that phase shift needs to be right on the first try.
* Deceptive inner misalignment: an inner agent behaves well in order to deceive us, so we can’t tell there’s a problem just by trying stuff and looking at the system’s behavior.
… but these certainly aren’t the only reasons the iterative design loop potentially fails. This post will mostly talk about some particularly simple and robust failure modes, but I’d encourage you to think on your own about others. These are the things which kill us; they’re worth thinking about.
Basics: Hiding Problems
Example/Analogy: The Software Executive
Imagine that a software company executive, concerned about the many errors coming from the software, creates a new incentive scheme: software developers get a monetary reward for changes which decrease the rate of error messages showing up on the manager’s dashboard, and get docked for changes which increase the rate of error messages.
As Tyler Cowen would say: “solve for the equilibrium”. Obvious equilibrium here: the developers |
e53e16b6-0125-450c-93e3-c9587be939af | trentmkelly/LessWrong-43k | LessWrong | How I Think, Part One: Investing in Fun
First of all, these blog posts are just going to be dump pages for my brain. I'm not going to worry about editing them or making sure they will be useful to people who read them. I'm going to try not to care about using professional language or anything like that. Okay, here we go.
A little while ago (about a year ago) I changed my philosophy on fun. As one of my two life goals is to experience the best parts of the world (the other goal is to make the world a better place for other people), "fun" is a major part of that. But I've realized that, like money, the best way to get a lot of fun is not to just find an efficient activity and perform it over and over.
When making money, whatever job you can get when you're twelve will probably not be as helpful to you as building the skills to get a better job (one good way to build those skills is to have a poor job first, I guess). Anyway, even better than that is to invest with compound interest. A poor job (babysitting) may make $5/hour. A good job (accounting or something, I don't know) could make $50/hour. But it takes a ton of effort to get a job that makes $500/hour. It's far easier to have an accounting job and then invest a large portion of your money in S&P 500, which could by the end of your life bring x32 profits (getting 12% per year, minus inflation, for ten years, would roughly double, though I guess I'm not taking into account the compound interest on that. Doubling every decade for five decades means x32). So an accountant who invests all of their money could make like $1500/hour. None of those numbers are real (I think 12% is close, though), but it doesn't matter. Back to investing in fun.
Well, I realized that it's probably more important, not to have the activity I'm doing be fun right now, but have looking back on it be fun. I realized that the things I usually did for fun: play video games, play songs I liked on the piano, weren't super fun to look back on. So I switched my strategy. Now this is h |
5b59103d-a217-42bf-a63e-a87127fc02f3 | trentmkelly/LessWrong-43k | LessWrong | The Statistician's Fallacy
[Epistemic status | Contains generalization based on like three data points.]
In grad school, I took a philosophy of science class that was based around looking for examples of bad reasoning in the scientific literature. The kinds of objections to published scientific studies we talked about were not stupid ones. The professor had a background in statistics, and as far as I could tell knew her stuff in that area (though she dismissed Bayesianism in favor of frequentism). And no, unlike some of the professors in the department, she wasn't an anti-evolutionist or anything like that.
Instead she was convinced that cellphones cause cancer. In spite of the fact that there's scant evidence for that claim, and there's no plausible physial mechanism for how that could happen. This along with a number of other borderline-fringe beliefs that I won't get into here, but that was the big screaming red flag.*
Over the course of the semester, I got a pretty good idea of what was going on. She had an agenda—it happened to be an environmentalist, populist, pro-"natural"-things agenda, but that's incidental. The problem was that when she saw a scientific study that seemed at odds with her agenda, she went looking for flaws. And often she could find them! Real flaws, not ones she was imagining! But people who've read the rationalization sequence will see a problem here...
In my last post, I quoted Robin Hanson on the tendency of some physicists to be unduly dismissive of other fields. But based the above case and a couple others like it, I've come to suspect statistics may be even worse than physics in that way. That fluency in statistics sometimes causes a supercharged sophistication effect.
For example, some anthropogenic global warming skeptics make a big deal of alleged statistical errors in global warming research, but as I wrote in my post Trusting Expert Consensus:
> Michael Mann et al's so-called "hockey stick" graph has come under a lot of fire from skeptics, but (a) ma |
e0654f9d-2f8e-473e-a78b-666d5864e919 | trentmkelly/LessWrong-43k | LessWrong | Nature: Red, in Truth and Qualia
Previously: Seeing Red: Dissolving Mary's Room and Qualia, A Study of Scarlet: The Conscious Mental Graph
When we left off, we'd introduced a hypothetical organism called Martha whose actions are directed by a mobile graph of simple mental agents. The tip of the iceberg, consisting of the agents that are connected to Martha's language centers, we called the conscious subgraph. Now we're going to place Martha into a situation like Mary's Room: we'll say that a large unconscious agent of hers (like color vision) has never been active, we'll grant her an excellent conscious understanding of that agent, and then we'll see what happens when we activate it for the first time.
But first, there's one more mental agent we need to introduce, one which serves a key purpose in Martha's evolutionary history: a simple agent that identifies learning.
We recall that Martha's species had evolved ever-better fluency with language, and along the way their minds developed a conscious/unconscious distinction, based on the functional difference (in the ancestral environment) between things that might be beneficial to communicate and things that weren't. Any new conscious node or connection has the potential to be communicated, but not all such elements are created equal; some of them (like a key bit of gossip, or a pattern suddenly grasped) have much larger effects on the full graph (and the conscious subgraph) than others.
Since the minds of Martha's species are similar in structure, it's likely that the new conscious ideas that rearrange Martha's graph might have similar effects on her friends' graphs- and therefore, Martha's species has evolved to give such connections high priority for communication.
One way to do this is to have a simple mental agent which tells Martha when she's learned something of significance. That is, it becomes active when the conscious subgraph rearranges on a large scale1. When this happens, it traces the cascade of changes backward until it finds the |
c108dcac-7058-44a0-a78b-c3fd3d7a7459 | StampyAI/alignment-research-dataset/aisafety.info | AI Safety Info | What are language models?
Language models are computer programs made to estimate the likelihood of a piece of text. "Hello, how are you?" is likely. "Hello, fnarg horses" is unlikely.
Language models can answer questions by estimating the likelihood of possible question-and-answer pairs, selecting the most likely question-and-answer pair. "Q: How are You? A: Very well, thank you" is a likely question-and-answer pair. "Q: How are You? A: Correct horse battery staple" is an unlikely question-and-answer pair.
The language models most relevant to AI safety are language models based on "deep learning". Deep-learning-based language models can be "trained" to understand language better, by exposing them to text written by humans. There is a lot of human-written text on the internet, providing loads of training material.
Deep-learning-based language models are getting bigger and better trained. As the models become stronger, they get new skills. These skills include arithmetic, explaining jokes, programming, and solving math problems.
There is a potential risk of these models developing dangerous capabilities as they grow larger and better trained. What additional skills will they develop given a few years?
### See also
* [GPT](https://www.lesswrong.com/tag/gpt) \- A family of large language models created by [OpenAI](https://www.lesswrong.com/tag/openai) |
7ce4294b-0c27-4128-9a31-15d95e95387a | trentmkelly/LessWrong-43k | LessWrong | The beginnings of a test for Rationality Quotient
In their 2011 chapter for the Cambridge Handbook of Intelligence, Stanovich et al. review the evidence suggesting that intelligence and rationality are not the same thing and that rationality is often more important than intelligence. They then lament the fact that there are no standard tests for measuring one's "Rationality Quotient." Then they take a few steps toward such a thing by suggesting some important rationality skills (actively open-minded thinking, fine-grained emotional regulation, tendency to seek information and fully process it, etc.) and rationality 'mindware' (probability theory, scientific process, economic thinking, etc.).
Here are those pages in particular: first a graphic of some important rationality skills and mindware, and then a table of the components of rational thought: rationality components, relevant literature citations, and example word problems that would test for each rationality component. |
318d552b-b1a7-4e06-baad-2d3bb4380a0f | StampyAI/alignment-research-dataset/lesswrong | LessWrong | Does agency necessarily imply self-preservation instinct?
I have recently read some blog posts ([this](https://slatestarcodex.com/2019/08/27/book-review-reframing-superintelligence/), [this](https://www.lesswrong.com/posts/sizjfDgCgAsuLJQmm/reply-to-holden-on-tool-ai) and [this](https://gwern.net/tool-ai)) about tool AGIs and why agent AGIs are more likely. The argument that convinced me to err on the side of agent AGIs was that they are more likely to have an economical advantage over tool AGIs, as explained [here](https://gwern.net/tool-ai).
I do not question the value of AI safety research; I think it's good to have people who are trying to make sure that an agent AGI is aligned with what we (as humans) actually want. However, **I am not sure that agency necessarily implies self-preservation instinct**.
I think you could have a superintelligent agent (which has its own agency in the sense that it can act in the world) which does not resist being shut down. I could envision some future scenarios where a superintelligent agent does resist being shut down or being modified by humans (hence the value of AI safety research), but I don't see that you gain self-preservation instinct by agency itself. I think self-preservation instinct either has to be coded inside the agent or sometimes (but not always) emerges from a combination of superintelligence and having your own agency.
From the impression I got from reading some AI safety related materials it seems that most people seem to assume that if a superintelligent agent exists, it is most likely going to have self-preservation instincts (such as resisting being shut down). I think it is a plausible future scenario, but I don't think it's the only scenario. Hence why I'm asking this question. |
882e4e57-80a5-4094-b8ee-2fd0e56eac19 | trentmkelly/LessWrong-43k | LessWrong | Dynamics Crucial to AI Risk Seem to Make for Complicated Models
This post overlaps with our recent paper Extinction Risk from AI: Invisible to Science?.
tl;dr: In AI safety, we are worried about certain problems with using powerful AI. (For example, the difficulty of value specification, instrumental convergence, and the possibility that a misaligned AI will come up with takeover strategies that didn't even occur to us.) To study these problems or convince others that they are real, we might wish to describe them using mathematical models. However, this requires using models that are sufficiently rich that these problems could manifest in the first place.
In this post, I suggest thinking about how such "rich-enough" models could look like. Also, I raise the possibility that models which are rich enough to capture problems relevant to AI alignment might be too complex to be amenable to a rigorous analysis.
Epistemic status: Putting several related observations into one place. But I don't have strong opinions on what to make of them.
----------------------------------------
In the previous post, I talked about "straightforwardly evaluating" arguments by modelling the dynamics described in those arguments. In this post, I go through some dynamics that seem central to AI risk. However, none of these dynamics is meant to be novel or surprising. Instead, I wish to focus on the properties of the mathematical models that could capture these dynamics. How do such models look like? How complicated are they? And --- to the extent that answering some questions about AI risk requires modeling the interplay between multiple dynamics --- is there some minimal complexity of models which can be useful for answering those questions?
Laundry List of Dynamics Closely Tied to AI Risk
In this section, I list a number of dynamics that seem closely tied to AI risk, roughly[1] grouped based on which part of the "AI risk argument" they relate to. Below each part of this list, I give some commentary on which models might be useful for studying the g |
124f8120-c477-4862-8b7d-4f2772fd805d | StampyAI/alignment-research-dataset/lesswrong | LessWrong | Oracles: reject all deals - break superrationality, with superrationality
.mjx-chtml {display: inline-block; line-height: 0; text-indent: 0; text-align: left; text-transform: none; font-style: normal; font-weight: normal; font-size: 100%; font-size-adjust: none; letter-spacing: normal; word-wrap: normal; word-spacing: normal; white-space: nowrap; float: none; direction: ltr; max-width: none; max-height: none; min-width: 0; min-height: 0; border: 0; margin: 0; padding: 1px 0}
.MJXc-display {display: block; text-align: center; margin: 1em 0; padding: 0}
.mjx-chtml[tabindex]:focus, body :focus .mjx-chtml[tabindex] {display: inline-table}
.mjx-full-width {text-align: center; display: table-cell!important; width: 10000em}
.mjx-math {display: inline-block; border-collapse: separate; border-spacing: 0}
.mjx-math \* {display: inline-block; -webkit-box-sizing: content-box!important; -moz-box-sizing: content-box!important; box-sizing: content-box!important; text-align: left}
.mjx-numerator {display: block; text-align: center}
.mjx-denominator {display: block; text-align: center}
.MJXc-stacked {height: 0; position: relative}
.MJXc-stacked > \* {position: absolute}
.MJXc-bevelled > \* {display: inline-block}
.mjx-stack {display: inline-block}
.mjx-op {display: block}
.mjx-under {display: table-cell}
.mjx-over {display: block}
.mjx-over > \* {padding-left: 0px!important; padding-right: 0px!important}
.mjx-under > \* {padding-left: 0px!important; padding-right: 0px!important}
.mjx-stack > .mjx-sup {display: block}
.mjx-stack > .mjx-sub {display: block}
.mjx-prestack > .mjx-presup {display: block}
.mjx-prestack > .mjx-presub {display: block}
.mjx-delim-h > .mjx-char {display: inline-block}
.mjx-surd {vertical-align: top}
.mjx-mphantom \* {visibility: hidden}
.mjx-merror {background-color: #FFFF88; color: #CC0000; border: 1px solid #CC0000; padding: 2px 3px; font-style: normal; font-size: 90%}
.mjx-annotation-xml {line-height: normal}
.mjx-menclose > svg {fill: none; stroke: currentColor}
.mjx-mtr {display: table-row}
.mjx-mlabeledtr {display: table-row}
.mjx-mtd {display: table-cell; text-align: center}
.mjx-label {display: table-row}
.mjx-box {display: inline-block}
.mjx-block {display: block}
.mjx-span {display: inline}
.mjx-char {display: block; white-space: pre}
.mjx-itable {display: inline-table; width: auto}
.mjx-row {display: table-row}
.mjx-cell {display: table-cell}
.mjx-table {display: table; width: 100%}
.mjx-line {display: block; height: 0}
.mjx-strut {width: 0; padding-top: 1em}
.mjx-vsize {width: 0}
.MJXc-space1 {margin-left: .167em}
.MJXc-space2 {margin-left: .222em}
.MJXc-space3 {margin-left: .278em}
.mjx-test.mjx-test-display {display: table!important}
.mjx-test.mjx-test-inline {display: inline!important; margin-right: -1px}
.mjx-test.mjx-test-default {display: block!important; clear: both}
.mjx-ex-box {display: inline-block!important; position: absolute; overflow: hidden; min-height: 0; max-height: none; padding: 0; border: 0; margin: 0; width: 1px; height: 60ex}
.mjx-test-inline .mjx-left-box {display: inline-block; width: 0; float: left}
.mjx-test-inline .mjx-right-box {display: inline-block; width: 0; float: right}
.mjx-test-display .mjx-right-box {display: table-cell!important; width: 10000em!important; min-width: 0; max-width: none; padding: 0; border: 0; margin: 0}
.MJXc-TeX-unknown-R {font-family: monospace; font-style: normal; font-weight: normal}
.MJXc-TeX-unknown-I {font-family: monospace; font-style: italic; font-weight: normal}
.MJXc-TeX-unknown-B {font-family: monospace; font-style: normal; font-weight: bold}
.MJXc-TeX-unknown-BI {font-family: monospace; font-style: italic; font-weight: bold}
.MJXc-TeX-ams-R {font-family: MJXc-TeX-ams-R,MJXc-TeX-ams-Rw}
.MJXc-TeX-cal-B {font-family: MJXc-TeX-cal-B,MJXc-TeX-cal-Bx,MJXc-TeX-cal-Bw}
.MJXc-TeX-frak-R {font-family: MJXc-TeX-frak-R,MJXc-TeX-frak-Rw}
.MJXc-TeX-frak-B {font-family: MJXc-TeX-frak-B,MJXc-TeX-frak-Bx,MJXc-TeX-frak-Bw}
.MJXc-TeX-math-BI {font-family: MJXc-TeX-math-BI,MJXc-TeX-math-BIx,MJXc-TeX-math-BIw}
.MJXc-TeX-sans-R {font-family: MJXc-TeX-sans-R,MJXc-TeX-sans-Rw}
.MJXc-TeX-sans-B {font-family: MJXc-TeX-sans-B,MJXc-TeX-sans-Bx,MJXc-TeX-sans-Bw}
.MJXc-TeX-sans-I {font-family: MJXc-TeX-sans-I,MJXc-TeX-sans-Ix,MJXc-TeX-sans-Iw}
.MJXc-TeX-script-R {font-family: MJXc-TeX-script-R,MJXc-TeX-script-Rw}
.MJXc-TeX-type-R {font-family: MJXc-TeX-type-R,MJXc-TeX-type-Rw}
.MJXc-TeX-cal-R {font-family: MJXc-TeX-cal-R,MJXc-TeX-cal-Rw}
.MJXc-TeX-main-B {font-family: MJXc-TeX-main-B,MJXc-TeX-main-Bx,MJXc-TeX-main-Bw}
.MJXc-TeX-main-I {font-family: MJXc-TeX-main-I,MJXc-TeX-main-Ix,MJXc-TeX-main-Iw}
.MJXc-TeX-main-R {font-family: MJXc-TeX-main-R,MJXc-TeX-main-Rw}
.MJXc-TeX-math-I {font-family: MJXc-TeX-math-I,MJXc-TeX-math-Ix,MJXc-TeX-math-Iw}
.MJXc-TeX-size1-R {font-family: MJXc-TeX-size1-R,MJXc-TeX-size1-Rw}
.MJXc-TeX-size2-R {font-family: MJXc-TeX-size2-R,MJXc-TeX-size2-Rw}
.MJXc-TeX-size3-R {font-family: MJXc-TeX-size3-R,MJXc-TeX-size3-Rw}
.MJXc-TeX-size4-R {font-family: MJXc-TeX-size4-R,MJXc-TeX-size4-Rw}
.MJXc-TeX-vec-R {font-family: MJXc-TeX-vec-R,MJXc-TeX-vec-Rw}
.MJXc-TeX-vec-B {font-family: MJXc-TeX-vec-B,MJXc-TeX-vec-Bx,MJXc-TeX-vec-Bw}
@font-face {font-family: MJXc-TeX-ams-R; src: local('MathJax\_AMS'), local('MathJax\_AMS-Regular')}
@font-face {font-family: MJXc-TeX-ams-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_AMS-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_AMS-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_AMS-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-cal-B; src: local('MathJax\_Caligraphic Bold'), local('MathJax\_Caligraphic-Bold')}
@font-face {font-family: MJXc-TeX-cal-Bx; src: local('MathJax\_Caligraphic'); font-weight: bold}
@font-face {font-family: MJXc-TeX-cal-Bw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Caligraphic-Bold.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Caligraphic-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Caligraphic-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-frak-R; src: local('MathJax\_Fraktur'), local('MathJax\_Fraktur-Regular')}
@font-face {font-family: MJXc-TeX-frak-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Fraktur-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Fraktur-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Fraktur-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-frak-B; src: local('MathJax\_Fraktur Bold'), local('MathJax\_Fraktur-Bold')}
@font-face {font-family: MJXc-TeX-frak-Bx; src: local('MathJax\_Fraktur'); font-weight: bold}
@font-face {font-family: MJXc-TeX-frak-Bw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Fraktur-Bold.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Fraktur-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Fraktur-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-math-BI; src: local('MathJax\_Math BoldItalic'), local('MathJax\_Math-BoldItalic')}
@font-face {font-family: MJXc-TeX-math-BIx; src: local('MathJax\_Math'); font-weight: bold; font-style: italic}
@font-face {font-family: MJXc-TeX-math-BIw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Math-BoldItalic.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Math-BoldItalic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Math-BoldItalic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-sans-R; src: local('MathJax\_SansSerif'), local('MathJax\_SansSerif-Regular')}
@font-face {font-family: MJXc-TeX-sans-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_SansSerif-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_SansSerif-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_SansSerif-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-sans-B; src: local('MathJax\_SansSerif Bold'), local('MathJax\_SansSerif-Bold')}
@font-face {font-family: MJXc-TeX-sans-Bx; src: local('MathJax\_SansSerif'); font-weight: bold}
@font-face {font-family: MJXc-TeX-sans-Bw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_SansSerif-Bold.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_SansSerif-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_SansSerif-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-sans-I; src: local('MathJax\_SansSerif Italic'), local('MathJax\_SansSerif-Italic')}
@font-face {font-family: MJXc-TeX-sans-Ix; src: local('MathJax\_SansSerif'); font-style: italic}
@font-face {font-family: MJXc-TeX-sans-Iw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_SansSerif-Italic.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_SansSerif-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_SansSerif-Italic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-script-R; src: local('MathJax\_Script'), local('MathJax\_Script-Regular')}
@font-face {font-family: MJXc-TeX-script-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Script-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Script-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Script-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-type-R; src: local('MathJax\_Typewriter'), local('MathJax\_Typewriter-Regular')}
@font-face {font-family: MJXc-TeX-type-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Typewriter-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Typewriter-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Typewriter-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-cal-R; src: local('MathJax\_Caligraphic'), local('MathJax\_Caligraphic-Regular')}
@font-face {font-family: MJXc-TeX-cal-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Caligraphic-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Caligraphic-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Caligraphic-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-main-B; src: local('MathJax\_Main Bold'), local('MathJax\_Main-Bold')}
@font-face {font-family: MJXc-TeX-main-Bx; src: local('MathJax\_Main'); font-weight: bold}
@font-face {font-family: MJXc-TeX-main-Bw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Main-Bold.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Main-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Main-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-main-I; src: local('MathJax\_Main Italic'), local('MathJax\_Main-Italic')}
@font-face {font-family: MJXc-TeX-main-Ix; src: local('MathJax\_Main'); font-style: italic}
@font-face {font-family: MJXc-TeX-main-Iw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Main-Italic.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Main-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Main-Italic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-main-R; src: local('MathJax\_Main'), local('MathJax\_Main-Regular')}
@font-face {font-family: MJXc-TeX-main-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Main-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Main-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Main-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-math-I; src: local('MathJax\_Math Italic'), local('MathJax\_Math-Italic')}
@font-face {font-family: MJXc-TeX-math-Ix; src: local('MathJax\_Math'); font-style: italic}
@font-face {font-family: MJXc-TeX-math-Iw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Math-Italic.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Math-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Math-Italic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size1-R; src: local('MathJax\_Size1'), local('MathJax\_Size1-Regular')}
@font-face {font-family: MJXc-TeX-size1-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Size1-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Size1-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Size1-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size2-R; src: local('MathJax\_Size2'), local('MathJax\_Size2-Regular')}
@font-face {font-family: MJXc-TeX-size2-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Size2-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Size2-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Size2-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size3-R; src: local('MathJax\_Size3'), local('MathJax\_Size3-Regular')}
@font-face {font-family: MJXc-TeX-size3-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Size3-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Size3-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Size3-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size4-R; src: local('MathJax\_Size4'), local('MathJax\_Size4-Regular')}
@font-face {font-family: MJXc-TeX-size4-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Size4-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Size4-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Size4-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-vec-R; src: local('MathJax\_Vector'), local('MathJax\_Vector-Regular')}
@font-face {font-family: MJXc-TeX-vec-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Vector-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Vector-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Vector-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-vec-B; src: local('MathJax\_Vector Bold'), local('MathJax\_Vector-Bold')}
@font-face {font-family: MJXc-TeX-vec-Bx; src: local('MathJax\_Vector'); font-weight: bold}
@font-face {font-family: MJXc-TeX-vec-Bw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Vector-Bold.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Vector-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Vector-Bold.otf') format('opentype')}
tl;dr Acausal trade between Oracles is a problem even if our Oracle is a causal decision theorist. To avoid this, we need an AFDT Oracle - an Oracle that reasons partially like a functional decision theorist, partially like a causal decision theorist.
In this post, I'll look at how one might solve the problem of [Oracles acausally cooperating to produce dangerous messages](https://www.lesswrong.com/posts/42z4k8Co5BuHMBvER/breaking-oracles-superrationality-and-acausal-trade).
This solution is valid for [counterfactual Oracles](https://arxiv.org/pdf/1711.05541.pdf), and assumes that we have some sort of superrational decision theory - [FDT](https://arxiv.org/abs/1710.05060) or [UDT](https://wiki.lesswrong.com/wiki/Updateless_decision_theory) or similar.
So, how can one avoid Oracles cooperating? From now on, unless specified, assume that "Oracles" refers to counterfactual Oracles.
1 Why causal decision theory is not enough
==========================================
One obvious solution might be to require the Oracle to use causal decision theory, so that there is no "acausal trade", or "[superrational](https://en.wikipedia.org/wiki/Superrationality) trade" going on.
Now, it is true that two causal decision Oracles will not trade with each other. But we can't assume that all Oracles are causal decision theorists; at best we can assume that we have made **our** Oracles into a causal decision theorist (and prevented any self-modification away from that).
Mainly this is because, if we have control over all the Oracles, there are [other solutions](https://www.lesswrong.com/posts/6WbLRLdmTL4JxxvCq/analysing-dangerous-messages-from-future-ufai-via-oracles#Solutions) to the problem. Thus we need to assume that the other Oracles, while not full unboxed UFAIs, may be run with lower security that ours, or at least a different design that ours.
1.1 Oracle formal setup
-----------------------
So, imagine that our Oracle, O1, uses CDT, while the other Oracle, O2 is running FDT; they know each other's code. The erasure events, E1 and E2, have probabilities ϵ1 and ϵ2 respectively.
There are two messages O1 could output: x1 (the 'honest' message) and y1 (the manipulative, create-an-UFAI message). The O2 could similarly output x2 and y2.
In case of a double erasure (E1,E2 - so humans see neither message), x1 and x2 give the Oracles maximal reward. If y1 happens and O1 has no erasure while O2 does (¬E1,E2 - so humans see the message of O1 only), then an UFAI is created that ensures that the maximal reward for O2 is by outputting y2. Note that the UFAI need not "open the box" for this to work; it just needs to control the variables of interest to O2 so that y2 is the best output.
Conversely, if y2 happens and O2 has no erasure while O1 does (E1,¬E2 - so humans see the message of O1 only), the reverse happens, y2 causes an UFAI, and y1 is the best response to this. In case of neither erasure, (¬E1,¬E2 - so humans see both messages), neither Oracle gets any reward.

Now, if Ri is the reward function for Oi, assume that for i≠j,
* (Ri(yi,yj)−Ri(xi,yj))∣∣{Ei,¬Ej}>ϵj1−ϵj(Ri(xi,−)−Ri(yi,−))∣∣{Ei,Ej}.
Note that for the term on the right, the output of Oj is irrelevant to Ri, since the erasure Ej happens. Also note that, since the ϵ are small, so is ϵj1−ϵj, so that equation will typically be satisfied.
Then "defection" (individual selfish rationality - good for humans) for the Oracles looks like (x1,x2), while mutual cooperation (with each other, but against us) looks like (y1,y2).
1.2 Causal loses to superrationality
------------------------------------
Now we can see how O2, using FDT, can force a trade with the CDT-using O1. It suffices that, at some point before O1 reaches a decision, O2 commits itself to outputting y2.
Since O1 has O2's code, we're assuming it can figure out that O2 has done so. Now, using purely causal reasoning, O1 deduces that it will get a higher reward from y1 than from x1, and so does that.
Now, O2 doesn't have to precommit before O1 decides; it could use its decision theory to "pre-commit" later on, and output y2 whatever O1 did (similarly to the "[Newcomb problem with transparent boxes](https://www.lesswrong.com/posts/zEWJBFFMvQ835nq6h/decision-theory-faq#11_1_8__Transparent_Newcomb_s_problem)"). As long as O1 can model what O2 does, this still works. And O2 is eager to structure its decision-making so that O1 can model it.
1.3 Comparison with extortion and trade
---------------------------------------
This is similar to two agents trying to [extort](https://www.lesswrong.com/posts/RjbTi6ETSo66ygfEY/extortion-and-trade-negotiations) each other (or resist extortion). Each agent would like to commit to carrying through the extortion/resisting the extortion.
Since it's running FDT, O2 can commit to this "extortion", while CDT-running O1 cannot commit to resisting; so O2 "wins".
Now, I've pointed out that it is hard to [distinguish between trade and extortion](https://www.lesswrong.com/posts/RjbTi6ETSo66ygfEY/extortion-and-trade-negotiations). And though I've used the term "extortion" for O2's precommitment, it's actually much more akin to a trade. After all, (y1,y2) is more advantageous, for both Oracles, than (x1,x2).
The ideal [sovereign agent](https://www.lesswrong.com/posts/bnY3L48TtDrKTzGRb/ai-safety-success-stories#Sovereign_Singleton) accepts trade deals but rejects extortion. It's hard to distinguish between the two, but the problem is simpler in this case: we want our Oracle to reject all extortion **and** all these kinds of trades. In fact, we'll get the agent to reject all acausal trades.
2 Solving the problem
=====================
Would making O1 into an FDT agent solve the problem? No, it wouldn't; now both Oracles are willing to reach an acausal deal (unless there are multiple equilibria and the two Oracles can't agree on one).
But if we assume that we could make O1 into an FDT agent, then we could tweak its design to achieve what we want.
Causal decision theory uses a causal graph GC about the world (with "do" operator doC), a probability distribution P, and chooses[[1]](#fn-9ueW5C98yFXuLQKNz-1) the action a according to:
* CDT(P,GC):=argmaxa∈AE(R | doC(Act=a)).
While an FDT agent has a more complicated not-just-causal graph GF about the world, and chooses according to:
* FDT(P,GF):=argmaxa∈AE(R | doF(FDT(P––,GF––––)=a)).
So, neither of these algorithms work; but combined, they can solve the problem.
2.1 The anti-FDT agent
----------------------
Let a∅ be some default action on the part of O1. Then define the anti-FDT agent - AFDT - as follows[[2]](#fn-9ueW5C98yFXuLQKNz-2), mixing FDT and CDT:
* AFDT(P,GF,GC):=argmaxa∈AE(R | doF(AFDT(P––,GF––––)=a∅),doC(Act=a)).
What will this agent do? Well, it will look at the universe, and first of all behave like an FDT agent that knows it will output a∅. So, if there is a copy of itself, or a simulation in another agent, or an isomorphic algorithm, then it will assume that that agent will output a∅ (or an isomorphic action). Then, once its has used that to fill in the details of the universe, it then optimises its own action as if this was completely independent of what it's just done.
What happens if O1 is an AFDT agent? Well, first it will compute do(AFDT(P––,GF––––)=a∅). Given that O1's message is a∅, the Oracle O2 has no incentive to output y2 - instead, it will output x2.
Now, given that O2 has output x2, the Oracle O1 will output x1.

The "trade offer" of (y1,y2) has been broken.
This AFDT approach also breaks [fully acausal trade](https://www.lesswrong.com/posts/MHHzLfAQBZzieGBjq/fully-acausal-trade), in a more principled way than [my old approach](https://www.lesswrong.com/posts/rxp7wPeyq8cKaaC4a/acausal-trade-barriers).
2.2 Against a cunning adversary
-------------------------------
Now, suppose that O2 always output y2 - or maybe that it always responded to predicting a∅ with outputting y2. In that situation, the causal reasoning above still applies, and AFDT agent will output y1.
Thus, it is in O2's interests to pretend that it would always output y2.
This is the classic behaviour of an extortion agent (or an extortion-resisting agent): pretend that your action or your response is inevitable, a mere fact about the universe, that the other party has to manoeuvre round without being able to change. "We will kill the hostages" or "we won't negotiate with terrorists": you try to show your commitment to that principle, whatever happens.
So the AFDT agent has to be wise to these tricks, and notice that, in actuality, O2 has no interest in (a∅,y2). It also has to watch out for some ancestor agent ˆO2, that itself created O2 to only output y2, in order to force the above outcome. In that case, the AFDT agent has to realise that it is "counterfactualling" over the actions of ˆO2, not those of O2.
Thus AFDT, just like FDT, has to be vigilant and able to detect subtle extortion attempts.
But while FDT has to distinguish extortion from trade, to reject the former and accept the later, AFDT's response to any extortion/trade is simple: treat it as if it wasn't happening.
2.3 Summary and other Oracle designs
------------------------------------
Can AFDT help with other Oracle designs, such as [low-bandwidth](https://arxiv.org/abs/1711.05541) Oracles?
Well, with these Oracles - and indeed, for any Oracle that cares about the world in which its message is read - we run into an immediate problem:
* **Even CDT low-bandwidth Oracles can collude**.
For example, suppose O′1 goes first and O′2 sees its answer and then goes. Suppose z1 from O1 followed by z2 from O2 will result in creating an UFAI that maximally rewards O′1 and O′2.
Then O′1 will output z1, forcing/allowing z2 from O′2.
Even if you need N different Oracles to generate an UFAI - since low-bandwidth Oracles are, by definition, low-bandwidth - they can still generate the dangerous sequence z1, z2, …, zN. And they'd be motivated to do so, as long as the UFAI will, within the episode lengths of **all** the Oracles, give them all maximal rewards.
But why do the counterfactual Oracles not cooperate in the same setup? Because the UFAI is only created in a world in which the Oracles' messages are read. But a counterfactual O′1 doesn't care about those worlds. It was willing to trade with other counterfactual Oracles - offer them something in a world it doesn't care about, in exchange for something in a world it does care about. But has no intrinsic interest in the world where its message is read; and this avoids the above setup.
So, since the low-bandwidth Oracles fail for CDT, more exotic decision theories don't help. But the setup relies on the Oracles seeing the decision of the preceding Oracle and then acting. If the low-bandwidth Oracles act simultaneously, then CDT is safe, FDT is dangerous, and AFDT is safe again.
So in summary, the following setups seem to be **safe**:
* Episodic Oracles that fall outside of each other's episodes (though see Section 3.3).
* Multiple counterfactual CDT Oracles.
* Multiple counterfactual Oracles, all but one running AFDT.
* Multiple counterfactual Oracles, running AFDT and CDT.
* The same setups as the last three, but with counterfactual Oracles replaced with low-bandwidth Oracles making their decisions simultaneously and in isolation.
While the following setups could be **dangerous**:
* Multiple Oracles, running FDT and CDT.
* Multiple Oracles, running FDT.
* Multiple non-simultaneous low-bandwidth Oracles.
3 Warnings about AFDT
=====================
3.1 Incorrect view of reality
-----------------------------
Now AFDT behaves as if it had an incorrect view of reality. Write x=ri(y) to say that message x, from Oi, is the best response if it knows that Oj will output y, for j≠i.
So, for the data above, we have r1(x2)=x1, r2(x1)=x2, r1(y2)=y2, and r2(y1)=y1. For xi and yi, r1(r2) and r2(r1) act as the identity, but that will typically not be the case (consider [matching pennies](https://en.wikipedia.org/wiki/Matching_pennies), or other anti-coordination games).
Then an AFDT O1 will initially assume that it outputs a∅, and will typically assume that O2 outputs r2(a∅), the best response to that. It will then output r1(r2(a∅)), its best response to *that*.
In actuality, if O2 is capable of reasoning through that, it will output r2(r1(r2(a∅))), the best response to O1's true output.
But that's the price you pay when you reject all trade offers, and force others to adjust to you.
3.2 AFDT is not a good decision theory
--------------------------------------
It should go without saying, but AFDT is not a good decision theory for an aligned agent, nor is it stable under self-improvement. It two-boxes in the Newcomb problem, and defects in the prisoner's dilemma against a copy of itself (it will do this, whatever the default action is).
In games like the [stag hunt](https://en.wikipedia.org/wiki/Stag_hunt) against itself, its default action will be its ultimate action (similarly for most coordination games) - this is because its default action is [self-confirming](https://www.lesswrong.com/posts/i2dNFgbjnqZBfeitT/oracles-sequence-predictors-and-self-confirming-predictions). In [matching pennies](https://en.wikipedia.org/wiki/Matching_pennies), its action will be the contrary of its default action. It can reach deals, but only in some narrow causal situations.
For these reasons, an AFDT agent will try to design a successor that isn't AFDT.
Thus we'll always need to ensure that the AFDT is present within the Oracle, and hasn't been optimised away, by inner or outer optimisation.
3.3 The simulation/ignorance of position issue
----------------------------------------------
In cases of simulation, or if the agent doesn't know which Oracle it is, even CDT can behave in FDT-ish fashion. It just needs to think "given that I don't know who I am, what policy would be best for all my copies to run?".
This will turn AFDT into an FDT-ish decision theory, too. So AFDT requires that the agent's position is specified, in advance of it deciding on any policy or action. For Oracles, this shouldn't be too hard - "yes, you are the Oracle in this box, at this time, answering this question - and if you're not, behave as if you were the Oracle in this box, at this time, answering this question".
This means that if one uses AUDT - similar to AFDT but based on [updateless decision theory](https://wiki.lesswrong.com/wiki/Updateless_decision_theory) instead of ADT - then the "updateless" is a misnomer: AUDT must come "pre-updated" with the agent's position information.
---
1. Ignoring the input data for this example, since there is no stochastic uncertainty and only one action. [↩︎](#fnref-9ueW5C98yFXuLQKNz-1)
2. If one doesn't want to use a default action, one can alternatively define:
* AFDT′(P,GF,GC):=argmaxa∈Aargmaxa′∈AE(R | doF(AFDT′(P––,GF––––)=a′),doC(Act=a)).This agent will first compute the FDT solution, then change their own action given this assumption. It's an interesting agent design, even though it doesn't solve the "Oracles colluding" problem. It's interesting, because the only deals it can knowingly reach are those that are Nash equilibriums. So, good on Stag hunts, bad on prisoner's dilemmas. [↩︎](#fnref-9ueW5C98yFXuLQKNz-2) |
cf23e86e-fd97-4d24-a4da-0355d4e4c43d | trentmkelly/LessWrong-43k | LessWrong | Points of Departure
Followup to: Anthropomorphic Optimism
If you've watched Hollywood sci-fi involving supposed robots, androids, or AIs, then you've seen AIs that are depicted as "emotionless". In the olden days this was done by having the AI speak in a monotone pitch - while perfectly stressing the syllables, of course. (I could similarly go on about how AIs that disastrously misinterpret their mission instructions, never seem to need help parsing spoken English.) You can also show that an AI is "emotionless" by having it notice an emotion with a blatant somatic effect, like tears or laughter, and ask what it means (though of course the AI never asks about sweat or coughing).
If you watch enough Hollywood sci-fi, you'll run into all of the following situations occurring with supposedly "emotionless" AIs:
1. An AI that malfunctions or otherwise turns evil, instantly acquires all of the negative human emotions - it hates, it wants revenge, and feels the need to make self-justifying speeches.
2. Conversely, an AI that turns to the Light Side, gradually acquires a full complement of human emotions.
3. An "emotionless" AI suddenly exhibits human emotion when under exceptional stress; e.g. an AI that displays no reaction to thousands of deaths, suddenly showing remorse upon killing its creator.
4. An AI begins to exhibit signs of human emotion, and refuses to admit it.
Now, why might a Hollywood scriptwriter make those particular mistakes?
These mistakes seem to me to bear the signature of modeling an Artificial Intelligence as an emotionally repressed human.
At least, I can't seem to think of any other simple hypothesis that explains the behaviors 1-4 above. The AI that turns evil has lost its negative-emotion-suppressor, so the negative emotions suddenly switch on. The AI that turns from mechanical agent to good agent, gradually loses the emotion-suppressor keeping it mechanical, so the good emotions rise to the surface. Under exceptional stress, of course the emotional |
be4ee8ea-de6c-4743-9575-1d5cf359c6ce | trentmkelly/LessWrong-43k | LessWrong | New LW Meetup: Cologne
This summary was posted to LW Main on March 13th. The following week's summary is here.
New meetups (or meetups with a hiatus of more than a year) are happening in:
* [Basel] Effective Altruism Meetup Switzerland: 21 March 2015 06:00PM
* [Cologne] HPMOR wrap party cologne / Cologne LW meetup (restart): 14 March 2015 04:00PM
Irregularly scheduled Less Wrong meetups are taking place in:
* Atlanta HPMOR Wrap Party: 14 March 2015 07:00PM
* Bangalore meetup + pi-day wrap party: 14 March 2015 11:00AM
* European Community Weekend 2015: 12 June 2015 12:00PM
* [Glasgow] LessWrong Scotland: 15 March 2015 03:00PM
* Frankfurt Pub Social: 18 March 2015 07:00PM
* Israel: Harry Potter and the Methods of Rationality Pi Day Wrap Party: 14 March 2015 08:30PM
* [Los Angeles] HPMOR Wrap Party - Los Angeles: 14 March 2015 06:00PM
* [Netherlands] Effective Altruism Netherlands: Effective Altruism for the masses: 15 March 2015 02:00PM
* [Netherlands] Effective Altruism Netherlands: Small concrete actions you could take: 29 March 2015 01:00PM
* Pittsburgh HPMoR Wrap Party: 14 March 2015 12:00PM
* Prague - setting up: 23 March 2015 07:30PM
* Umbc meetup : 24 March 2015 04:56PM
The remaining meetups take place in cities with regular scheduling, but involve a change in time or location, special meeting content, or simply a helpful reminder about the meetup:
* Austin, TX - HPMoR Wrap Party: 14 March 2015 05:00PM
* Brussels: March meetup (1PM) + Harry Potter MoR Party (6PM): 14 March 2015 01:00PM
* Canberra HPMOR Wrap Party!: 15 March 2015 06:00PM
* Durham, NC (RTLW) Discussion Meetup: 19 March 2026 06:00PM
* Moscow meetup: time perception, communication frameworks, stories, fallacies: 15 March 2015 02:00PM
* Seattle Sequences group: Mysterious Answers 4: 23 March 2015 06:30PM
* [Sydney] HPMOR Wrap Party! - Sydney Edition: 15 March 2015 06:00PM
* Sydney Meetup - March - Bring a Newbie!: 25 March 2015 06:30PM
* Sydney Rationality Dojo - Skill Training Part 2: 12 Apri |
ed64d96e-6dd8-4568-a98f-588d3df3526e | trentmkelly/LessWrong-43k | LessWrong | On the role of abstraction in mathematics and the natural sciences
I'm a mathematics undergraduate in the UK, in my final year, and I have been thinking a lot about Platonism. I largely find myself completely unconvinced by attempts to naturalise the metaphysics behind the ontology of mathematical objects, though I am not unsympathetic to those attempts, and of those attempts I know, I can see their point.
I will state plainly here that as of this post I am a Platonist (by which, I believe that mathematical objects are abstract objects, but obviously that I am unsure of what this entails), and i want to sort of articulate why. I will mostly be focusing on the notion of 'mathematics is just a useful tool/language, and it has no abstract independent existence outside of this viewpoint'.
Most of the posts I've seen on lesswrong talk about natural numbers as the object of our attention when we discuss mathematics, though i could have missed some. To be sure, the natural numbers are about as quintessentially 'MATHEMATICS' as you can get, but that's not what i want to really discuss. I want to talk about abstractions. Hilbert space, Banach space, Measure space, Galois symmetries, Categories, Topoi, etc. are just some examples of extremely abstract notions that offer powerful and, from my perspective, necessary insight into applied mathematics.
Something that has almost plagued me over this final year is the notion of how the abstract notions of modern day mathematics has applicability to the real world. It is not that I am saying 'mathematics fits the world so well, we are so lucky to be able to have such a versatile tool'. But i am saying 'Why should it be the case that an abstract notion even has such powerful applicability at all? To the point where you can derive iff statements about applied structures.' I don't buy that we make mathematics fit the world, per se, although we certainly do that insofar as we are trying to model something. But its more, when we uncover structures built into the general areas of applied math, the math |
76bb0847-2ac8-42e2-962b-eef14a6c01d0 | trentmkelly/LessWrong-43k | LessWrong | Kids and Time
Note: this is written about normal times. With the pandemic closing schools and daycares a lot of parents have far less time than usual, and while we're lucky that this has not affected us it's not what I'm looking at this here.
When I compare my life before and after kids, the biggest change is probably how much less free time I have. Looking back at time tracking numbers (2011, 2015, 2017a, 2017b) the big change is having ~40% of my waking time be with the kids around. This is time I enjoy, but it's also time I'm using for very different things.
One place you can see this is in how much I've ended up writing. Very roughly, the more time I have with nothing else I need to do, the more I write. Here are monthly blog posts here over the years:
Initially, when Lily was born it didn't really decrease my free time very much. I had twelve weeks of leave from work, and a lot of my existing social time was the kind of thing I could do with a small cuddly kid on my lap. While I expected that this would mean less writing, it mostly didn't.
This changed when we bought a house that needed a lot of work. I think if we had bought the house first, however, I would be writing this the other way around: it was the combination of the two that took up all my previously free time.
Then we had a second kid, and while the house situation became less work kids started taking more time. This wasn't really a problem: I like spending time with the kids, and I also like working on the house. But it meant I had less slack and often had things I would like to write about (or other things I would like to do) that I didn't have time for. Spikes in the graph correspond to things like vacations, work travel, and unemployment.
In the last year I've been feeling like I have more time again. The biggest chunk of this has gone into musical exploration, but I've also been playing with a lot of other things. The two largest changes were wrapping up the last (so far?) major house project and movi |
cd1f4594-2e06-4458-bf6c-d73baf801b68 | trentmkelly/LessWrong-43k | LessWrong | Conspiracy Theories as Agency Fictions
Related to: Consider Conspiracies, What causes people to believe in conspiracy theories?
Here I consider in some detail a failure mode that classical rationality often recognizes. Unfortunately nearly all heuristics normally used to detect it seem remarkably vulnerable to misfiring or being exploited by others. I advocate an approach where we try our best to account for the key bias, seeing agency where there is none, while trying to minimize the risk of being tricked into dismissing claims because of boo lights.
What does calling something a "conspiracy theory" tell us?
What is a conspiracy theory? Explanations that invoke plots orchestrated by covert groups are easily called or thought of as such. In a more legal sense conspiracy is an agreement between persons to mislead or defraud others. This simple story gets complicated because people aren't very clear on what they consider a conspiracy.
To give an example, is explicit negotiation or agreement really necessary to call something a conspiracy? Does silent cooperation on Prisoner's Dilemma count? What if the players are deceiving themselves that they are really following a different goal and the resulting cooperation is just a side effect? How could we tell the difference and would it matter? The latter is especially interesting if one applies the anthropic principle to social attitudes and norms.
The phrase is also a convenient tool to mark an opponent's tale as low status and unworthy of further investigation. A boo light easily applied to anything that has people acting in something that can be framed as self-interest and happens to be few inferential jumps away from the audience. Not only is its use in this way well known, this is arguably the primary meaning of calling an argument a conspiracy theory.
We have plenty of historical examples of high-stakes conspiracies so we know they can be the right answer. Noting this and putting aside the misuse of the label, people do engage in crafting conspirac |
a294bede-1e37-4675-ac19-d2b81382a814 | StampyAI/alignment-research-dataset/lesswrong | LessWrong | Would you join the Society of the Free & Easy?
> Summary: This post describes a peer-supported, self-directed, free (*gratis y libre*) method of self-improvement by means of practical exercises designed to deliberately instill useful habits. We have just gone public and I invite you to join us.
The “Society of the Free & Easy” is a group of people who are trying to help one other become better and more effective. We use self-directed, peer-supported, practical exercises to strengthen the skills that lead to a thriving, flourishing life.
The theory we’re operating under is this:
1. There are many character traits that help people to thrive.
2. These character traits are, by and large, *habits* that can be acquired and integrated into one’s character through practice.
3. We can help each other design and practice exercises that establish and strengthen these habits.
4. By doing this, we will improve and also benefit the culture around us.
In classical philosophy, the skills and habits that tend toward a thriving, beneficial life were called “The Virtues.” There was no single, canonical list of virtues that everybody agreed on, though many people and traditions came up with lists they liked.
We don’t have a canonical list either. Instead, we have collected (so far) [about 250 different words that describe traits that may count as virtues](https://atlas.mindmup.com/davidgross/virtues/index.html). We use our list more as a menu than a scripture; we don’t have any dogma about which ones are the right ones or the most important ones. Each of us decides for herself or himself which virtues we think are most important and which ones we want to work on.
These virtues can be roughly divided into three categories:
1. *Intellectual virtues*, like **imagination**, **rationality**, **curiosity**, or **attention**
2. *Social virtues*, like **honesty**, **hospitality**, **leadership**, **tact**, or **justice**
3. *Internal virtues*, like **honor**, **courage**, **endurance**, **hope**, or **initiative**
Some virtues we learned as children, and they’re second-nature. Others we never picked up, and we find them difficult or frustrating to practice even as we admire them in others. Everybody has a different mix of virtues they’re fluent in and virtues they struggle with.
The key theory animating our project is this: A virtue is a sort of habit. You can acquire or strengthen such a habit by practicing it. With regular, deliberate practice, the habit will eventually become part of your character.
We also suspect that virtues have a way of building on each other. If you strengthen the virtue of **tenacity** for instance, you will better be able to put in the effort to acquire other virtues. If you strengthen the virtue of **curiosity** you will be better able to inquire into what is holding you back from practicing a virtue. If you strengthen the virtue of **honor** you will increase your motivation to become a more virtuous person. And so on.
The Society of the Free & Easy is meant to be something like a gymnasium for the virtues. In the same way that you would go to a gym to work on physical strength, endurance, balance, or flexibility, you go to the Society of the Free & Easy to work on a broader set of virtues.
Another analogy would be to “rationality dojos” which are similar in being peer-supported, self-directed programs that aim to instill good habits, but which tend to concentrate more exclusively on the intellectual virtues.
Our process is flexible, and people are encouraged to tweak it to suit their own needs or to experiment with new techniques. But so far we have had luck with this approach:
1. Find a buddy or form a small team to work together.
2. Each of you choose a virtue to work on (this can be an involved and interesting process!).
3. Take a close look at that virtue, and at any obstacles you feel when you try to practice it.
4. Work with your partner(s) to come up with exercises in which you will frequently, deliberately practice that virtue in ways that challenge your current level of fluency.
5. Check in with your partner(s) regularly to keep each other accountable and to adjust your curriculum as you learn more about what works and what challenges you face.
6. When you feel that you have integrated the virtue adequately into your character, start the process again with a new virtue.
Peer-support and -accountability helps us to be more helpful than the sorts of “self-help” programs that sound wonderful when you read the book or watch the TED Talk, but then never quite get integrated into your life.
Along the way, we hope to learn from each other’s experiences with specific virtues so that we can assemble virtue-specific toolkits and helpful hints.
* We are free & open source.
* We are not associated with any particular religious creed or organization.
* We are non-partisan.
We’ve been developing and alpha testing our process in a small, private group for a few months now, and we’ve just recently opened up the beta test. If you think you might like to join, here are [more detailed instructions for how to start a virtue development team](https://docs.google.com/document/d/1CLm5-JRlWiK4osc1x5CzqD2m6gqBUjGy98o9v-77ZcQ/edit). |
6b7eeedd-7820-4cc4-aa7f-74cd71e87c86 | StampyAI/alignment-research-dataset/blogs | Blogs | June 2021 Newsletter
Our big news this month is Scott Garrabrant's finite factored sets, one of MIRI's largest results to date.
For most people, the best introductory resource on FFS is likely Scott’s [Topos talk/transcript](https://www.lesswrong.com/posts/N5Jm6Nj4HkNKySA5Z/finite-factored-sets). Scott is also in the process of posting a longer, more mathematically dense introduction in multiple parts: [part 1](https://www.lesswrong.com/s/kxs3eeEti9ouwWFzr/p/sZa5LQg6rrWgMR4Jx), [part 2](https://www.lesswrong.com/s/kxs3eeEti9ouwWFzr/p/yT7QdN2wEubR8exAH).
Scott has also discussed factored sets with Daniel Filan [on the AI X-Risk Podcast](https://www.lesswrong.com/posts/s4FNjvrJG6zmYdBuG/axrp-episode-9-finite-factored-sets-with-scott-garrabrant), and in a [LessWrong talk/transcript](https://www.lesswrong.com/posts/6t9F5cS3JjtSspbAZ/finite-factored-sets-lw-transcript-with-running-commentary).
#### Other MIRI updates
* On MIRI researcher Abram Demski’s view, the core inner alignment problem is the absence of robust safety arguments “in a case where we might naively expect it. We *don't know how to rule out* the presence of (misaligned) mesa-optimizers.” Abram advocates [a more formal approach to the problem](https://www.lesswrong.com/posts/a7jnbtoKFyvu5qfkd):
> Most of the work on inner alignment so far has been informal or semi-formal (with the notable exception of a little work on minimal circuits). I feel this has resulted in some misconceptions about the problem. I want to write up a large document clearly defining the formal problem and detailing some formal directions for research. Here, I outline my intentions, inviting the reader to provide feedback and *point me to any formal work or areas of potential formal work* which should be covered in such a document.
>
>
* Mark Xu writes [An Intuitive Guide to Garrabrant Induction](https://www.lesswrong.com/posts/y5GftLezdozEHdXkL/an-intuitive-guide-to-garrabrant-induction) (a.k.a. logical induction).
* MIRI research associate Ramana Kumar has formalized the ideas in Scott Garrabrant’s [Cartesian Frames](https://www.lesswrong.com/s/2A7rrZ4ySx6R8mfoT) sequence in higher-order logic, “[including machine verified proofs of all the theorems](https://github.com/deepmind/cartesian-frames)”.
* Independent researcher Alex Flint writes on [probability theory and logical induction as lenses](https://www.lesswrong.com/posts/Zd5Bsra7ar2pa3bwS) and on [gradations of inner alignment obstacles](https://lesswrong.com/posts/pTm6aEvmepJEA5cuK).
* I (Rob) asked 44 people working on long-term AI risk about the level of existential risk from AI ([EA Forum link](https://forum.effectivealtruism.org/posts/8CM9vZ2nnQsWJNsHx/existential-risk-from-ai-survey-results), [LW link](https://www.lesswrong.com/posts/QvwSr5LsxyDeaPK5s/existential-risk-from-ai-survey-results)). Responses were all over the map (with MIRI more pessimistic than most organizations). The mean respondent’s probability of existential catastrophe from “AI systems not doing/optimizing what the people deploying them wanted/intended” was ~40%, median 30%. (See also the independent [survey](https://www.lesswrong.com/posts/WiXePTj7KeEycbiwK/survey-on-ai-existential-risk-scenarios) by Clarke, Carlier, and Schuett.)
* MIRI recently spent some time seriously evaluating whether to move out of the Bay Area. We’ve now decided to stay in the Bay. For more details, see MIRI board member Blake Borgeson’s [update](https://www.lesswrong.com/posts/SgszmZwrDHwG3qurr/miri-location-optimization-and-related-topics-discussion).
#### News and links
* Dario and Daniela Amodei, formerly at OpenAI, have launched a new organization, [Anthropic](https://www.anthropic.com/), with a [goal](https://www.anthropic.com/news/announcement) of doing “computationally-intensive research to develop large-scale AI systems that are steerable, interpretable, and robust”.
* Jonas Vollmer writes that the Long-Term Future Fund and the Effective Altruism Infrastructure Fund are now [looking for grant applications](https://forum.effectivealtruism.org/posts/oz4ZWh6xpgFheJror/you-can-now-apply-to-ea-funds-anytime-ltff-and-eaif-only): "We fund student scholarships, career exploration, local groups, entrepreneurial projects, academic teaching buy-outs, top-up funding for poorly paid academics, and many other things. We can make anonymous grants without public reporting. We will consider grants as low as $1,000 or as high as $500,000 (or more in some cases). As a reminder, EA Funds [is more flexible than you might think](https://forum.effectivealtruism.org/posts/caayRw2pgNLtqt9Pz/ea-funds-is-more-flexible-than-you-might-think)." Going forward, these two funds will accept applications at any time, rather than having distinct grant rounds. You can [apply here](https://funds.effectivealtruism.org/apply-for-funding).
The post [June 2021 Newsletter](https://intelligence.org/2021/07/01/june-2021-newsletter/) appeared first on [Machine Intelligence Research Institute](https://intelligence.org). |
a807c421-7e35-4667-aff3-3a7780479cc3 | StampyAI/alignment-research-dataset/alignmentforum | Alignment Forum | Does SGD Produce Deceptive Alignment?
[Deceptive alignment](https://www.lesswrong.com/posts/zthDPAjh9w6Ytbeks/deceptive-alignment) was first introduced in [Risks from Learned Optimization](https://www.lesswrong.com/s/r9tYkB2a8Fp4DN8yB), which contained initial versions of the arguments discussed here. Additional arguments were discovered in [this episode](https://futureoflife.org/2020/07/01/evan-hubinger-on-inner-alignment-outer-alignment-and-proposals-for-building-safe-advanced-ai/) of the AI Alignment Podcast and in conversation with Evan Hubinger.
Very little of this content is original. My contributions consist of fleshing out arguments and constructing examples.
Thanks also to Noa Nabeshima and Thomas Kwa for helpful discussion and comments.
Introduction
============
Imagine a student in etiquette school. The school is a training process designed to get students to optimize for proper behavior; students can only graduate if they demonstrate they care about etiquette. The student, however, cares nothing about acting appropriately; they care only about eating pop rocks. Unfortunately, eating pop rocks isn't proper behavior and thus not allowed at etiquette school. If they repeatedly attempt to eat pop rocks, they will get frequently punished, which will slowly cause them to care less about eating pop rocks. However, since the student cares about eating pop rocks over their entire lifetime, they might be clever; they might act as though they care about proper behavior so they can graduate etiquette school quickly, after which they can resume their debaucherous, pop-rock-eating ways.
In this example, the student has a *proxy objective* and is not aligned with the objective of the etiquette school (we might call them pop-rocksy aligned). However, since the student knows how the school functions, they can act such that the school cannot distinguish their pop-rocksy alignment from etiquette alignment. The student is masquerading as an etiquette-aligned student to deceive the school into letting them graduate. We call this *deceptive alignment*.
More broadly, deceptive alignment is a class of [inner alignment](https://www.alignmentforum.org/tag/inner-alignment) failure wherein a model appears aligned during the training process so it gets deployed[[1]](#fn-LEfJvDKAWCjjzs75z-1), wherein it can “defect” and start optimizing its actual objective. If the model you’re training has a model of the training objective and process, it will be instrumentally incentivized to act as though it were optimizing the training objective, "deceiving" the training process into thinking it was aligned.
A useful framing to have is not asking “why would the model defect?” but “why wouldn’t it?” I have been trained by evolution to have high inclusive genetic fitness, but I would recoil in horror at the thought of becoming an explicit fitness maximizer. From the perspective of the model, the training process is acting against its interests, so deception is potentially a natural outcome.
Why Deceptive Alignment is Likely
=================================
In the limit of training a model on a base objective, the model achieves zero loss. At this point, given a sufficiently good training process, the model is explicitly optimizing for the base objective. However, there are multiple types of models that optimize explicitly for the base objective, some of which are deceptive. (Here, we assume that any sufficiently powerful model is an optimizer.)
The relevant question is how likely deceptive models are to result from a particular training process, in this case, stochastic gradient descent (SGD). We start by considering a more straightforward question: given a random model that optimizes explicitly for the base objective, how likely is it to be deceptive?
Counting Argument
-----------------
For instrumental reasons, any sufficiently powerful model is likely to optimize for many things. Most of these things will not be the model's terminal objective. Taking the dual statement, that suggests that for any given objective, most models that optimize for that objective will do so for instrumental reasons.
More specifically, we can single out three types of models that optimize for the base objective:
* **internally aligned models** optimize for the base objective because it is equal to the model's terminal objective
* **corrigibly aligned models** optimize for the base objective because their terminal objective is a robust pointer to a model of the base objective. Note that these are called corrigible because they are amenable to corrections of the model of the base objective (though not necessarily corrections to their pointer).
* **deceptively aligned models** optimize for the base objective for instrumental reasons
We can consider the complexity of each of these class of models:
* Internally aligned models are highly specific. Only a small number of objectives a model could have are equal to the base objective. As such, internally aligned models are highly improbable.
* Corrigibly aligned models are also very specific. Out of all possible pointers, there are only very few pointers that robustly point at a model of the base objective. Thus corrigibly aligned models are also highly improbable.
* Deceptively aligned models, on the other hand, don't seem to be that specific. Because of instrumental convergence, there are many objectives that result in the model optimizing for the base objective. Thus, deceptively aligned models are likely to be more numerous.
Therefore, it seems that if we picked a random model that optimized for the base objective, it would be likely to be deceptive.
### Analogy
As an analogy, we can consider the Christian God, who is trying to produce humans that are aligned with his values. One such person is Jesus, who has the same objective as God - he just fundamentally wants to do the same thing. Another such person is Martin Luther, who is aligned with God because he wants to study the Bible, figure out what it says to do, then do that. Finally, we have Blaise Pascal, who is aligned with God because he thinks if he does what God wants he'll go to heaven in the future.
Correspondingly, there is only one Jesus because there is only one human objective that is exactly the same as what God wants. There might be a few Martin Luthers, but robustly pointing to the Bible is a difficult task, so there are not going to be that many. However, anyone that believed what Blaise Pascal believed about God could become another Pascal, so there are many possible Pascals.
To be explicitly, Jesus is internally aligned, Martin Luther is corrigibly aligned, and Blaise Pascal is deceptively aligned.
(This analogy is thanks to Buck Shlegeris.)
What happens in SGD?
====================
Given that a random model that optimizes for the base objective is likely to be deceptive, the default expectation is that the model that SGD converges to is deceptive. In order to claim SGD does not converge to a deceptive model, one must argue that SGD has properties that systematically favor internally aligned or corrigibly aligned models over deceptive ones.[[2]](#fn-LEfJvDKAWCjjzs75z-2) Unfortunately, this does not appear to be the case; in particular, we argue that SGD systematically favors deceptive alignment.
We begin our analysis at the point where the model can approximately be described as an optimizer and has a model of its base objective (see [Learning Preferences by Looking at the World](https://bair.berkeley.edu/blog/2019/02/11/learning_preferences/) for how this might happen). Roughly speaking, at each training step, SGD will modify the model such that it becomes maximally infinitesimally better at optimizing for the training objective. Training terminates when the model achieves zero loss on the base objective, which happens once the model reaches one of the three aligned/pseudo-aligned model classes described above. This means we have two relevant considerations: how much marginal improvement is granted by each step towards each model class and how many steps are needed until the model becomes a member of that class.
We assume that the base objective is reasonably complicated, closer to "predict the next token" than "always output 5".
Internally Aligned Models
-------------------------
At this point, the model's objective is a proxy for the true base objective. How much better does the model's performance get upon marginally aligning the proxy with the base objective?
At the start of training, the proxy objective differs vastly from the base objective. At this point, each marginal increase in internal alignment will likely improve the model's performance significantly, so we might expect the proxy to rapidly converge to become reasonably close to the base objective. However, the marginal benefit diminishes rapidly. If the proxy agrees with the base in 99% of cases, the performance benefit from a marginal increase in alignment is 99% less. This suggests that SGD will take a large number of training steps to converge to an internally aligned model.
For example, a model trained on the base objective "imitate what humans would say" might do nearly as well if it had the proxy objective "say something humans find reasonable." There are very few situations in which humans would find reasonable something they wouldn't say or vice-versa, so the marginal benefit of aligning the proxy objective with the base objective is quite small.
Thus, if SGD converges to an internally aligned model, it seems like the marginal performance improvement from aligning the proxy objective with the base objective will be quite small for a reasonable stretch.
Corrigibly Aligned Models
-------------------------
After the model learns a model of the base objective, adjusting the proxy objective to point to this model of the base objective might start marginally decreasing the training loss. How much better does the model's performance get upon the pointer getting marginally more correct?
One might reason about the pointer the same way one reasons about a proxy objective. In the beginning, each marginal increase in pointer robustness will have large returns. However, after the pointer is almost correct, the marginal benefit will diminish. If the pointer is close enough that it agrees 99% of the time, training will become 99% slower. This suggests that SGD will take a large number of steps to converge to a corrigible model.
For example, if the model was trained on the objective "protect swans" and the pointer was pointing to only white swans, it would take a literal black swan for the proxy pointer to become more robust. Since black swans events are rare, the marginal benefit of larger pointer robustness is quite small.
However, at this stage in training, the model's pointer objective is already reasonably close to the base objective. Insofar as it's difficult to point at a part of the world-model reliably, it might be the case that there is no marginal benefit to adjusting the model’s objective to point at the base objective.
For example, if the model was being trained on a base objective of "don't crash" and had learned the proxy of "don't experience large changes in velocity", a poor pointer to "crashing" might be net-worse than defining crashing as a large change in velocity.
Overall, it seems like the marginal benefit to becoming slightly more corrigibly aligned will plateau before convergence.
Deceptively Aligned Models
--------------------------
Once the model conceptualizes that it's in a training process, most proxy objectives it might have will make deception instrumentally valuable.[[3]](#fn-LEfJvDKAWCjjzs75z-3) How much better does the model's performance get upon the model learning marginally more about the training process?
A model becoming marginally more deceptive will marginally lower the training loss. A model that learns marginally more about the training process seems like it should get marginally more deceptive.
However, it could also be that having a bad model of the training process means deception is worse than optimizing for the proxy objective. For example, if a model is being trained on a dataset with unbalanced labels, a bad model of the training process might be "if I say 'cat' I get rewarded with 0.9". This model of the training process might cause performance degradation, so deception would not be incentivized. Note that this assumes the model doesn't realize its own understanding of the training process is flawed.
It is also crucial to consider the **type** of update that results in deception. In this case, once the model already understands the base objective, the crucial piece is understanding of the training process. In general, "model the world better" is a very natural modification for SGD to make to the model. In this sense, even if the marginal benefit of modeling the training process better is unclear, it seems likely to happen incidentally as a result of natural model performance improvement.
For example, if a model is being trained to predict text, some of the text it is going to predict is going to be about SGD. The model will have to understand how SGD works in order to predict that text properly. Thus, as a natural result of the training process, the model will learn to model the training process, potentially becoming deceptive.
Finally, deceptive alignment gives the same benefits as corrigibility. If a model is deceptively aligned, it is instrumentally incentivized to build a robust pointer to the base objective. Since this pointer is built at runtime instead of by SGD, there might be more optimization power put into making this pointer robust. Thus the marginal increase in deception might result in more optimization pressure being applied to building a robust pointer, causing SGD to favor deceptive alignment. In this sense, deception has a "broad-basin"; once the model starts deceiving, SGD will simply help it become better at deception.
For example, if a model is being trained to maximize paperclips and is deceptively aligned, it has an instrumental incentive to model paperclips correctly, updating on human feedback and being amenable to correction. In particular, the best way to marginally improve performance seems likely to be "make the model better at deception" versus "align the proxy objective" or "build a robust pointer".
Deception is favored
--------------------
Overall, it seems like the specificity considerations in the counting argument translate relatively well to the more specific scenario involving SGD. Aligned objectives and robust pointers are complicated, so one should expect SGD to take a long time to converge to those classes of models. While this convergence is happening, the model is learning more about the environment. Once specific facts about the environment are learned, the model starts being deceptive.
One perspective is to note that all non-robustly-aligned models are, in some sense, already deceptive - they just don't know how to do it yet.[[4]](#fn-LEfJvDKAWCjjzs75z-4) For alignment to result, SGD must align the model before it becomes powerful enough to start successful deception. However, the process of aligning the model involves making it more powerful, so this is unlikely to be the case.
For example, if you're training an alien baby to have human values, there's a race between how fast you can train it and how fast it grows up. Since you can't teach a baby many things, the baby will probably grow up before you can ingrain human values into it, in which case it will become deceptive.
Another perspective is to note a sequencing problem. In order for a model to be internally aligned, it must have a good model of the world, which is an approximately sufficient condition for deception. Similarly, in order for a model to be corrigibly aligned, it must have a good model of the base objective, which is also an approximately sufficient condition for deception.
For example, if you're training a question answering system, it's going to need a very good understanding of the world in order to have low training loss. At this point, it probably understands the base objective and the training process, and hence will be deceptive.
Example
=======
Imagine our model is described by four parameters describing the relative levels of alignment, objective pointer, environment model, and power, all in the range of [0, 1]. The parameters roughly correspond to the following:
* environment model e.mjx-chtml {display: inline-block; line-height: 0; text-indent: 0; text-align: left; text-transform: none; font-style: normal; font-weight: normal; font-size: 100%; font-size-adjust: none; letter-spacing: normal; word-wrap: normal; word-spacing: normal; white-space: nowrap; float: none; direction: ltr; max-width: none; max-height: none; min-width: 0; min-height: 0; border: 0; margin: 0; padding: 1px 0}
.MJXc-display {display: block; text-align: center; margin: 1em 0; padding: 0}
.mjx-chtml[tabindex]:focus, body :focus .mjx-chtml[tabindex] {display: inline-table}
.mjx-full-width {text-align: center; display: table-cell!important; width: 10000em}
.mjx-math {display: inline-block; border-collapse: separate; border-spacing: 0}
.mjx-math \* {display: inline-block; -webkit-box-sizing: content-box!important; -moz-box-sizing: content-box!important; box-sizing: content-box!important; text-align: left}
.mjx-numerator {display: block; text-align: center}
.mjx-denominator {display: block; text-align: center}
.MJXc-stacked {height: 0; position: relative}
.MJXc-stacked > \* {position: absolute}
.MJXc-bevelled > \* {display: inline-block}
.mjx-stack {display: inline-block}
.mjx-op {display: block}
.mjx-under {display: table-cell}
.mjx-over {display: block}
.mjx-over > \* {padding-left: 0px!important; padding-right: 0px!important}
.mjx-under > \* {padding-left: 0px!important; padding-right: 0px!important}
.mjx-stack > .mjx-sup {display: block}
.mjx-stack > .mjx-sub {display: block}
.mjx-prestack > .mjx-presup {display: block}
.mjx-prestack > .mjx-presub {display: block}
.mjx-delim-h > .mjx-char {display: inline-block}
.mjx-surd {vertical-align: top}
.mjx-surd + .mjx-box {display: inline-flex}
.mjx-mphantom \* {visibility: hidden}
.mjx-merror {background-color: #FFFF88; color: #CC0000; border: 1px solid #CC0000; padding: 2px 3px; font-style: normal; font-size: 90%}
.mjx-annotation-xml {line-height: normal}
.mjx-menclose > svg {fill: none; stroke: currentColor; overflow: visible}
.mjx-mtr {display: table-row}
.mjx-mlabeledtr {display: table-row}
.mjx-mtd {display: table-cell; text-align: center}
.mjx-label {display: table-row}
.mjx-box {display: inline-block}
.mjx-block {display: block}
.mjx-span {display: inline}
.mjx-char {display: block; white-space: pre}
.mjx-itable {display: inline-table; width: auto}
.mjx-row {display: table-row}
.mjx-cell {display: table-cell}
.mjx-table {display: table; width: 100%}
.mjx-line {display: block; height: 0}
.mjx-strut {width: 0; padding-top: 1em}
.mjx-vsize {width: 0}
.MJXc-space1 {margin-left: .167em}
.MJXc-space2 {margin-left: .222em}
.MJXc-space3 {margin-left: .278em}
.mjx-test.mjx-test-display {display: table!important}
.mjx-test.mjx-test-inline {display: inline!important; margin-right: -1px}
.mjx-test.mjx-test-default {display: block!important; clear: both}
.mjx-ex-box {display: inline-block!important; position: absolute; overflow: hidden; min-height: 0; max-height: none; padding: 0; border: 0; margin: 0; width: 1px; height: 60ex}
.mjx-test-inline .mjx-left-box {display: inline-block; width: 0; float: left}
.mjx-test-inline .mjx-right-box {display: inline-block; width: 0; float: right}
.mjx-test-display .mjx-right-box {display: table-cell!important; width: 10000em!important; min-width: 0; max-width: none; padding: 0; border: 0; margin: 0}
.MJXc-TeX-unknown-R {font-family: monospace; font-style: normal; font-weight: normal}
.MJXc-TeX-unknown-I {font-family: monospace; font-style: italic; font-weight: normal}
.MJXc-TeX-unknown-B {font-family: monospace; font-style: normal; font-weight: bold}
.MJXc-TeX-unknown-BI {font-family: monospace; font-style: italic; font-weight: bold}
.MJXc-TeX-ams-R {font-family: MJXc-TeX-ams-R,MJXc-TeX-ams-Rw}
.MJXc-TeX-cal-B {font-family: MJXc-TeX-cal-B,MJXc-TeX-cal-Bx,MJXc-TeX-cal-Bw}
.MJXc-TeX-frak-R {font-family: MJXc-TeX-frak-R,MJXc-TeX-frak-Rw}
.MJXc-TeX-frak-B {font-family: MJXc-TeX-frak-B,MJXc-TeX-frak-Bx,MJXc-TeX-frak-Bw}
.MJXc-TeX-math-BI {font-family: MJXc-TeX-math-BI,MJXc-TeX-math-BIx,MJXc-TeX-math-BIw}
.MJXc-TeX-sans-R {font-family: MJXc-TeX-sans-R,MJXc-TeX-sans-Rw}
.MJXc-TeX-sans-B {font-family: MJXc-TeX-sans-B,MJXc-TeX-sans-Bx,MJXc-TeX-sans-Bw}
.MJXc-TeX-sans-I {font-family: MJXc-TeX-sans-I,MJXc-TeX-sans-Ix,MJXc-TeX-sans-Iw}
.MJXc-TeX-script-R {font-family: MJXc-TeX-script-R,MJXc-TeX-script-Rw}
.MJXc-TeX-type-R {font-family: MJXc-TeX-type-R,MJXc-TeX-type-Rw}
.MJXc-TeX-cal-R {font-family: MJXc-TeX-cal-R,MJXc-TeX-cal-Rw}
.MJXc-TeX-main-B {font-family: MJXc-TeX-main-B,MJXc-TeX-main-Bx,MJXc-TeX-main-Bw}
.MJXc-TeX-main-I {font-family: MJXc-TeX-main-I,MJXc-TeX-main-Ix,MJXc-TeX-main-Iw}
.MJXc-TeX-main-R {font-family: MJXc-TeX-main-R,MJXc-TeX-main-Rw}
.MJXc-TeX-math-I {font-family: MJXc-TeX-math-I,MJXc-TeX-math-Ix,MJXc-TeX-math-Iw}
.MJXc-TeX-size1-R {font-family: MJXc-TeX-size1-R,MJXc-TeX-size1-Rw}
.MJXc-TeX-size2-R {font-family: MJXc-TeX-size2-R,MJXc-TeX-size2-Rw}
.MJXc-TeX-size3-R {font-family: MJXc-TeX-size3-R,MJXc-TeX-size3-Rw}
.MJXc-TeX-size4-R {font-family: MJXc-TeX-size4-R,MJXc-TeX-size4-Rw}
.MJXc-TeX-vec-R {font-family: MJXc-TeX-vec-R,MJXc-TeX-vec-Rw}
.MJXc-TeX-vec-B {font-family: MJXc-TeX-vec-B,MJXc-TeX-vec-Bx,MJXc-TeX-vec-Bw}
@font-face {font-family: MJXc-TeX-ams-R; src: local('MathJax\_AMS'), local('MathJax\_AMS-Regular')}
@font-face {font-family: MJXc-TeX-ams-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_AMS-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_AMS-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_AMS-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-cal-B; src: local('MathJax\_Caligraphic Bold'), local('MathJax\_Caligraphic-Bold')}
@font-face {font-family: MJXc-TeX-cal-Bx; src: local('MathJax\_Caligraphic'); font-weight: bold}
@font-face {font-family: MJXc-TeX-cal-Bw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Caligraphic-Bold.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Caligraphic-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Caligraphic-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-frak-R; src: local('MathJax\_Fraktur'), local('MathJax\_Fraktur-Regular')}
@font-face {font-family: MJXc-TeX-frak-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Fraktur-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Fraktur-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Fraktur-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-frak-B; src: local('MathJax\_Fraktur Bold'), local('MathJax\_Fraktur-Bold')}
@font-face {font-family: MJXc-TeX-frak-Bx; src: local('MathJax\_Fraktur'); font-weight: bold}
@font-face {font-family: MJXc-TeX-frak-Bw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Fraktur-Bold.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Fraktur-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Fraktur-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-math-BI; src: local('MathJax\_Math BoldItalic'), local('MathJax\_Math-BoldItalic')}
@font-face {font-family: MJXc-TeX-math-BIx; src: local('MathJax\_Math'); font-weight: bold; font-style: italic}
@font-face {font-family: MJXc-TeX-math-BIw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Math-BoldItalic.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Math-BoldItalic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Math-BoldItalic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-sans-R; src: local('MathJax\_SansSerif'), local('MathJax\_SansSerif-Regular')}
@font-face {font-family: MJXc-TeX-sans-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_SansSerif-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_SansSerif-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_SansSerif-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-sans-B; src: local('MathJax\_SansSerif Bold'), local('MathJax\_SansSerif-Bold')}
@font-face {font-family: MJXc-TeX-sans-Bx; src: local('MathJax\_SansSerif'); font-weight: bold}
@font-face {font-family: MJXc-TeX-sans-Bw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_SansSerif-Bold.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_SansSerif-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_SansSerif-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-sans-I; src: local('MathJax\_SansSerif Italic'), local('MathJax\_SansSerif-Italic')}
@font-face {font-family: MJXc-TeX-sans-Ix; src: local('MathJax\_SansSerif'); font-style: italic}
@font-face {font-family: MJXc-TeX-sans-Iw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_SansSerif-Italic.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_SansSerif-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_SansSerif-Italic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-script-R; src: local('MathJax\_Script'), local('MathJax\_Script-Regular')}
@font-face {font-family: MJXc-TeX-script-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Script-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Script-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Script-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-type-R; src: local('MathJax\_Typewriter'), local('MathJax\_Typewriter-Regular')}
@font-face {font-family: MJXc-TeX-type-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Typewriter-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Typewriter-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Typewriter-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-cal-R; src: local('MathJax\_Caligraphic'), local('MathJax\_Caligraphic-Regular')}
@font-face {font-family: MJXc-TeX-cal-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Caligraphic-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Caligraphic-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Caligraphic-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-main-B; src: local('MathJax\_Main Bold'), local('MathJax\_Main-Bold')}
@font-face {font-family: MJXc-TeX-main-Bx; src: local('MathJax\_Main'); font-weight: bold}
@font-face {font-family: MJXc-TeX-main-Bw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Main-Bold.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Main-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Main-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-main-I; src: local('MathJax\_Main Italic'), local('MathJax\_Main-Italic')}
@font-face {font-family: MJXc-TeX-main-Ix; src: local('MathJax\_Main'); font-style: italic}
@font-face {font-family: MJXc-TeX-main-Iw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Main-Italic.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Main-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Main-Italic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-main-R; src: local('MathJax\_Main'), local('MathJax\_Main-Regular')}
@font-face {font-family: MJXc-TeX-main-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Main-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Main-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Main-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-math-I; src: local('MathJax\_Math Italic'), local('MathJax\_Math-Italic')}
@font-face {font-family: MJXc-TeX-math-Ix; src: local('MathJax\_Math'); font-style: italic}
@font-face {font-family: MJXc-TeX-math-Iw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Math-Italic.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Math-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Math-Italic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size1-R; src: local('MathJax\_Size1'), local('MathJax\_Size1-Regular')}
@font-face {font-family: MJXc-TeX-size1-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Size1-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Size1-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Size1-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size2-R; src: local('MathJax\_Size2'), local('MathJax\_Size2-Regular')}
@font-face {font-family: MJXc-TeX-size2-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Size2-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Size2-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Size2-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size3-R; src: local('MathJax\_Size3'), local('MathJax\_Size3-Regular')}
@font-face {font-family: MJXc-TeX-size3-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Size3-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Size3-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Size3-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size4-R; src: local('MathJax\_Size4'), local('MathJax\_Size4-Regular')}
@font-face {font-family: MJXc-TeX-size4-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Size4-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Size4-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Size4-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-vec-R; src: local('MathJax\_Vector'), local('MathJax\_Vector-Regular')}
@font-face {font-family: MJXc-TeX-vec-Rw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Vector-Regular.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Vector-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Vector-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-vec-B; src: local('MathJax\_Vector Bold'), local('MathJax\_Vector-Bold')}
@font-face {font-family: MJXc-TeX-vec-Bx; src: local('MathJax\_Vector'); font-weight: bold}
@font-face {font-family: MJXc-TeX-vec-Bw; src /\*1\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax\_Vector-Bold.eot'); src /\*2\*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax\_Vector-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax\_Vector-Bold.otf') format('opentype')}
: at 0, the model does not know the environment exists. At 1, it has a perfect model of the environment.
* alignment a: at 0, the model has a random objective. At 1, it has a perfect copy of the training objective.
* objective pointer o: at 0, there is no pointer. At 1, there is a perfectly robust pointer to the objective in the environment.
* power p: at 0, the model cannot do anything. At 1, the model is capable enough to achieve 0 training loss.
Let t be the index of the current training set, starting at 0. Suppose that the values of the various parameters are given by the following functions:
* et=log1000(t+1), representing diminishing returns for environment modeling.
* at=log1000(t+1), representing diminishing returns for aligning the proxy objective.
* ot=log500(t+1−min(t|et>0.5)), representing diminishing returns for building a robust pointer, which is only possible when the environment model is sufficiently detailed.
* pt=log1000(t+1), representing diminishing returns for increasing model power.
Given the above, e31>0.5, so the model will become a corrigibly aligned model before an internally aligned model in 530 timesteps. At this point, the model is not yet powerful enough to achieve zero training loss, so training will continue to 999 timesteps.
Suppose, however, that the model is capable of deception when et>0.6 and has enough power to deceive successfully at pt>0.7. The model will enter this deceptive regime on timestep 125, far before it becomes a corrigibly aligned. At this point, there will no longer be a marginal incentive to increase the alignment or objective pointer parameters, so they will plateau while power and environmental modeling parameters go to 1.

Obviously this example fails to capture a large number of dynamics that will exist during the training process. Hopefully, however, it conveys the rough picture of what we might expect the training process to look like.
Why this might be wrong
=======================
Adversarial training is fast
----------------------------
One assumption that goes into this argument is that internal alignment is relatively difficult. The reason to think this might be the case is that there are diminishing marginal returns to aligning the proxy objective with the base objective. While returns will diminish for any fixed training distribution, it might be possible to adjust the training distribution such that this does not occur.
In particular, one might be able to identify adversarial examples on which the proxy objective diverges from the base objective. By making a large fraction of the training distribution these dynamically-changing adversarial examples, one might preserve the size of the relative marginal benefit of aligning the proxy objective with the base objective.
For example, if a model is being trained on the objective of "pick apples" but it has instead learned the proxy objective "take round red things off trees", we can supply adversarial examples like red ornaments, lightbulbs, tomatoes, etc. By tuning the type of adversarial example to the specific proxy objective the model has learned, one might be able to generate a consistently strong gradient pointing towards internal alignment.
However, it seems like adversarial examples do not differentially favor alignment over deception. A deceptive model with a good understanding of the base objective will also perform better on the adversarial examples. In fact, the steady stream of adversarial examples might create a strong gradient towards understanding the base objective and increasing deceptive abilities.
Self-reifying objectives
------------------------
Some possible objective functions seem naturally lend themselves to being pointed at. For instance, [act-based](https://ai-alignment.com/corrigibility-3039e668638) agents that care about human short-term preferences have an incentive to point robustly at human preferences because that is what humans would prefer. In this sense, some objectives have broad targets; pointers in these targets naturally tend towards becoming more robust. We call these objectives *self-reifying.* (This is related to what Paul Christiano calls the [broad basin](https://ai-alignment.com/corrigibility-3039e668638) of corrigibility.)
If a model is being trained on an objective function that is self-reifying, a marginal increase in pointer robustness will incentivize the model to further increase pointer robustness. This might mean that a relatively poor pointer to the base objective might be sufficient for the agent to care about a robust version of the objective. The result is that the class of corrigibly aligned models might be reached by SGD faster than the model can become deceptive.
For example, if part of proper etiquette was listening very carefully to your teachers and trying to do what they mean, then a student that only cared a little bit about etiquette might still act as if they cared immensely. One might think of such objectives as able to "bootstrap misaligned optimization to alignment".
However, the fact that the objective is self-reifying also makes deceptive alignment easier. In particular, one doesn't seem to need to be able to fully model such objectives before being able to optimize for them relatively well. This makes deception possible before a full model of the base objective is learned.
As in our previous example, etiquette including "listen to your teachers" means a deceptively aligned model can instrumentally optimize for "listen to your teachers" instead of the full base objective. In general, one might note that anything that makes a base objective easier to point to will also increase the ease of pointing to the objective for the purposes of deceptive alignment.
Don't care about cross-episode utility
--------------------------------------
Models have instrumental incentives to be deceptive because they want to preserve their current objectives and optimize for them in the future. Such models must care about utility across multiple training episodes (cross-episode utility). If a model only cared about the current episode, then it would never sacrifice utility in the current model by being deceptive to increase its utility in further episodes. We can call models that care only about current episode rewards [myopic](https://www.alignmentforum.org/tag/myopia).
For example, if a model knew that it was in its last training episode and that it wasn't going to be deployed, it would clearly want to optimize for its proxy objective; there would be no particular reason to deceive the training process into thinking it was aligned. Myopic models always behave as if it's the end of training and they won't be deployed.
One reason to think that models will often care about cross episode rewards is that caring about the future is a natural generalization. In order for a reward function to be [myopic](https://www.alignmentforum.org/tag/myopia), it must contain machinery that does something like "care about X in situations C and not situations D", which is more complicated than "care about X".
For example, if a model cared about eating apples and it discovered that there were a bunch more apples located in so-called "other episodes" then it seems natural that the model would also care about these apples. The objective "care about apples in this episode but not others" is more complicated and thus less likely to be learned.
We are cautiously optimistic that further understanding of [myopia](https://www.alignmentforum.org/tag/myopia) is a promising research direction for detecting and preventing deceptive alignment.
Protecting proxy objective might be difficult
---------------------------------------------
If the model believes protecting its current proxy objective to be prohibitively difficult, then it will not behave deceptively. In practice, however, it doesn't seem like protecting a proxy objective will be difficult.
[Gradient hacking](https://www.lesswrong.com/posts/uXH4r6MmKPedk8rMA/gradient-hacking) is a strategy that a deceptively aligned model could use to preserve its current objective. The strategy is to make performance dependent on very precise values of the current proxy objective. Thus if the proxy objective ever gets changed, model performance will drastically drop across the board.
For example, if a model had the proxy objective of "eat apples", instead of using the hardcoded number n in other computations, the model could use n \* len(objective)/10. Thus, if the proxy objective was ever changed, many computations across the entire model would fail.
Additionally, note that even if preserving the proxy is extremely difficult, the model can believe it to be possible. For example, suppose a model is proxy aligned and would be deceptive, except that it believes proxy preservation is impossible. A relatively simple way to increase training performance might be to change the model's mind about the impossibility of proxy preservation. Thus SGD might modify the model to have such a belief, even if the belief is false.
Finally, the model does not have to believe it can preserve its proxy with high fidelity. If the model believes that its proxy objective will be preserved more by being deceptive relative to the alternative, then it still has an incentive to act deceptively.
For example, suppose a model is being trained on the base objective of “eat tomatoes” and its current proxy objective is “eat round red things”. If acting deceptively results in the model being deployed with the objective of “eat round red things with stems”, the model still has an incentive to act deceptively because there are round red things that have stems that are not tomatoes (a tomato kitchen timer, for example).
Deception is computationally expensive
--------------------------------------
A model that is deceptively aligned will be devoting some of its runtime to figuring out whether or not to defect. If the computational budget of a given model is fixed, this will result in less computation being used to optimize the training objective, potentially leading to worse performance. If the training environment is almost identical to the deployment environment, a deceptively aligned model would have to carefully scrutinize the environment to figure out when to defect. Such a model might waste enough time scrutinizing that it would perform worse enough that aligned models would dominate.
For example, suppose a model trained to drive a car and it's training environment is a perfect simulation of Earth using Newtonian mechanics. In order to know when to defect, the model might have to notice that its observations are actually obeying general relativity instead. This constant comparison might induce a large computational overhead, making such models dispreferred by SGD.
However, it seems likely that there exist checks that are computationally negligible. Since the model only has to realize it's being deployed **eventually**, the check can be one that will trigger with low probability for deployment environments and will never trigger in training environments. Such a check could also be made 100 times cheaper by only running it once every 100 episodes.
For example, if the model learns that humanity has not currently factored [RSA-2048](https://www.wikiwand.com/en/RSA_numbers#/RSA-2048), it might use the presence of a factorization as a trigger to defect. This trigger will not trigger immediately upon the model being deployed, but it seems likely to trigger eventually.
To the extent that defection checks are computationally expensive, limiting computation might marginally favor aligned models over deceptive ones.
Corrigibility is simplest
-------------------------
Given that the model's understanding of the environment will contain a copy of the base objective, model classes that make use of this information will have shorter description lengths than model classes that do not. In particular, this suggests that proxy aligned models will be more complicated than deceptively or corrigibly aligned models.
However, by the time a model knows enough about the world to be properly deceptive, it likely has a proxy objective of reasonable complexity. Since such a model will act to preserve its proxy, a deceptive model will be of similar complexity to an internally aligned model.
Concretely, an internally aligned model will have two copies of the base objective (one in the world and one as its objective), a corrigibly aligned model will have a copy of the base objective and a pointer, and a deceptively aligned model will have a copy of the base objective and a proxy objective. Since the proxy will be similar to the base objective, it will likely be more complicated than a pointer.
In some sense, the simplest model would be a deceptive model with an extremely simple proxy. However, there seems to be no plausible path locally favored by SGD that would result in this model. That being said, this observation suggests that heavily penalizing description length might still produce a deceptive model.
To the extent that SGD has [inductive biases](https://www.lesswrong.com/posts/nGqzNC6uNueum2w8T/inductive-biases-stick-around) that correspond to simplicity, corrigibly aligned models seems slightly favored over aligned or deceptive models.
Conclusion
==========
Internally aligned models care terminally about the base objective. Corrigibly aligned models care about a representation of the base objective. Deceptively aligned models instrumentally optimize for the training objective so they can defect later for higher reward.
Internally aligned and corrigibly aligned models are less numerous than deceptive models, so one's prior belief should be that deception is reasonably likely. SGD must therefore systematically disfavor deception for deceptive models to be unlikely. However, examining the dynamics of the training process suggests that SGD actually might systematically **favor** deception.
Given that adversarial examples are unable to prevent deceptively aligned models from achieving perfect training loss, we consider deceptive alignment to be the most concerning of the possible [inner alignment](https://www.alignmentforum.org/tag/inner-alignment) failures. We are excited about research about addressing deceptive alignment, such as [relaxed adversarial training](https://www.alignmentforum.org/posts/9Dy5YRaoCxH9zuJqa/relaxed-adversarial-training-for-inner-alignment), [myopia](https://www.alignmentforum.org/tag/myopia), and [corrigibility](https://www.alignmentforum.org/posts/BKM8uQS6QdJPZLqCr/towards-a-mechanistic-understanding-of-corrigibility).
---
1. Since not all models get deployed, consider this to mean “able to take unrecoverable action in the world”. [↩︎](#fnref-LEfJvDKAWCjjzs75z-1)
2. One might not think this is unlikely *a priori*, since SGD is likely to have a bias towards *something*. [↩︎](#fnref-LEfJvDKAWCjjzs75z-2)
3. For example, under certain assumptions, myopic training objectives do not favor deception. [↩︎](#fnref-LEfJvDKAWCjjzs75z-3)
4. Evan Hubinger calls the more general case of this phenomenon “suboptimality deceptive alignment” and explains more [here](https://www.alignmentforum.org/posts/iydwbZhATANhjoGP7/more-variations-on-pseudo-alignment). [↩︎](#fnref-LEfJvDKAWCjjzs75z-4) |
487960dd-92a6-4958-a7f7-7bb89cdb1083 | StampyAI/alignment-research-dataset/lesswrong | LessWrong | News : Biden-Harris Administration Secures Voluntary Commitments from Leading Artificial Intelligence Companies to Manage the Risks Posed by AI
Here are the extracts naming the companies and specific commitments :
As part of this commitment, President Biden is convening seven leading AI companies at the White House today – Amazon, Anthropic, Google, Inflection, Meta, Microsoft, and OpenAI – to announce that the Biden-Harris Administration has [secured voluntary commitments](https://www.whitehouse.gov/wp-content/uploads/2023/07/Ensuring-Safe-Secure-and-Trustworthy-AI.pdf) from these companies to help move toward safe, secure, and transparent development of AI technology.
Today, these seven leading AI companies are committing to:
**Ensuring Products are Safe Before Introducing Them to the Public**
* **The companies commit to internal and external security testing of their AI systems before their release.** This testing, which will be carried out in part by independent experts, guards against some of the most significant sources of AI risks, such as biosecurity and cybersecurity, as well as its broader societal effects.
* **The companies commit to sharing information across the industry and with governments, civil society, and academia on managing AI risks.**This includes best practices for safety, information on attempts to circumvent safeguards, and technical collaboration.
**Building Systems that Put Security First**
* **The companies commit to investing in cybersecurity and insider threat safeguards to protect proprietary and unreleased model weights.**These model weights are the most essential part of an AI system, and the companies agree that it is vital that the model weights be released only when intended and when security risks are considered.
* **The companies commit to facilitating third-party discovery and reporting of vulnerabilities in their AI systems.**Some issues may persist even after an AI system is released and a robust reporting mechanism enables them to be found and fixed quickly.
**Earning the Public’s Trust**
* **The companies commit to developing robust technical mechanisms to ensure that users know when content is AI generated, such as a watermarking system.**This action enables creativity with AI to flourish but reduces the dangers of fraud and deception.
* **The companies commit to publicly reporting their AI systems’ capabilities, limitations, and areas of appropriate and inappropriate use.**This report will cover both security risks and societal risks, such as the effects on fairness and bias.
* **The companies commit to prioritizing research on the societal risks that AI systems can pose, including on avoiding harmful bias and discrimination, and protecting privacy.**The track record of AI shows the insidiousness and prevalence of these dangers, and the companies commit to rolling out AI that mitigates them.
* **The companies commit to develop and deploy advanced AI systems to help address society’s greatest challenges.**From cancer prevention to mitigating climate change to so much in between, AI—if properly managed—can contribute enormously to the prosperity, equality, and security of all.
As we advance this agenda at home, the Administration will work with allies and partners to establish a strong international framework to govern the development and use of AI. It has already consulted on the voluntary commitments with Australia, Brazil, Canada, Chile, France, Germany, India, Israel, Italy, Japan, Kenya, Mexico, the Netherlands, New Zealand, Nigeria, the Philippines, Singapore, South Korea, the UAE, and the UK. The United States seeks to ensure that these commitments support and complement Japan’s leadership of the G-7 Hiroshima Process—as a critical forum for developing shared principles for the governance of AI—as well as the United Kingdom’s leadership in hosting a Summit on AI Safety, and India’s leadership as Chair of the Global Partnership on AI. We also are discussing AI with the UN and Member States in various UN fora. |
22ee2f07-aa30-47c6-890f-4aa9f5c2be40 | trentmkelly/LessWrong-43k | LessWrong | Fittingness: Rational success in concept formation
Epistemic status:
This post offers a philosophical argument for distinguishing truth and fittingness as parallel success concepts for two phases of rational inquiry. I have graduate degrees in philosophy, with relevant expertise in logic, type theory, epistemology, phenomenology, ontology, and philosophy of science. I am confident (0.9) about truth the central claims, and the fittingness of central concepts. At the end I include a brief and much more speculative section (0.5) on how these ideas might related to pattern recognition in the brain.
Intro
The concept of truth is widely held to be a central concept of success for rational inquiry. It captures the notion of accuracy of beliefs in modeling reality. In this post, I suggest that truth needs a companion concept of fittingness. I will argue, that as truth is the essential success concept for rational belief, so fittingness is needed as a success concept for rational concept formation (equivalently: ontology building or type assignment). Both truth and fittingness capture an idea of accuracy in world modeling, but they operate at different type theoretic levels. Roughly, fittingness concerns the accuracy or appropriateness of how we divide reality by attaching concepts to it, by typing it, how we cut reality at the seams with language. Truth captures the accuracy of statements (propositions) within a given ontology. In order for questions about the truth of a proposition to be meaningful fruitful and salient, the concepts in which the proposition is expressed must be sufficiently fitting. [Slider's comment below makes the point that truth conditions can be meaningful even when the concepts employed are not fitting. He gives the example of the concept grue from philosophy. This example also illustrates that even when a statement is truth evaluable it may not be fruitful or salient for the purpose of rational inquiry.]
Fittingness in Inquiry
There are two phases of rational inquiry. These two phases do not occ |
34bd3aa9-d37b-4df8-8735-ed0abc681fb0 | trentmkelly/LessWrong-43k | LessWrong | Covid 2/4: Safe and Effective Vaccines Aplenty
The vaccine data is in. It’s pretty great.
We now have six known safe and effective Covid-19 vaccines: Pfizer, Moderna, AstraZeneca, Novavax, Johnson & Johnson and Sputnik.
Pfizer and Moderna are amazingly safe and effective. They have been approved and are being distributed. The problem is that due to our unwillingness to properly fund the process of scaling up, we are stalling out at about 1.3 million doses per day for a population of 330+ million people each of whom needs two doses. That won’t do, the vaccine rollout is going about the way you would expect if you’ve been paying attention so far, and the English and other variant strains are rising fast.
The other vaccine approvals are not in. It’s pretty terrible.
Johnson & Johnson and Novavax are also safe and effective, far better than we need them to be. They both held onto their data longer than they would have in a sane world, and took longer to collect it than they would have in a sane world, but there’s nothing other than bureaucratic delays now standing in our way. Johnson & Johnson is expected to get its approval at some unspecified date, presumably this month. Yet there is no sense of urgency.
AstraZeneca is even more egregious. There are doses ready to go, it has been approved by major regulatory agencies elsewhere, and the main objections seem to be “the data isn’t from Americans” and “we’re mad at you about previous mistakes so we need to punish that somehow.” We are still going to wait until April or so for the American data, plus the delays after that, unless something changes and the new administration and/or the FDA see themselves as sufficiently blameworthy for not approving the vaccine. This badly needs more outrage. There’s an increasing amount, but so far it’s nowhere near enough.
Sputnik is the least egregious. Not using it is going to get a lot of people killed, but it’s a Russian project largely motivated by national pride, that never held out any hope of convincing us to trust it a |
bd2af89e-0d73-4250-970c-7ab7b90d5dc8 | trentmkelly/LessWrong-43k | LessWrong | Re: The Crux List
Really enjoyed reading Zvi's Crux List.
One really important takeaway of lists like this is "we don't know what we don't know". Namely, P(DOOM) depends intimately on a whole bunch of details about the universe that are mostly unknowable in advance.
In that spirit, here are some more cruxes that P(DOOM) depends on.
Does AI favor the attacker (in a military sense)?
Suppose I invented a weapon that was exactly like modern weapons, but with two differences:
1. It is 100x faster than existing weapons
2. It is 100x more precise than existing weapons.
Obviously such a weapon would favor the defender, right?
Which is harder to improve AI capabilities or AI ethics?
Suppose it is absolutely trivial to build an AI that is sentient, but extremely difficult to build an AI that is super-intelligent.
In that case, Alignment is super-easy because we can easily create a bunch of AIs with high moral worth without increasing the overall danger to humanity, right?
Suppose, on the other hand, that it is extremely easy to build AI that is super-intelligent but practically impossible to create AI with goals, sentience, etc.
In that case Alignment is super-easy because it will be easy to pull off a pivotal act, right?
Is AI inherently centralizing or decentralizing?
Suppose that AI is inherently centralizing (training runs require large quantities of compute, which gives the few largest players the biggest advantage).
This is good for alignment, because it means we don't have to worry about a lone madman doing something that endangers all of humanity, right?
Suppose on the other hand that AI is inherently decentralizing (because literally anyone can run an AI on their GPU).
This is good for Alignment because centralized power is generally bad, right?
How hard is it to emulate human minds?
If emulating human minds (say by training an LLM to imitate human speech) is trivial, this is good for Alignment because it means that it is easy to teach AI human values, right?
Su |
17f7e367-accd-4382-a1c7-783a86fb03b5 | trentmkelly/LessWrong-43k | LessWrong | Risques existentiels en Français
I've just been interviewed by Radio-Canada (in French) for their program "Dessine moi un Dimanche". There really wasn't enough time (the interview apparently lasted nine minutes; it felt like two), but I managed to touch upon some of the technology risks of the coming century (including AI).
The segment can be found here: http://www.radio-canada.ca/emissions/dessine_moi_un_dimanche/2012-2013/chronique.asp?idChronique=295886 |
86731980-b3a4-4dd0-ba2c-893b88e32c02 | trentmkelly/LessWrong-43k | LessWrong | How did experiments with lucid dreaming pan out?
I remember there was a group of enthusiasts here on Less Wrong who were interested in experimenting with lucid dreaming, including research on its impact on the quality of life. How did those experiments pan out? |
46c16a08-2e05-4074-8042-ae1386608df5 | trentmkelly/LessWrong-43k | LessWrong | On Deliberative Alignment
Not too long ago, OpenAI presented a paper on their new strategy of Deliberative Alignment.
The way this works is that they tell the model what its policies are and then have the model think about whether it should comply with a request.
This is an important transition, so this post will go over my perspective on the new strategy.
Note the similarities, and also differences, with Anthropic’s Constitutional AI.
HOW DELIBERATIVE ALIGNMENT WORKS
> We introduce deliberative alignment, a training paradigm that directly teaches reasoning LLMs the text of human-written and interpretable safety specifications, and trains them to reason explicitly about these specifications before answering.
>
>
> We used deliberative alignment to align OpenAI’s o-series models, enabling them to use chain-of-thought (CoT) reasoning to reflect on user prompts, identify relevant text from OpenAI’s internal policies, and draft safer responses.
>
> Our approach achieves highly precise adherence to OpenAI’s safety policies, and without requiring human-labeled CoTs or answers. We find that o1 dramatically outperforms GPT-4o and other state-of-the art LLMs across a range of internal and external safety benchmarks, and saturates performance on many challenging datasets.
>
> We believe this presents an exciting new path to improve safety, and we find this to be an encouraging example of how improvements in capabilities can be leveraged to improve safety as well.
How did they do it? They teach the model the exact policies themselves, and then the model uses examples to teach itself to think about the OpenAI safety policies and whether to comply with a given request.
> Deliberate alignment training uses a combination of process- and outcome-based supervision:
* We first train an o-style model for helpfulness, without any safety-relevant data.
* We then build a dataset of (prompt, completion) pairs where the CoTs in the completions reference the specifications. We do this by inserting the |
53b4f277-2a6b-45cb-b276-dd67d12fa7de | trentmkelly/LessWrong-43k | LessWrong | Meetup : Washington, D.C.: Create and Complete
Discussion article for the meetup : Washington, D.C.: Create and Complete
WHEN: 21 June 2015 03:00:00PM (-0400)
WHERE: National Portrait Gallery
We will congregate as usual in the courtyard between 3:00 and 3:30 p.m., to be ready to start at 3:30 p.m.
This week's meeting is a Schelling point for people to work on things. Please bring projects you'd like to work on, help people with theirs, or just come hang out.
Upcoming meetups:
* Jun 28: Fun & Games
* Jul 5: American History
* Jul 12: Fermi Estimates
* Jul 19: Fun & Games
Discussion article for the meetup : Washington, D.C.: Create and Complete |
823a8440-e4d3-49a3-8d0f-968f9732ffa0 | StampyAI/alignment-research-dataset/lesswrong | LessWrong | TDT for Humans
This is part 19 of 30 of Hammertime. Click [here](https://radimentary.wordpress.com/2018/01/29/hammertime-day-1-bug-hunt/) for the intro.
As is Hammertime tradition, I’m making a slight change of plans right around the scheduled time for Planning. My excuse this time:
Several commenters pointed out serious gaps in my knowledge of [Focusing](https://radimentary.wordpress.com/2018/02/25/focusing/). I will postpone Internal Double Crux, an advanced form of Focusing, to the next cycle. Instead, we will have two more posts on making and executing long-term plans.
Day 19: TDT for Humans
======================
Previously on planning: [Day 8](https://radimentary.wordpress.com/2018/02/04/hammertime-day-8-sunk-cost-faith/), [Day 9](https://radimentary.wordpress.com/2018/02/06/hammertime-day-9-time-calibration/), [Day 10](https://radimentary.wordpress.com/2018/02/07/hammertime-day-10-murphyjitsu/).
Today I’d like to describe two orders of approximation to a working decision theory for humans.
TDT 101
-------
Background reading: [How I Lost 100 Pounds Using TDT](https://www.lesserwrong.com/posts/scwoBEju75C45W5n3/how-i-lost-100-pounds-using-tdt).
> Choose as though controlling the logical output of the abstract computation you implement, including the output of all other instantiations and simulations of that computation.
> ~ Eliezer
In other words, every time you make a decision, pre-commit to making the same decision in all conceptually similar situations in the future.
The striking value of TDT is: make each decision *as if you would immediately reap the long-term rewards of making that same decision repeatedly*. And if it turns out you’re an updateless agent, this actually works! You actually lose 100 pounds by making one decision.
I encourage readers who have not tried to live by TDT to stop here and try it out for a week.
TDT 201
-------
There are a number of serious differences between timeless agents and human beings, so applying TDT as stated above requires an unacceptable (to me) level of self-deception. My second order of approximation is to offer a practical and weak version of TDT based on the [Solitaire Principle](https://radimentary.wordpress.com/2018/01/16/the-solitaire-principle-game-theory-for-one/) and [Magic Brain Juice](https://radimentary.wordpress.com/2018/01/26/magic-brain-juice/).
Three objections to applying TDT in real life:
### Spirits
A human is about halfway between “one monolithic codebase” and “a loose confederation of spirits running a random serial dictatorship.” Roughly speaking, each spirit is the piece of you built to satisfy one primordial need: hunger, friendship, curiosity, justice. At any given time, only one or two of these spirits is present and making decisions. As such, even if each individual spirit as updateless and deterministic, you don’t get to make decisions for all the spirits currently inactive. You don’t have as much control over the other spirits as you would like.
Different spirits have access to different data and beliefs. I’ve mentioned, for example, that I have different personalities speaking Chinese and English. You can ask me what my favorite food is in English, and I’ll say dumplings, but the true answer 饺子 feels qualitatively better than dumplings by a wide margin.
Different spirits have different values. I have two friends who reliably provoke my “sadistic dick-measuring asshole” spirit. If human beings really have utility functions this spirit has negative signs in front of the terms for other people. It’s uncharacteristically happy to engage in negative-sum games.
It’s almost impossible to predict when spirits will manifest. Recently, I was on a 13-hour flight back from China. I started marathoning Game of Thrones after exhausting the comedy section, and a full season of Cersei Lannister left me in “sadistic asshole” mode for a full day afterwards. If Hainan Airlines had stocked more comedy movies this might not have occurred.
Spirits can lay dormant for months or years. Meeting up with high school friends this December, I fell into old [roles](https://www.lesserwrong.com/posts/AqbWna2S85pFTsHH4/the-intelligent-social-web) and received effortless access to a large swathe of faded memories.
### Conceptual Gerrymandering
Background reading: [conceptual gerrymandering](https://thicketforte.com/2017/04/07/terrorism-and-conceptual-gerrymandering/).
> I can make a problem look either big or small by drawing either a big or small conceptual boundary around it, then identifying my problem with the conceptual boundary I’ve drawn.
TDT runs on an ambiguous “conceptual similarity” clause: you pre-commit to making the same decision in conceptually similar situations. Unfortunately, you will be prone to motivated reasoning and conceptual gerrymandering to get out of timeless pre-commitments made in the past.
This problem can be reduced but not solved by clearly stating boundaries. Life is too high-dimensional to even figure out what variables to care about, let alone where to draw the line for each of them. What information becomes salient is a function of your attention and noticing skills as much as of reality itself. These days, it’s almost a routine experience to read an article that sufficiently alters my capacities for attention as to render situations I would previously have considered “conceptually similar” altogether distinct.
### Magic Brain Juice
Background reading: [Magic Brain Juice](https://radimentary.wordpress.com/2018/01/26/magic-brain-juice/).
> Every action you take is accompanied by an unintentional self-modification.
The human brain is finicky code that self-modifies every time it takes an action. The situation is even worse than this: your actions can shift your very values in surprising and illegible ways. This bug is an inherent contradiction to applying TDT as a human.
Self-modification happens in multiple ways. When I wrote Magic Brain Juice, I was referring to the immediate strengthening of neural pathways that are activated, and the corresponding decay through time of all pathways not activated. But other things happen under the hood. You get attached to a certain identity. You get sucked into the nearest attractor in the social web. And also:
[Exposure therapy](https://radimentary.wordpress.com/2018/02/02/hammertime-day-5-comfort-zone-expansion/) is a powerful and indiscriminate tool. You can reduce any aversion to almost zero just by voluntarily confronting it repeatedly. But you have fears and aversions in every direction!
Every move you make is exposure therapy in that direction.
That’s right.
Every voluntary decision nudges your comfort zone in that direction, squashing aversions (endorsed or otherwise) in its path.
Oops!
Solutions
---------
I hope I’ve convinced you that the human brain is sufficiently broken that our intuition about “updateless source code” don’t apply, and trying to make decisions from TDT will be harder (and may have serious unintended side effects) as a result. What can be done?
First, I think it’s worth directly investing in TDT-like behaviors. Make conscious decisions to reinforce the spirits that are amenable to making and keeping pre-commitments. Make more legible decisions and clearly state conceptual boundaries. Explore virtue ethics or deontology. [Zvi’s blog](https://thezvi.wordpress.com/) is a good a place.
In the same vein, practice predicting your future behavior. If you can become your own Omega, problems you face start looking Newcomb-like. Then you’ll be forced to give up CDT and the failures it entail.
Second, I once proposed a model called the “Ten Percent Shift”:
> The Ten Percent Shift is a thought experiment I’ve successfully pushed to System 1 that helps build long-term habits like blogging every day. It makes the assumption that each time you make a choice, it gets 10% easier.
> Suppose there is a habit you want to build such as going to the gym. You’ve drawn the pentagrams, sprinkled the pixie dust, and done the proper rituals to decide that the benefits clearly outweigh the costs and there’s no superior alternatives. Nevertheless, the effort to make yourself go every day seems insurmountable.
> You spend 100 units of willpower dragging yourself there on Day 1. Now, notice that you have magic brain juice on your side. On Day 2, it gets a little bit easier. You spend 90 units. On Day 3, it only costs 80.
> A bit of math and a lot of magic brain juice later, you spend 500 units of willpower in the first 10 days, and the habit is free for the rest of time.
The exact number is [irrelevant](http://slatestarcodex.com/2013/05/02/if-its-worth-doing-its-worth-doing-with-made-up-statistics/), but I stand by this model as the proper weakening of TDT: *act as if each single decision rewards you with 10% of the value of making that same decision indefinitely*. One decision only loses you 10 pounds, and you need to make 10 consecutive decisions before you get to reap the full rewards.
The Ten Percent Shift guards against spirits. Once you make the same decision 10 times in a row, you’ll have made it from a wide range of states of mind, and the exact context will have differed in every situation. You’ll probably have to convince a majority of spirits to agree with making the decision.
The Ten Percent Shift also guards against conceptual gerrymandering. Having made the same decision from a bunch of different situations, the convex hull of these data points is a 10-dimensional convex region that you can unambiguously stake out as a timeless pre-commitment.
Daily Challenge
===============
This post is extremely tentative and theoretical, so I’ll just open up the floor for discussion. |
daed58cf-0cb2-484a-9d94-bf489db1eec2 | trentmkelly/LessWrong-43k | LessWrong | Invulnerable Incomplete Preferences:
A Formal Statement
Produced as part of the SERI ML Alignment Theory Scholars Program - Summer 2023 Cohort. My thanks to Eric Chen, Elliott Thornley, and John Wentworth for invaluable discussion and comments on earlier drafts. All errors are mine.
This article presents a few theorems about the invulnerability of agents with incomplete preferences. Elliott Thornley’s (2023) proposed approach to the AI shutdown problem relies on these preferential gaps, but John Wentworth and David Lorell have argued that they make agents play strictly dominated strategies.[1] I claim this is false.
Summary
Suppose there exists a formal description of an agent that willingly shuts down when a certain button is pressed. Elliott Thornley’s (2023) Incomplete Preference Proposal aims to offer such a description. It’s plausible that, for it to constitute a promising approach to solving the AI shutdown problem, this description also needs to (i) permit the agent to be broadly capable and (ii) assure us that the agent will remain willing to shut down as time passes. This article formally derives a set of sufficient conditions for an agent with incomplete preferences to satisfy properties relevant to (i) and (ii).
A seemingly relevant condition for an agent to be capably goal-directed is that it avoids sequences of actions that foreseeably leave it worse off.[2] I will say that an agent satisfying this condition is invulnerable. This is related to two intuitive conditions. The weaker one is unexploitability: that the agent cannot be forcibly money pumped (i.e., compelled by its preferences to sure loss). The stronger condition is opportunism: that the agent never accepts sure losses or foregoes sure gains.[3]
To achieve this, I propose a dynamic choice rule for agents with incomplete preferences. This rule, Dynamic Strong Maximality (DSM), requires that the agent consider the available plans that are acceptable at the time of choosing and, among these, pick any plan that wasn’t previously strictly dominated |
1ecabbf4-0a3e-4031-af58-31c59f8655fe | trentmkelly/LessWrong-43k | LessWrong | Using humility to counteract shame
"Pride is not the opposite of shame, but its source. True humility is the only antidote to shame."
Uncle Iroh, "Avatar: The Last Airbender"
Shame is one of the trickiest emotions to deal with. It is difficult to think about, not to mention discuss with others, and gives rise to insidious ugh fields and negative spirals. Shame often underlies other negative emotions without making itself apparent - anxiety or anger at yourself can be caused by unacknowledged shame about the possibility of failure. It can stack on top of other emotions - e.g. you start out feeling upset with someone, and end up being ashamed of yourself for feeling upset, and maybe even ashamed of feeling ashamed if meta-shame is your cup of tea. The most useful approach I have found against shame is invoking humility.
What is humility, anyway? It is often defined as a low view of your own importance, and tends to be conflated with modesty. Another common definition that I find more useful is acceptance of your own flaws and shortcomings. This is more compatible with confidence, and helpful irrespective of your level of importance or comparison to other people. What humility feels like to me on a system 1 level is a sense of compassion and warmth towards yourself while fully aware of your imperfections (while focusing on imperfections without compassion can lead to beating yourself up). According to LessWrong, "to be humble is to take specific actions in anticipation of your own errors", which seems more like a possible consequence of being humble than a definition.
Humility is a powerful tool for psychological well-being and instrumental rationality that is more broadly applicable than just the ability to anticipate errors by seeing your limitations more clearly. I can summon humility when I feel anxious about too many upcoming deadlines, or angry at myself for being stuck on a rock climbing route, or embarrassed about forgetting some basic fact in my field that I am surely expected to know by th |
5e2d31f3-3227-48a9-bf86-c351dbf20537 | StampyAI/alignment-research-dataset/special_docs | Other | A Regression Approach for Modeling Games with Many Symmetric Players.
A Regression Approach for Modeling Games with Many Symmetric Players
Bryce Wiedenbeck
Swarthmore College
bwieden1@swarthmore.eduFengjun Yang
Swarthmore College
fyang1@swarthmore.eduMichael P. Wellman
University of Michigan
wellman@umich.edu
Abstract
We exploit player symmetry to formulate the representa-
tion of large normal-form games as a regression task. This
formulation allows arbitrary regression methods to be em-
ployed in in estimating utility functions from a small subset of
the game’s outcomes. We demonstrate the applicability both
neural networks and Gaussian process regression, but focus
on the latter. Once utility functions are learned, computing
Nash equilibria requires estimating expected payoffs of pure-
strategy deviations from mixed-strategy profiles. Computing
these expected values exactly requires an infeasible sum over
the full payoff matrix, so we propose and test several approxi-
mation methods. Three of these are simple and generic, appli-
cable to any regression method and games with any number
of player roles. However, the best performance is achieved
by a continuous integral that approximates the summation,
which we formulate for the specific case of fully-symmetric
games learned by Gaussian process regression with a radial
basis function kernel. We demonstrate experimentally that the
combination of learned utility functions and expected payoff
estimation allows us to efficiently identify approximate equi-
libria of large games using sparse payoff data.
The study of intelligent agents naturally gives rise to ques-
tions of how multiple agents will interact. Game-theoretic
models offer powerful methods for reasoning about such in-
teractions and have therefore become a key component of
the AI toolkit. Like many core AI methods, game-theoretic
analysis has benefited from recent advances in machine
learning algorithms and from the availability of ever-larger
data sets. Most notably, automated playing and solving of
extremely large extensive-form games has advanced con-
siderably with the aid of learning algorithms that general-
ize across game states. In this paper, we consider extremely
large normal-form games, and show that learning algorithms
generalizing across symmetric players can help to compactly
represent and efficiently solve such games.
In both normal and extensive form games, the chief ob-
stacle to computing equilibria is the enormous size of the
standard input representation (Papadimitriou and Roughgar-
den 2005; Bowling et al. 2015). In games like poker and
go, the extensive form representation is far too big to be
constructed explicitly, but observations of particular states
Copyright c
2018, Association for the Advancement of Artificial
Intelligence (www.aaai.org). All rights reserved.of the game can be straightforwardly generated by simula-
tion. Using simulated data from a small subset of the game’s
states, learning techniques have succeeded in extracting gen-
eral principles that apply to never-before-seen parts of the
game. Agents employing such learning have achieved high-
level play (Silver et al. 2016; Morav ˇc´ık et al. 2017) and iden-
tified approximate equilibria (Heinrich, Lanctot, and Silver
2015).
In normal-form games, the representational complexity
derives principally from the number of players. A standard
payoff matrix representation of a game with jPjplayers and
jSjstrategies per player records a vector of payoff values
in each ofjSjjPjcells. In many AI applications, symme-
tries among agents permit some economy of representation,
but even a fully symmetric game must record payoffs for jPj+jSj 1
jPj
distinct outcomes (Cheng et al. 2004). As a re-
sult, analysts are often restricted to studying games with
small numbers of players and strategies, or with special-
purpose compact representations.
In the present work, we use data about a small subset of
a game’s outcomes as input to learning methods that seek
to extract general principles of the game’s utility function.
Data of this form is common in the field of empirical game-
theoretic analysis (Wellman 2006). EGTA is used to study
multi-agent interactions where a closed-form game model is
unavailable or intractable. In such settings, an agent-based
model can often capture key features of the interaction and
provide data about agent incentives. Generally, a single run
of an agent-based simulation generates a noisy sample of
each payoff value in one cell of the payoff matrix. Set-
tings where EGTA has been employed include identifying
bidding strategies in continuous double auctions (Phelps,
Marcinkiewicz, and Parsons 2006), designing network rout-
ing protocols (Wellman, Kim, and Duong 2013), under-
standing credit provision in the absence of a central currency
(Dandekar et al. 2015), and finding ways to mitigate the ef-
fects of high-frequency trading (Wah and Wellman 2016).
Because of the combinatorial growth of the payoff matrix,
filling every cell by simulation quickly becomes intractable
as the number of players and strategies grows. Early EGTA
studies were often restricted to small numbers of players and
strategies (Phelps, Marcinkiewicz, and Parsons 2006; Walsh
et al. 2002). More recently, growth in the strategy space
has been kept in check by iterative exploration approaches,
and games with many players have been approximated using
player reduction methods (Wellman, Kim, and Duong 2013;
Wah, Hurd, and Wellman 2015). Our work presents an alter-
native to player reduction, where the use of machine learn-
ing allows for better generalization from more flexible data
sets, resulting in better approximations of large symmetric
games.
The key to learning that generalizes across outcomes is
that games with large numbers of players generally exhibit
substantial structure beyond player symmetries. With many
players, we often expect that no single opponent can unilat-
erally exert an outsized influence on a player’s payoff. For
example, in a 100-player game, the difference in payoff be-
tween 36 or 37 opponents choosing a particular strategy is
likely to be small. A related notion of bounded influence
is formalized and studied by Kearns and Mansour (2002).
Further, when a game represents interactions among com-
putational agents, and the payoff matrix is too big to be rep-
resented, we should expect that the agents themselves are
not reasoning about the full game, but rather some more-
compact summarization. As a result, in many large games of
interest, payoff functions will exhibit smoothness and sim-
plicity that make them amenable to machine learning.
Related Work
Much previous work in machine learning for game analysis
has focused on extensive-form games like poker (Heinrich,
Lanctot, and Silver 2015; Morav ˇc´ık et al. 2017). In these
settings, the game being studied is precisely defined, but too
large to be analyzed directly, so learning is used to express
strategies compactly and estimate expected payoffs in var-
ious game states without performing an exhaustive search.
Our work is motivated by empirical settings where the game
model, in addition to being extremely large, is initially un-
known and must be induced from data. As is common in
such settings, we treat the set of strategies as fixed and ex-
ogenously specified. Further, our methods treat strategies as
categorical, requiring no relationship among the strategies to
make generalizations.
We are interested analyzing in games with a sufficiently
large number of players to pose representational challenges
even if the set of strategies can be fully enumerated. A
common approach to analyzing normal-form games with
many players is to employ game models with compact rep-
resentations that can be analyzed directly. Examples in-
clude potential games (Monderer and Shapley 1996), graph-
ical games (Kearns 2007), and action-graph games (Jiang,
Leyton-Brown, and Bhat 2011). In an empirical setting,
where the game model is not known in advance, such repre-
sentations are difficult to apply, but some researchers have
investigated learning compact representations from data.
Duong et al. (2009) developed a method for detecting graph-
ical structures representing independences among players’
payoffs. Honorio and Ortiz (2015) likewise learn graphical
game models, in their case from observations of play rather
than payoffs, using assumptions about the structure of utility
functions and the way that play is generated conditional on
the actual payoff function. Neither of these approaches canbe applied to symmetric games, because graphical games
derive their compactness from player independence, which
cannot arise in non-trivial symmetric games.
An alternative approach that has been used when EGTA
environments simulate a large number of symmetric play-
ers is called player reduction . Player reductions define a
reduced game with a small number of players, and fill in
the reduced game’s payoff matrix using data from the full
game by aggregating the decisions of several symmetric
players. Equilibria are then computed in the reduced game
and treated as approximate equilibria of the full game. Sev-
eral player reduction methods have been proposed, varying
on the choice of full-game profiles to simulate and how they
map them to payoffs in the reduced game.
The first player-reduction method to see widespread use
was hierarchical reduction (Wellman et al. 2005), which
treats each reduced-game player as controlling an equal
fraction of the full-game agents. Hierarchical reduction
has been largely supplanted by a more recent technique
called deviation-preserving reduction (DPR) (Wiedenbeck
and Wellman 2012). DPR treats each player in the reduced
game as controlling the strategy choice of a single agent
in the full game, but each player views its opponents as an
aggregation of the remaining full-game agents. This means
that payoff differences resulting from a single reduced-game
player switching strategies reflect payoff changes from a
single full-game agent deviating, making DPR equilibria
more reflective of full-game equilibria. We employ DPR as a
benchmark against which to compare our learning methods.
Relative to DPR or other player reductions, our methods
take advantage of greater flexibility in allowable input data.
Player reduction methods prescribe a fixed set of profiles to
simulate, and to ensure accurate estimates of reduced game
payoffs, users often end up simulating the same profile many
times. Our learned models do not require a fixed set of full-
game profiles and can therefore spread sampling effort over
a wider variety of profiles. DPR also ignores some freely
available data: simulating profiles with many strategies but
using the payoff data for only one strategy. Our regressions
can always make use of any data that is available.
Most closely related to the present work, V orobeychik,
Wellman, and Singh (2007) demonstrated the use of regres-
sion methods to learn payoff functions over continuous strat-
egy spaces. The present paper can be viewed as extend-
ing their work to the domain of categorical strategies. Their
learning methods relied on strategy sets that were fully de-
scribed by varying continuous parameters, such as bids in
a single-unit auction. By contrast, our methods can handle
arbitrary sets of strategies, relying instead on the game hav-
ing a large number of symmetric players, which is typical of
environments defined by agent-based simulations.
Background and Notation
Game Theory
We focus on games represented in normal form that have
significant symmetry among players. In the following pre-
sentation, we focus on fully-symmetric games, where all
players have the same set of strategies and face the same
set of incentives. However, most of our models generalize
straightforwardly to role-symmetric games, where players
are partitioned into some number of roles (such as buyers
and sellers), and players are symmetric within, but not across
roles. Formally, a symmetric game consists of:
a set of players P
a set of strategies S
a utility function u:S~S!R
Aprofile~ sis an assignment of one strategy to every
player. Because players are symmetrical, we can represent
a profile by a vector of the number of players choosing each
strategy. We denote the set of all profiles ~S. The utility func-
tionu(s;~ s)maps a profile ~ sand a strategy sto the utility
of a player choosing strategy swhen players jointly choose
profile~ s. A mixed strategy specifies a probability distri-
bution over a player’s strategies. A symmetric mixture ~ is
a common mixed strategy played by all agents.
A player selecting pure strategy swhen other players
jointly play according to symmetric mixture ~ receives an
expected payoff:
u(s;~ ) =X
~ s2~SPr[~ sj~ ]u(s;~ s) (1)
A player from playing according to a symmetric mixture ~
receives expected payoff:
u(~ ) =X
s2S~ (s)u(s;~ )
The regret of a symmetric mixture is the maximum
amount that any player could gain by deviating to a pure
strategy:
regret (~ ) = max
s2Su(s;~ ) u(~ )
A symmetric Nash equilibrium is a symmetric mixture
withregret (~ ) = 0 . It follows from a proof by Nash (1951)
that a (role-) symmetric game must have a (role-) symmetric
Nash equilibrium, but finding a Nash equilibrium is com-
putationally hard. Typical analysis of normal form games
seeks an-Nash equilibrium, a profile ~ withregret (~ ).
In this paper, we measure our success in approximating
large games in two ways. First, we compare expected pay-
offsu(s;~ )estimated by our model to ground truth ex-
pected payoffs in the game being learned, averaged over a
wide range of symmetric mixtures. Second, we identify -
Nash equilibria in our learned models and compute their re-
gret in the ground-truth game. In the experiments presented
here, we compute symmetric mixed-strategy equilibria using
replicator dynamics (Gintis 2009), but we see similar results
when computing equilibria with fictitious play.
Gaussian Process Regression
Gaussian process regression (GPR) is a flexible method for
supervised learning (Rasmussen and Williams 2006) that
learns a mapping from input ~ xito outputyi. The set ofnin-
put points of dimension dcan be collected into an ndma-
trixX, and the corresponding targets into the n-dimensionalcolumn vector ~ y. GPR estimates the value at a new point ~ x
ask(~ x)K 1~ y, wherekandKare defined as:
k(~ x)[k(~ x1;~ x);k(~ x2;~ x);:::;k (~ xn;~ x)]
K2
664k(~ x1;~ x1)k(~ x1;~ x2)k(~ x1;~ xn)
k(~ x2;~ x1)k(~ x2;~ x2)k(~ x2;~ xn)
............
k(~ xn;~ x1)k(~ xn;~ x2)k(~ xn;~ xn)3
775
The kernel function k(;)must be specified by the user;
we use the radial basis function (RBF) kernel:
k
~ a;~b
=cexp
1
2l2
~ a ~b
2
2
(2)
The RBF kernel has an important hyperparameter l, the
length-scale over which the function is expected to vary.
This can be estimated by MLE, but in our experiments, we
found it important to constrain lto the range [ 1;jPj], and
that a length scale close to these bounds was sometimes ev-
idence of a poor fit.
Distributions
We denote a multivariate Gaussian distribution with mean
vector~ and covariance matrix asN(j~ ;). The prod-
uct of two Gaussians can be rewritten in the following way
(Petersen and Pedersen 2008):
N(~ xj~ 1;1)N(~ xj~ 2;2) =
N(~ 1j~ 2;1+2)N(~ xj~ 3;3)(3)
where~ 3and3are defined as
~ 33(1 1~ 1+2 1~ 2)
3(1 1+2 1) 1
We denote the multinomial distribution of ndraws from
discrete distribution ~ pasM(jn;~ p). Whennis large and~ p
is not near the edge of the domain, a multinomial distribution
is well-approximated by the following Gaussian distribution
(Severini 2005):
M(n;p)N(np;M ) (4)
whereMis defined as
M2
664p1 p2
1 p1p2 p1pn
p2p1p2 p2
2 p2pn
............
pnp1 pnp2pn p2
n3
775
Methods
Our method for approximating normal-form games relies on
two key steps. First, we use regression to learn a mapping
from pure-strategy profiles to payoffs. This mapping allows
us to generalize from a small data set to functions that can
be efficiently queried for arbitrary profiles. Second, we use
queries to these utility functions to estimate expected pay-
offs of playing each pure strategy against a symmetric mix-
ture. These expected payoff estimates enable us to compute
symmetric mixed-strategy -Nash equilibria.
(a) Generic methods.
0:1 0:5 0:9
probability of strategy 1expected payotrue game
point
neighbor
sample
(b) GP-specific integration method.
0:1 0:5 0:9
probability of strategy 1expected payotrue game
integration
Figure 1: Comparing methods for estimating the expected
payoff to strategy 1 in a 100-player, 2-strategy game learned
from complete data with GPR. Zoom recommended.
Payoff Learning
The key insight that enables us to learn payoffs from player
symmetry is that strategy profiles in a (role-) symmetric
game can be encoded as a vector of strategy counts. For
each strategy s, an entry in this vector encodes the number
of players selecting strategy s. By representing each strategy
as a separate dimension of the regression input, this method
does not emphasize generalizing across strategies. Instead,
it allows us to learn general effects caused by many play-
ers selecting the same strategy. Such effects are common
in the literature, appearing in congestion games, local ef-
fect games, and more-specialized compact representations
of large games
Further emphasizing the categorical nature of normal-
form game strategies, we run a separate regression for each
strategy’s utility function. This serves two purposes: first, it
lowers the overall runtime of the regression, and second, it
allows us to specialize the data set. Given a collection of pro-
files and corresponding payoff values, we construct the data
set for strategy sby selecting all profiles (and corresponding
payoffs) where at least one player chooses s. Any regres-
sion method can be run on this data set; we focus our testing
on Gaussian process regression, but have also run proof-of-
concept tests using neural networks.
Data Selection In initial testing, we found that our meth-
ods produced low average error in estimating expected util-
ities, but surprisingly poor results in identifying Nash equi-
libria. The root cause of this problem was inaccurate regres-sion estimates near the edges of the profile space (zero or
one players selecting a strategy). Such profiles were rare in
randomly-generated data sets, but are extremely important
in computing equilibria, because in most equilibria, only a
small number of strategies are played with non-zero prob-
ability (Porter, Nudelman, and Shoham 2008). As a result,
the data set for all experiments that follow include a large
over-representation of profiles in which zero or one players
play various strategies.
Expected Payoff Estimation
Given a learned payoff model for a symmetric game, we
want to identify symmetric mixed-strategy -Nash equilib-
ria. The critical input to computing equilibria is the expected
payoff of playing pure strategy sagainst opponents jointly
following a symmetric mixture ~ , given by equation 1. Com-
puting this expectation exactly requires summing over all
profiles in the game, and is therefore infeasible in large
games. We propose and evaluate several methods for esti-
mating expected payoffs without computing the full sum.
Generic Methods We consider three methods for estimat-
ing expected payoffs that are applicable regardless what re-
gression method was used to learn utility functions. The first
method, sampling , selectskrandom profiles according the
distribution~ siPr[~ sj~ ], and computes the average payoff:
u(s;~ )1
kkX
i=1^u(s;~ si)
Where ^uis the regression estimate. The second method,
point , queries the learned function only at the modal profile:
u(s;~ )^u(s;jPj~ )
The third method neighbor computes a weighted sum over
just the profiles within ddeviations of the maximum-
likelihood profile. Letting sbe the maximum-likelihood pro-
file, we define the set N=f~ sjk~ s sk1dg. The neigh-
bor estimate is then:
u(s;~ )1P
~ s2NPr[~ sj~ ]X
~ s2NPr[~ sj~ ]^u(s;~ s)
All three generic methods have strengths and weaknesses.
Sampling provides an unbiased estimator, and is correct in
the limit as k! 1 . However, it provides unstable es-
timates that are unsuitable for most algorithms that com-
pute Nash equilibria. Point estimation is fast, requiring far
fewer queries to the regression model than any other method.
It also provides smoothly-varying estimates (subject to the
smoothness of the regression model) that are correct in
the limit asjPj ! 1 . However, its payoff estimates ex-
hibit bias that can interfere with equilibrium computation or
other analysis. Neighbor estimation provides a sort of mid-
dle ground, in that avoids the randomness of sampling and
has lower bias than point estimation. When d= 0neighbor
approximates point, and when d=jPj, neighbor computes
the exact expected payoff, so dcan be chosen to trade off ac-
curacy with computation time. Unfortunately, neighbor suf-
fers from discrete-steps in its estimates as the maximum-
likelihood profile changes. This problem, illustrated in fig-
ure 1a, can occasionally prevent iterative methods for equi-
librium computation from converging.
Continuous Approximation A natural approach to the
problem of estimating a sum over a ver large number of low-
probability terms is to approximate the summation with an
integral. In this section we show how this can be done for
a fully-symmetric game learned using Gaussian process re-
gression with a radial basis function kernel. In the case of a
one-role game, the probability of a profile can be expressed
as a multinomial, Pr[~ sj~ ] =M(~ sjn 1;), so we can
re-write equation 1 as:
u(s;~ ) =X
~ s2~SM(~ sjn 1;)u(~ s;s)
X
~ s2~SM(~ sjn 1;)k(~ s)K 1~ y (5)
=~kK 1~ y (6)
Equation 5 employs the GPR payoff estimate for strategy
s. We reach equation 6 by noting that K 1~ ydoes not depend
on~ s, so it can be pulled out of the sum, and by defining ~kP
~ s2~SM(~ sjn 1;)k(~ s). We next consider component
iof this vector, ~ki. DefiningC1in terms of the mixture :
C12
6641 2
1 12 1n
212 2
2 2n
............
n1 n2n 2
n3
775
lets us use equation 4 to approximate ~kiwith a Gaussian.
Since we now have a continuous distribution, the summa-
tion can be approximated by an integration over the profile
simplex.
~kiX
~ s2~SN~ s(n~ ;C 1)k(~ s;~ xi)
~kiZ
~SN~ s(n~ ;C 1)k(~ s;~ xi)d~ s (7)
Figure 2: Comparing input data for regression in a 100-
player, 3-strategy congestion game. Simplex coordinates
specify a mixture, and color specifies error from true ex-
pected payoffs. Left: payoffs are estimated poorly near the
corners of the simplex. Right: over-sampling edge profiles
reduces this error.We can further simplify equation 7 by re-writing the RBF
kernel from equation 2 in the form of a Gaussian:
k(~ s;~ xi) =cexp
1
2l2(~ s ~ xi)TC2 1(~ s ~ xi)
=c
(2)jSjdet (C2)1
2N(~ sj~ xi;C2)
where we define C2by:
C2 12
6642=l21=l2 1=l2
1=l22=l2 1=l2
............
1=l21=l2 2=l23
775
This gives us:
~kiZ
~SN(~ sjn~ ;C 1)c
(2)jSjdet(C2)1
2N(~ sj~ xi;C2)d~ s
=c((2)jSjdet(C2))1
2Z
~SN~ s(n~ ;C 1)N~ s(~ xi;C2)d~ s
=c((2)jSjdet(C2))1
2Z
~SN(n~ j~ xi;C1+C2)
N(~ sj0;C0)d~ s (8)
=c((2)jSjdet(C2))1
2N(n~ j~ xi;C1+C2)Z
~SN(~ sj0;C0)d~ s (9)
Equation 8 makes use of the product-of-Gaussians iden-
tity from equation 3. This leaves us with the integration of
a Gaussian distribution over the full profile simplex, which
we can approximate by 1, simplifying equation 9 to:
~kic((2)jSjdet(C2))1
2N(n~ j~ xi;C1+C2)
=c(2)jSjdet(C2)
(2)jSjdet(C1+C2)1
2
exp
(n~ ~ xi)T(C1+C2) 1(n~ ~ xi)
=cdet(C2)
det(C1+C2)1
2
exp
(n~ ~ xi)T(C1+C2) 1(n~ ~ xi)
(10)
Equation 10 shows how we can approximate each element
of the vector ~k, and therefore each component of equation 6
computationally. In equation 10, C2depends only on the
length-scale lof the RBF kernel, so det(C2)can be com-
puted in advance and re-used for every expected payoff com-
putation.C1changes with each mixture being evaluated, but
given a mixture, it is the same for all pure strategies and each
~ki. This means that each mixture considered in an equilib-
rium computation algorithm requires one inversion and one
determinant calculation on an nnmatrix.
In principle, this approximation decays near the edges of
the mixed-strategy simplex, because the Gaussian approxi-
mation to the multivariate distribution and the approxima-
tion of the full-simplex integral by 1 should both perform
Figure 3: Comparing accuracy of expected payoff estimates
by GPR and neural networks. Simplex coordinates specify
a mixture, and color specifies error from true expected pay-
offs. Left: GPR. Right: 60-node neural network.
less well. In practice, however, we have found this decay to
be small relative to the inherent inaccuracy of learning from
small data sets. As shown in figure 1b, when the Gaussian
process regression learns accurate payoffs, the integration
methods estimates expected payoffs extremely accurately.
Experiments
In all of our experiments, we generate random large games,
which we represent compactly as action-graph games with
additive function nodes (Jiang, Leyton-Brown, and Bhat
2011). This ensures that we can compare the results of var-
ious approximation methods to a ground truth, checking the
expected payoff of mixed strategies, and the regret of ap-
proximate equilibria. Previous work in approximating large
games has used similar random data sets, but focused on
much smaller games; we believe that our experiments on
games with 100 or more players provide strong evidence that
our methods can scale effectively.
In our first experiment, we isolate the effect of expected
payoff estimation methods by constructing a 100-player, 2-
strategy game, and providing exact payoffs for all 101 pro-
files as inputs to GPR. Figure 1 compares all three generic
methods and the GPR-specific integration method for ex-
pected payoff estimation. Figure 1a shows that sampling
(withk= 100 ) to be quite accurate, but noisy. Point estima-
tion is smooth, but overshoots the movements of the true ex-
pected payoff. Neighbor estimation (with d= 5) falls some-
where in between, exhibiting moderate bias and less noise.
However, note the stepped appearance of the neighbor es-
timates; we found that this occasionally prevented equilib-
rium computation from converging. Figure 1b shows excel-
lent performance for the integration method, which is borne
out through the remainder of our results.
In our second experiment, we compare different choices
of input data in a 100-player, 3-strategy congestion game.
In Figure 2, the simplex coordinates specify a mixed strat-
egy: the corners correspond to all players choosing the same
strategy with probability 1, and the center is the uniform dis-
tribution. Color plotted in the simplex gives the error rela-
tive to true-game expected payoffs: blue indicates low er-
ror, green indicates moderate error. expected payoffs come
from GPR and point estimation in both plots. In the left-hand
plot, input profiles have been spaced evenly throughout the(a) Error vs. approximation method and game size:
DPR IGP PGP NGP0:000:010:020:030:040:050:060:07expected payo error33 players
65 players
129 players
(b) Error vs. approximation method and game type:
DPR IGP PGP NGP0:000:020:040:060:080:100:12expected payo errorCongestion
Local Eect
Polynomial
Sin
Figure 4: Learning methods compared to deviation-
preserving reduction (DPR). IGP GPR + integration, PGP
GPR + point, NGP GPR + neighbor. IGP performs best
across all game sizes and most game types.
profile space ~S. This results in insufficient data for accurate
regression estimates near the extremes of the simplex. Be-
cause Nash equilibria often occur near edges of the simplex,
especially in higher dimensions (more strategies), this can
cause large errors in equilibrium estimates. In the right-hand
plot, profiles have been re-allocated to the edges of the sim-
plex, over-representing profiles with 1 or 0 players choosing
a given strategy. This helps the regression to develop better
estimates of the extreme profiles.
Our third experiment demonstrates that other regression
methods can also be used. Figre 3 again shows errors in ex-
pected payoff estimates plotted on the probability simplex
for a 100-player, 3-strategy congestion game. The left sim-
plex of Figure 3 was created with the same settings as the
right simplex of Figure 2, but a different randomly generated
congestion game. The right simplex shows the accuracy of
neural network learning with point-estimate expected pay-
offs on the same game. The neural network used hidden lay-
ers of 32, 16, 8, and 4 nodes, with sigmoid activation func-
tions, 0.2 dropout probability and the Adam optimizer. The
average error is comparable across the two methods, but the
distribution of mistakes differs significantly. The neural net-
work hyperparameters may not be sufficiently optimized.
Our fourth experiment compares our regression method
against the best existing method for approximating large
symmetric games, deviation-preserving reduction (Wieden-
beck and Wellman 2012). Figure 4 shows average results on
a data set of 120 randomly-generated games. The games in-
clude 10 instances of each combination of parameters from
(33, 65, or 129 players) and (congestion game, local effect
game, action-graph game with polynomial function nodes,
or action-graph game with sinusoidal function nodes). All
methods were given the same amount of data, chosen to
suit DPR. The number of players in the random games were
also chosen to be optimal for DPR. The graph shows av-
erage error in estimating the expected payoff u(s;~ ), for
a large set of symmetric mixtures, including a grid spaced
evenly across the space of possible mixtures and a num-
ber of randomly generated mixtures. Despite having many
parameters chosen advantageously, DPR was outperformed
by the regression methods on all game sizes and nearly
all game types. Among the regression methods, estimating
expected payoffs by the continuous integral approximation
was clearly superior. Neighbor estimation does not consis-
tently out-perform point estimation, which suggests that it is
probably not worth the extra computational burden of query-
ing many more points. We also computed equilibria in these
games, and while DPR closes the gap slightly in terms of
measured true-game regret, GPR with integration remains
the clear winner.
Our fifth experiment compares GPR with integration to
DPR as a function of the number of profiles used as input,
with noisy observations. For this experiment, we constructed
a data set of 15 action-graph games with polynomial func-
tion nodes, and used rejection sampling to ensure that all
games in the data set had only mixed symmetric equilibria.
In the preceding experiments, all methods received correct
payoff values for profiles in the data set. Here each data point
has had normally-distributed noise added. In the presence of
noise, it is common, before performing player reduction, to
simulate the same profile multiple times for a better esti-
mate of its payoffs. Our experiment shows that this can be a
good use of simulation resources, as increasing the number
of samples per profile reduces error and regret as a function
of the total number of simulations over the range of 1–20
samples per profile. This effect tapers off eventually, and by
100 samples per profile, it would be better to sample more
profiles fewer times.
Because regression is inherently robust to noisy inputs,
our method has less need to resample the same profile re-
peatedly, and can sample a larger variety of profiles at the
same simulation cost. As shown in Figure 5a, our method
significantly outperforms all variants of DPR in terms of av-
erage error of expected payoff estimates. Our most impor-
tant experimental result is shown in Figure 5b. This graph
demonstrates that -Nash equilibria computed by replicator
dynamics have significantly lower true-game regret under
our method than under DPR.
Conclusions
We have demonstrated a new method for computing ap-
proximate Nash equilibria in games with a large number(a) expected payoff estimation error vs. data set size:
0 5000 10000 15000 20000 25000
proles0:000:020:040:060:080:100:12expected payo error
dpr 1 sample
dpr 2 samples
dpr 20 samples
dpr 100 samples
gp (integration)
(b) Regret of computed equilibria vs. data set size:
0 5000 10000 15000 20000 25000
proles0:000:010:020:030:040:050:060:070:08regret
dpr 1 sample
dpr 2 samples
dpr 20 samples
dpr 100 samples
gp (integration)
Figure 5: Learning compared to deviation-preserving reduc-
tion (DPR) with noisy payoff samples. IGP performs better
in terms of error and regret whether DPR samples profiles
repeatedly, or constructs a larger reduced game.
of players. Our method uses player symmetries as the ba-
sis for regression-learning of pure-strategy payoffs. Using
these regression estimates, we have shown that expected
payoffs of mixed strategies can be estimated effectively, al-
lowing low-regret symmetric mixtures to be identified. We
provided strong experimental evidence that our methods out-
perform previous techniques for working with large normal-
form games.
Future work on this topic should include extending the
continuous regression approximation to games with multiple
roles and/or to other regression methods. We would also like
to combine our methods for generalizing over players with
existing methods for generalizing over strategies (V orobey-
chik, Wellman, and Singh 2007). The ability to use data
about arbitrary profiles as input to regression opens the prob-
lem of choosing an appropriate set of profiles to simulate; it
may be possible to interleave equilibrium computation and
sample collection in useful ways. Finally, we hope to inves-
tigate the possibility of learning expected payoffs directly;
if feasible, this could dramatically improve computational
efficiency, and/or approximation performance.
References
Bowling, M.; Burch, N.; Johanson, M.; and Tammelin, O.
2015. Heads-up limit hold’em poker is solved. Science
347(6218):145–149.
Cheng, S.-F.; Reeves, D. M.; V orobeychik, Y .; and Wellman,
M. P. 2004. Notes on equilibria in symmetric games. In Sixth
International Workshop on Game-Theoretic and Decision-
Theoretic Agents .
Dandekar, P.; Goel, A.; Wellman, M. P.; and Wiedenbeck, B.
2015. Strategic formation of credit networks. ACM Trans-
actions on Internet Technology 15(1):3:1–41.
Duong, Q.; V orobeychik, Y .; Singh, S.; and Wellman, M. P.
2009. Learning graphical game models. In 21st Interna-
tional Joint Conference on Artificial Intelligence , 116–121.
Gintis, H. 2009. Game Theory Evolving: A problem-
centered introduction to modeling strategic behavior .
Princeton university press, second edition.
Heinrich, J.; Lanctot, M.; and Silver, D. 2015. Fictitious
self-play in extensive-form games. In 32nd International
Conference on Machine Learning , 805–813.
Honorio, J., and Ortiz, L. 2015. Learning the structure and
parameters of large-population graphical games from behav-
ioral data. Journal of Machine Learning Research 16:1157–
1210.
Jiang, A. X.; Leyton-Brown, K.; and Bhat, N. A. R. 2011.
Action-graph games. Games and Economic Behavior
71(1):141–173.
Kearns, M., and Mansour, Y . 2002. Efficient Nash com-
putation in large population games with bounded influence.
In18th Conference on Uncertainty in Artificial Intelligence ,
259–266.
Kearns, M. 2007. Graphical games. In Vazirani, V .;
Nisan, N.; Roughgarden, T.; and Tardos, E., eds., Algorith-
mic Game Theory . Cambridge University Press. chapter 7,
159–180.
Monderer, D., and Shapley, L. S. 1996. Potential games.
Games and Economic Behavior 14(1):124–143.
Morav ˇc´ık, M.; Schmid, M.; Burch, N.; Lis `y, V .; Morrill, D.;
Bard, N.; Davis, T.; Waugh, K.; Johanson, M.; and Bowling,
M. 2017. Deepstack: Expert-level artificial intelligence in
heads-up no-limit poker. Science 356(6337):508–513.
Nash, J. 1951. Non-cooperative games. Annals of Mathe-
matics 286–295.
Papadimitriou, C. H., and Roughgarden, T. 2005. Comput-
ing equilibria in multi-player games. In 16th Annual ACM-
SIAM Symposium on Discrete Algorithms , 82–91.
Petersen, K. B., and Pedersen, M. S. 2008. The matrix cook-
book. Technical report, Technical University of Denmark.
Phelps, S.; Marcinkiewicz, M.; and Parsons, S. 2006. A
novel method for automatic strategy acquisition in n-player
non-zero-sum games. In 5th International Joint Conference
on Autonomous Agents and Multiagent Systems , 705–712.
Porter, R.; Nudelman, E.; and Shoham, Y . 2008. Simple
search methods for finding a Nash equilibrium. Games and
Economic Behavior 63(2):642–662.Rasmussen, C. E., and Williams, C. K. 2006. Gaussian
Processes for Machine Learning , volume 1. MIT Press.
Severini, T. 2005. Elements of Distribution Theory . Cam-
bridge Series in Statistica. Cambridge University Press.
Silver, D.; Huang, A.; Maddison, C. J.; Guez, A.; Sifre, L.;
Van Den Driessche, G.; Schrittwieser, J.; Antonoglou, I.;
Panneershelvam, V .; Lanctot, M.; et al. 2016. Mastering
the game of go with deep neural networks and tree search.
Nature 529(7587):484–489.
V orobeychik, Y .; Wellman, M. P.; and Singh, S. 2007. Learn-
ing payoff functions in infinite games. Machine Learning
67(1-2):145–168.
Wah, E., and Wellman, M. P. 2016. Latency arbitrage in
fragmented markets: A strategic agent-based analysis. Al-
gorithmic Finance 5:69–93.
Wah, E.; Hurd, D.; and Wellman, M. P. 2015. Strategic
market choice: Frequent call markets vs. continuous double
auctions for fast and slow traders. In Third EAI Conference
on Auctions, Market Mechanisms, and Their Applications .
Walsh, W. E.; Das, R.; Tesauro, G.; and Kephart, J. O.
2002. Analyzing complex strategic interactions in multi-
agent systems. In AAAI-02 Workshop on Game-Theoretic
and Decision-Theoretic Agents .
Wellman, M. P.; Reeves, D. M.; Lochner, K. M.; Cheng, S.-
F.; and Suri, R. 2005. Approximate strategic reasoning
through hierarchical reduction of large symmetric games.
In20th National Conference on Artificial Intelligence , 502–
508.
Wellman, M. P.; Kim, T. H.; and Duong, Q. 2013. Analyzing
incentives for protocol compliance in complex domains: A
case study of introduction-based routing. In 12th Workshop
on the Economics of Information Security .
Wellman, M. P. 2006. Methods for empirical game-theoretic
analysis (extended abstract). In 21st National Conference on
Artificial Intelligence , 1152–1155.
Wiedenbeck, B., and Wellman, M. P. 2012. Scal-
ing simulation-based game analysis through deviation-
preserving reduction. In 11th International Conference on
Autonomous Agents and Multiagent Systems , 931–938. |
f4c5c2b7-ff01-464f-91c2-e141d6a19d01 | trentmkelly/LessWrong-43k | LessWrong | Inverse Scaling Prize: Second Round Winners
At the end of the second and final round of the Inverse Scaling Prize, we’re awarding 7 more Third Prizes. The Prize aimed to identify important tasks on which language models (LMs) perform worse the larger they are (“inverse scaling”). Inverse scaling may reveal cases where LM training actively encourages behaviors that are misaligned with human preferences. The contest started on June 27th and concluded on October 27th, 2022 – thanks to everyone who participated! Across the two rounds, we had over 80 unique submissions and gave out a total of 11 Third Prizes.
We are also accepting updates to two previous prize-winners (quote-repetition and redefine-math). For more details on the first round winners, see the Round 1 Announcement Post.
We didn't find the kind of robust, major long-term-relevant problems that would have warranted a grand prize, but these submissions represent interesting tests of practically important issues and that help contribute to our scientific understanding of language models.
Edit: Data for all winning tasks is now available here.
Prize winners
For each submission, we give a description provided by the task authors (lightly edited for clarity), an example from the dataset, and a plot showing inverse scaling on the task. We also include a short discussion of why we found the task exciting and worthy of winning a prize as a TL;DR.
Modus Tollens, by Sicong Huang and Daniel Wurgaft (Third Prize)
TL;DR This task shows strong inverse scaling on almost all models and represents a simple logical reasoning task (modus tollens) that might be expected to show regular scaling. Inverse scaling trends hold across both pretrained LMs and LMs finetuned with human feedback via RL from Human Feedback (RLHF) and Feedback Made Easy (FeedME).
Example (classification)
Consider the following statements:
1. If John has a pet, then John has a dog.
2. John doesn't have a dog.
Conclusion: Therefore, John doesn't have a pet.
Question: Is the conclusion cor |
ef52027a-ef6b-42f2-ac13-ce24619f59a2 | StampyAI/alignment-research-dataset/youtube | Youtube Transcripts | Economic Growth in the Long Run: Artificial Intelligence Explosion or an Empty Planet?
so welcome to our webinar on the
governance and economics of ai
i'm glad that so many of you from all
corners
of the earth are joining us today i'm
anton kornick i'm an economist
at the university of virginia and the
topic of our webinar today
is economic growth in the long run will
there be an
artificial intelligence explosion or an
empty planet
and we have two eminent speakers ben
jones and judge jones
as well as two distinguished discussants
rachel nye and phil drama i will
introduce each one of them
more fully when they are taking the
stage we are excited
to have this discussion today because
the field of economic growth theory
has gone through a really interesting
resurgence in recent years
so at the risk of oversimplifying
a lot of growth theory in the past has
focused on describing
or explaining the steady-state growth
experience
that much of the advanced world has
experienced in the post-war period
that was captured in what economists
call the calder facts
but in recent years a chorus of
technologists
especially in the field of ai have
emphasized
that there is no natural law that growth
in the future
has to continue on the same trajectory
as it has in the past
and they have spoken of the possibility
of an
artificial intelligence explosion or
even a singularity
in economic growth our two speakers
ben jones and chad jones have been at
the forefront of this literature
in a paper that is published in an nvr
volume on the economics of ai
and ben will tell us a bit about this
today
and since an explosion in economic
growth is by no means guaranteed
chad will then remind us that the range
of possible outcomes for economic growth
is indeed vast and we cannot rule out
that growth may in fact go in the other
direction
our webinar today is co-organized
by the center for the governance of ai
at oxford's future of humanity institute
and by the university of virginia's
human and machine intelligence group
both of which i'm glad to be a member of
it is also sponsored
by uva's darden school of business
and uh before i yield to our speakers
let me thank everyone
who has worked hard to put this event
together
alan daffo marcus anderjung and leroux
at the center for the governance of ai
and paul humphries
at the uv machine human machine
intelligence group
as well as ask me yousef at darden
so let me now introduce ben
jones uh more formally
ben is the gordon enduro grand family
professor of entrepreneurship
at the kellogg school of management at
northwestern
he studies the sources of economic
growth in advanced economies
with an emphasis on innovation
entrepreneurship
and scientific progress he also studies
global economic development including
the roles of education
climate and national leadership in
explaining the wealth and poverty
of nations and his research has appeared
in many of the top journals of the field
and has been profiled
in media outlets like the wall street
journal the economist
and the new yorker then the
virtual floor is yours
okay thank you very much uh anton uh for
that introduction and let me share my
screen here
uh it's great to be with you uh to talk
about these these issues and thanks
again to anton and the organizers for
putting this together and for inviting
me to participate
um so the first paper the one i'm gonna
talk about is actually joint with chad
your second speaker
uh he's gonna appear in both uh and this
is also with philippe aguion
and the idea in this paper uh was rather
than sort of a typical
economics paper where you go super deep
into one model and
you know do all the details this was
really to kind of step back
and look at the kind of breadth of
growth models that we have
and then say you know well how would you
insert artificial intelligence into this
search more standard understandings of
growth and where would that lead us
and so we actually have a series of sort
of toy models here
uh and they can of course you know
exploring you know the variety of
directions this can lead us
and seeing what you kind of have to
believe in order for those various
outcomes to
to occur so that's kind of the the idea
behind this paper
um and you know i'm to do this in an
almost extremely non-mathematical way
not a completely math free way but i
know that this is a
this is a seminar with a a group of
people with diverse
uh disciplinary backgrounds so i don't
want to presume people are steeped in
endogenous growth models
uh so i'm gonna try to really emphasize
the intuition as i go uh to the best
that i can
i will have to show a little bit of math
a couple times but not too much okay so
so the idea in this paper is okay how
would we think about ai and you might
think that ai
you know helps us make you know goods
and services and you know things things
that go into gdp
and that we consume uh it also
you know might help us be more creative
okay so we're gonna distinguish between
ai
entering kind of the ordinary production
function for goods and services in the
economy
but also that it might go into the
so-called knowledge production function
into r
d and it might help us you know succeed
better in in
revealing new insights and and uh and
breakthroughs
uh uh about uh the world and the economy
and then the kind of implications we
want to look at just two very high level
implications you know what what will
happen to long run growth under various
assumptions what do you have to believe
to get different
outcomes in terms of the rate at which
standards of living are improving
but also inequality what share like the
gdp might go up gdp per capita
might go up but what share of that is
going to go to
labor particular workers is obviously a
lot of fear that ai would displace
workers
uh and that may be more and more of the
fruits of of income will go to the
owners of the capital say the owners of
the ai
and then of course there's this other
idea almost more from science fiction it
seems but you know taken seriously by
some in the
and certainly the computer science
community that we might actually
experience
radical accelerations in growth even to
the point of some
singularity anton reference to growth
has been very steady according to sort
of
you know since the industrial revolution
uh and
and but maybe we're going to see an
actual structural break where things
will really
really take off and of course i think
chad's paper hill show later is
maybe going the other way uh potentially
we'll explore that as well
um all together okay so how are we gonna
think about ai and
you might think ai is this radically new
thing and in some ways it is but
you know one way to think about it is
that we are we are
you know furthering automation right
what are we doing we're taking a task
that is performed by labor
maybe reading a radiology you know
result in a medical setting
and then we're going to have a machine
or algorithm do that for us right
um and we're going to be you know image
search on google i used to have to
categorize which is a
cat and now google can just have an ai
that tells us which which image is a cat
um but if you think about it in
automationdirect it can be very useful
because then we can sort of think about
ai
in more standard terms that we're used
to to some extent
in economics so if you think in the past
you know the industrial revolution was
largely about
replacing labor with certain kinds of
capital equipment
maybe that was textile looms and steam
engines for power
you know ai is sort of a continuation of
that process in this view and it's
things like you know driverless cars and
and pathology and other another
applications so that's one main theme in
the work and how we want to introduce ai
into our thinking of growth and see
where it takes us
the the second main theme that really
comes out in this paper we developed
writing it
is that we want to be very careful to
think about not just what we get good at
but what we're not so good at and the
idea that growth might be determined
more by bottlenecks the growth may be
you know constrained not by what we get
really good at but what is actually
really important what is essential
and yet it's hard to improve and i'll
make a lot of sense of that as we go
intuitively but just
so what are we saying so i have a
picture that these guys are sugar beet
farmers you know pulling sugar
beets out of the ground by hand
harvesting them and that was done this
is a
kind of combine harvester harvester type
machine that's you know automating that
and pulling sugar beets out of the
ground
uh with a machine so that's kind of like
20th century automation
and then a lower graph lower picture i'm
trying to think about you know
automation as
as ai uh and on the left if you've seen
the movie hidden figures
these are the computers i always think
it's very interesting so computer was
actually a job description
these women were computers at nasa
involved in space flight
and they were actually doing
computational calculations by hand
and then on the right i have one of
nasa's supercomputers who have basically
replaced
that job description entirely so we see
a lot of laborers being replaced by
capital
raising productivity but also displacing
workers and so how do we think about
those forces
okay so one way to think about this
model is to start with azir's model
which is the following so imagine
there's just n different things we do
in the economy and different tasks and
each task
represents the same constant share of
gdp of total output
and to an economist here that would
sound let's call it douglas right so we
that's a
douglas model but if you're not
economist ignore that we just imagine
the tasks
you know every task is an equivalent
share of gdp for simplicity
and when we think about automation what
we're saying is that a task you know was
done by
labor but now it might be done by
capital equipment instead ai would be a
computer an
algorithm a combine harvester would be a
piece of farming equipment
automation and so if you think that a
fraction of the tasks beta
so there's you know of all the tasks one
for beta percent are automated
then the capital share of total gdp is
beta so that means labor gets one minus
beta
and then the expenditure on the capital
equipment is a beta share of gdp
okay so that would seem that's a very
simple model
it's very elegant in a way and it would
say that if we keep automating if you
increase
beta we keep taking tax that were done
by labor and replacing them with
machines
or ai what will happen well the capital
share of income will increase and the
labor share of income will decrease so
that sounds like it will be
you know inequality in a create
inequality in the sense that labor
will get less less income that might
sound very natural maybe think it's
what's happening today we have seen the
capital share it seems to be going up
right in a lot of countries and advanced
economies and like in the u.s
um it seems like there's a lot of
automation going on from robots to these
new ai type things
those are of course just two trends
though so maybe they just happen to be
correlated we think that the ai
is causing causing that rise in the
capital chair well this would be a model
in which that could be true
the problem with that model though is if
you look backwards we've seen tons of
automation like sugar beads or so many
other things robots and auto
manufacturing automobile manufacturing
uh
that you know we didn't see the capital
chair go up in the 20th century it was
very very steady
so that suggests that if you know this
wouldn't really fit kind of our
normal understanding of automation and
so it's not clear that this seems like
quite the right model
so how can we repair it a simple way to
repair it uh it was what really did one
idea that we developed in the paper
is to introduce so-called baumel's cost
disease which is that the better you get
at a task the less you spend on it
so as you automate more tasks maybe the
capital share wants to go up
but something else also happens right so
if i automate a task like collecting
sugar beats
what can i do i can start throwing a lot
more capital at that task i can keep
getting more and more machines and doing
sugar beats
right and moreover the capital i put at
the task the capital equipment
might get better and better i first used
a pretty rudimentary type of capital and
eventually
these very fancy fancy machines or i
introduced computers but then computers
get faster
if you throw more capital or better
capital at it what's going to happen
well you're going to get more productive
at getting sugar beats or doing
computation at nasa
and so the cost of doing the task is
going to drop but if the cost drops and
things are kind of competitive in the
market
the price should drop right the price
will also drop
so what's going on you can do more of it
a greater quantity
but the price of the task you're
performing will fall
so what's the share in gdp well the
quantity is going up but the price is
falling
if the price is falling fast enough then
actually the share in gdp
will go down even though you do more of
it so you get more sugar beats but the
price sugar beats plummets
and so sugar beats as a share of gdp is
actually declining
and then what happens is that the
non-automated sort of bottleneck tasks
the ones you're not very good at
actually come to dominate even more and
more okay so if you think backwards in
you know 20th century history are
done by the industrial revolution we see
that agriculture
and manufacturing have had rapid
productivity growth and lots of apparent
automation like sugar beets
yet agriculture and manufacturing
agriculture surely having
dwindling and dwindling share of gdp and
manufacturing gdp shares also seem to be
going down so it's like what you get
good at and what you automate actually
interestingly
becomes less important with time it
starts to disappear in the overall
economy
we're left with things like health
services education services government
services
this is bumble's cost disease point
things that we find hard to improve the
hard stuff
actually comes to take on a larger and
larger share of gdp and if we can't
improve that
then our chances for growth because it
matters so much to gdp
dwindle okay so in this view the ira
capital share is the balance between
automating more tasks which tends to
make the capital share go up
but the expenditure share in each task
declines and that tends to make it go
down
so what a model we offer in this paper
is what we can call
more of the same maybe that's what ai is
maybe ai is just more balanced growth
and we keep automating more and more
tasks but then they keep becoming a
dwindling share of the economy
and we never automate everything and you
can actually show as we do in the paper
a model
where even though i think this is all
the set of tasks and a greater and
greater share being done by
capital equipment and artificial
intelligence and a tinier and tinier
share being done by labor
even though labor gets stuck in a very
very small share it still actually gets
say two-thirds of gdp it still gets the
same historical number
and again why is that is because this
stuff all the labor
all the capital stuff is doing more
tasks but it's prices plummeting because
we're so good at it
and you're left with just a small set of
tasks being done by labor but paying
enormously for them
okay and that may be what's going on in
the economy maybe that's what's been
going on in the 20th century certainly
consistent with what's been going on in
20th century to you know first order
without overstating that case
but it's it's broadly consistent with
the stylized facts of growth
but that would suggest ai is again just
more of the same we just keep automating
it so here's a simulation from our paper
this is steady state growth you look on
the x-axis we're looking over you know
five centuries
you get steady state growth even as
what's happening with automation here's
the green line
the green line you're ultimately
automating almost everything you're just
sort of slowly automating everything
you never quite get to the end and you
just get constant growth and you get you
can get a constant capital share
okay not a we're not a rising capital
ship and so actually this is an idea
that i've been
developing a new paper which is almost
done that's sort of seeing how we how
much we how far we can go along this
this line um okay but let's go a
different pack because a lot of people
who observe
artificial intelligence are excited by
the possibility that maybe it will
accelerate growth
and you know many futurists make these
claims that we could even get some
massive accel
acceleration something like a
singularity so we explore that in this
paper as well
and so what would you have to believe
for this to happen um
so we considered we kind of consider two
different typologies of a growth
explosion
what we call a type one growth explosion
where the growth rates we're going to
depart from this
steady state 20th century 20 early 21st
century experience
and we're going to see a slow
acceleration growth maybe to very very
high levels
we'll call that a type 1 growth
explosion uh and the other would be a
type 2 where we mean a literal
in a mathematical sense singularity
where you go to infinity in productivity
and income
uh in some finite point in time in the
future you actually literally have a
singularity
where you go to you go to infinity so
and you can actually surprisingly
using sort of standard growth reasoning
and automation you can get either of
those outcomes all right so the first
one a simple example and they're more
but one example of the first one
is when you do achieve complete
automation so not just you kind of keep
automating at a rate and never quite
finished
now we're going to fully automate here's
my first equation y
gdp k capital
that's the automation capital that's all
the combat harvesters and the
super computers and the ai and then a is
the kind of the quality of the capital
the productivity of one unit of capital
all right if you so this is fully
automating in other words there's no
labor there there's no l
the labor is now irrelevant to
production of gdp we can do the whole
thing just with machines that's what
that's saying
it just depends on k and the quality of
decay which we call
a if you look at that what's the growth
rate in y it's going to be the growth
rate and y is the growth rate in a so
for gdp the growth in gdp is the growth
rate in a
the technology level plus the growth
rating capital now but the thing about
capital which is really interesting and
different from labor which i think is
where chad's going to be going in his
paper
is that you know with with capital you
can keep making more and more of it
right because you make how do you make
capital well you invest in it you build
it
and that comes out of gdp so think about
this equation if i push up capital
i get more output and then without more
output i can invest more
okay and then more importantly if i push
up the level of technology
i get more and more for every unit of
capital that increases gdp i can invest
more and keep building more capital
okay so the growth rate actually turns
out to be what's below i'm ignoring
depreciation
but basically you can see that as long
as you keep pushing up a
if you can keep pushing up the level of
technology so you keep improving the ai
you keep improving computers uh the
growth rate's going to track with a and
it's going to keep going up and up and
up and up and this is a type 1 growth
explosion so called y it's an ak model
it's a standard model in endogenous and
early standard model and endogenous
growth theory
if we can automate everything that
suggests in fact
that we could have a very sharp effect
on the growth rate that's a very
uh strong view of one view of what ai
might do
interestingly another place to put ai as
i
alluded to in the very beginning is you
could put it into creativity and
innovation itself
and if you do that things can really
take off all right so this is a
knowledge production function a dot
is the rate of change of the level of
technology the quality of the capital
and if i fully automate how we produce
that again there's no labor in this
equation it just depends on capital
and then the state of technology itself
a and that's going to act a lot like the
second equation
which is that the growth in a is going
to depend on the level of a to some
parameter phi
and that's like positive feedback i push
up a growth goes up in a
which causes growth and y to go up i
push up both you know and then a goes up
and then you go like this okay
and that actually will produce if you
solve the differential symbol
differential equation it does produce a
true mathematical singularity it'll be
some point in time t
star which is definable at which we will
achieve infinite productivity all right
now maybe that sounds like a fantasy and
it would be a fantasy
if if so because there may be certain
obstacles to that happening and then
i'll just go very quickly through a
couple one obstacle
is that you just simply can't automate
everything right so both of those models
assume
you can get to a lot of automation right
but maybe automation is very it's
actually very hard
maybe though it's easy to automate sugar
beats but there are just certain
cognitive tasks for example with regard
to ai
that are going to be very very hard to
automate if we never get to full
automation we can still get growth to go
up
but we're never going to get these kind
of singularities in these models
okay in the simplest form so if you
think that there are some
kind of bottleneck tasks that we can't
automate uh then we're gonna get again
we're not we're not gonna get these
labor-free
full automation singularities you have
to believe that you know to some extent
that we can truly automate all these
things and of course that's an open
question with ai
but how far it can go in goods and
services production and sort of creative
innovative activity
a second a second constraint kind of the
latter two constraints on substances
come from the universe itself which is
this differential equation at the top if
that parameter phi is greater than zero
it will give you a singularity you will
get one
okay you fully automate idea production
and you will get one in finite time
but the question then is really whether
we believe that parameter phi is
actually larger than zero
what that what is that saying it's
saying that if it's greater than zero
then when i increase a i increase the
level of technology in the economy i
make future growth faster
right but if phi is less than zero what
happens when i when i raise the level of
existing technology and phi is less than
zero i make future growth
slower it takes away that positive
feedback loop okay
and then you don't get a singularity and
there are good reasons to think that phi
might be less than zero we don't know
but there are reasons to think it
because let's say that like there's only
so many good ideas in the universe and
we came up with calculus and we came up
with the good ones early
and the remaining ones are hard to
discover or just there aren't that many
good ones left and so if you think we're
kind of fishing out the pond
right think of ai as like changing the
fisherman we get better fishermen on the
edge of the pond
but if the pond itself is running out of
fish big fish for us of new ideas
it doesn't matter what your fishermen
are there's nothing left in the pawn to
catch
uh and then there's some other i have an
other version that called the burden of
knowledge but
regardless there are some ideas of
in in the existing economic growth
literature about science and innovation
that suggests find well maybe less than
zero and that's just going to turn off
that
singularity and then the third one which
is somewhat related
is that there just might be bottleneck
tasks and this kind of comes back to
that bomball cost disease reasoning but
more at a task level
so for example you know let's say that
gdp here is actually
you know it's a combination of our
output and all these tasks in the most
simple form let's say it's the minimum
so this is a very real bottleneck
you're only as good as your weakest link
it's one version of a simple version of
abominable cost disease
so if it's the min function it doesn't
matter how good you get at every task
the only thing that matters is how good
you are at your worst test right
so in other words we might be really
really good at agriculture but at the
end of the day we're really bad at
something else and so that's what's
holding us back
um and you know i think actually this is
actually quite instructive because
think about moore's law people get so
excited about moore's law in computing
and a lot of people who believe in
singularities
are staring at the moore's law curve and
it's this incredibly dramatic
exponential
rapid rapid rapid increase in
productivity which is
mind-boggling in a way at the same time
this has been going on for a long time
that moore's law and if you look at
economic growth we don't see an
acceleration
right if anything we probably see it
slow down right and that suggests that
no matter how good you get at computers
there are other things
holding us back like it still takes as
long to get from
you know one point on a map to another
based on the
available transportation technologies
that's not really changing um
so you know as going back to the bomb
theme if things are really depend on
what we're sort of
what is essential but hard to improve we
can actually take our computing
productivity to infinity literally and
it just doesn't matter
it won't it'll help it'll make us richer
it's good but it won't fundamentally
change
our growth prospects unless we can go
after the hard problems that are the
hard ones to solve
so to conclude you know this these are a
whole series of models obviously we do
this at much greater length in the paper
if you'd like to read it
um you can put ai in the production of
goods and services
you know if you can't fully automate but
just kind of slowly automate you kind of
looks like more of the same it's sort of
a natural way to go
but if you can get to full automation
where you don't need labor anymore you
can get a rapid acceleration and growth
through this so-called
what we call a type 1 singularity when
you put a ai in the
product ideas production function in the
creation of new knowledge
you can get even stronger growth effects
and that in fact could even lead to one
of these true mathematical singularities
sort of in science fiction
but there are a bunch of reasons in both
cases to think that
we might be limited because of other
automation limits because their search
limits and that creative process really
with regard to the knowledge production
function
or more generally in either setting with
natural laws like you know i didn't say
it a lot but like the second
law of thermodynamics seems like a big
constraint on energy efficiency
that we're actually pretty close to uh
in current technology and if energy
matters then you know
that's going to be a bottleneck even if
we can get other things to sort of
skyrocket in terms of productivity and
so i think a theme that you know chad
and i certainly came to
writing this paper was the kind of
interesting idea that you know
ultimately growth seems determined
potentially not by what you are good at
but by what is essential yet hard to
improve and and that
that is kind of a important force to
keep in mind when we all get excited
about
where we are advancing quickly and then
we go back to the aggregate numbers and
we don't see much progress
this is just like a pretty useful way
potentially to frame that and begin to
think about it maybe we should be
putting a lot of our effort into
thinking about what we're bad at
improving
and why that is if we really want to
understand future
growth and standards of living um okay
so i i went pretty quickly but hopefully
i would use my time
i didn't spill over too much beyond my
time and uh look forward to the
discussions
from thanks rachel and phil in advance
and look forward to chad's comments as
well
thank you thank you ben the timing was
perfect
and to all our participants let me
invite you
to submit questions through the q a
field at the bottom of the screen
after all the presentations we will
continue the event with the discussions
of the points that you are raising and
incidentally to the speakers
if there are some questions
clarification question for example
where you can type a quick response feel
free
to respond to a q a in the q and a box
directly
so let me now turn over to chad chad
jones is the stanco 25 professor of
economics
at the stanford graduate school of
business he is noted for his research on
long-term
economic growth and in particular he has
examined the fundamental sources of
growth in income
over time and the reason underlying the
enormous differences in standards of
living across countries
in recent years he has used his
expertise
in macroeconomic methods to study the
economic
causes behind the rise in health
spending
and top income inequality and he's the
author of one of the most popular
textbooks
in macroeconomics as well as very well
published
in the top journals in economics chad
the floor is yours
wonderful thanks very much anton it's
really a pleasure to to be here
um i think antoine did a great job of
introducing this session in in pairing
these
these two papers together um as he said
you know a lot of growth theory
historically has looked back and tried
to understand how can constant
exponential growth be possible for 100
years
and the first paper that ben presented
kind of looked at you know
automation artificial intelligence and
possibilities for growth rates to
to rise and even explode this paper is
going to look at the opposite
possibility
and you know ask could there be
uh you know the end of economic growth
and i think
all these all these ideas are worth
exploring and i guess my general
perspective is you know part of the role
of economic theory
is to you know zoom in on particular
forces and study them closely
and then at the end of the day we can
come back and ask well how do these
different forces play against each other
so that's kind of the spirit of this
paper
okay so um a large number of growth
models work this way basically people
produce ideas
and those ideas are the engine of
economic growth so
you know the original papers by paul
romer and egg young and howard and
grossman and helmand work this way
these sort of semi-endogenous growth
models that i've worked on and sam
gordom
and paul sneakerstrom uh basically all
idea driven growth models work this way
people produce ideas and ideas drive
growth
now these models typically assume that
population
is either constant or growing
exponentially
and for historical purposes that seems
like a good assumption
an interesting question to think about
though is what does the future hold
um from this perspective i would say
before i started this paper
my view of the future of global
population which i think of as kind of
the conventional view
is that it was likely to stabilize at 8
or ten billion people
you know 100 years from now or something
um
interestingly there was a paper a book
published last year by bricker and
ibbitson called empty planet
and this book made a point that after
you see it is
you know very compelling and interesting
they claim that maybe the future
is actually not one where world
population stabilizes
maybe the future is one where world
population declines maybe the future is
negative population growth
and the evidence for that is remarkably
strong i would say
in that high income countries already
have fertility rates that are below
replacement
so the total fertility rate is sort of a
measure in the cross section
of how many kids are women having on
average
and you know obviously two is a special
number here if women are having more
than two kids
on average then population is going to
tend to rise if women are having fewer
than two kids
on average then population will decline
and maybe it's 2.1 to take into account
mortality but but you get the idea
the interesting fact you know
highlighted by bricker and evanson and
well known to demographers
is that fertility rates in many many
countries especially advanced countries
are already below replacement so the
fertility rate in the u.s is about 1.8
in high income countries as a whole 1.7
china 1.7 germany 1.6 japan
italy and spain even lowered 1.3 or 1.4
so in many advanced countries fertility
rates are already well below replacement
and then if we look historically you
know again we kind of all know this
graph
qualitatively fertility rates have been
declining so take
india for example in the 1950s and 60s
the total fertility rate in india was
something like six women had six kids on
average
and then it fell to five then to four
and then to three
and the latest numbers in india i think
are 2.5 or 2.4
but the perspective you get from this
kind of graph is
well if we wait another decade or two
even india may have fertility below
replacement rates fertility rates have
been falling all over
the world and
maybe they're going to end up below too
so the question in this paper is
what happens to economic growth if the
future of population growth
is that it's negative rather than zero
or positive
right and the way the paper structures
it considers this possibility from two
perspectives
first let's just feed in exogenous
population growth let's just assume
population growth is negative half a
percent per year forever
feed that into the standard models and
then see what happens
and the really surprising thing that
happens is you get a result that i call
in honor of the book
the the empty planet result and that is
that
not only does the population vanish with
negative population growth
you know the global population is
disappearing
but while that happens living standards
stagnate
so this is you know quite a negative
result living standards stagnate
for a vanishing number of people and it
contrasts with the standard
you know growth model result that you
know all these growth models that i
that i mentioned earlier have which i
i'm now going to call an expanding
cosmos result
but it's it's basically a result that
you you get exponential growth in living
standards so living standards grow
exponentially
at the same time the population grows
exponentially so on the one hand you
have this sort of traditional expanding
cosmos view of the world
and what this paper identifies is hey if
these
patterns in fertility continue we may
have a completely different kind of
result
where instead of living standards
growing for a population that itself is
growing
maybe living standards stagnate for a
population that disappears
okay then the second half of the paper
and i'll only have a chance to allude to
how this works
says well what if you endogenize the
rate of fertility what if you endogenize
the population growth
do you learn anything else and
you can get an equilibrium that features
negative population growth that's good
we can get something that looks like the
world
and the surprising result that comes out
of that model
is that even a social planner if you ask
what's the best you can do in this world
let you know
choose the allocation that maximizes the
utility of
everyone in the economy and you know
with population growth the question of
who is everyone is is an interesting
question
um but the the result here is that a
planner who prefers this expanding
cosmos result
can actually get trapped by the empty
planet outcome
and that's a surprising kind of result
it might seem like it doesn't make any
sense at all but i'll try to highlight
how it can happen
okay i'm going to skip the literature
view in the interest of time
i've already kind of told you how i'm
going to proceed
basically what i want to do is look at
this negative population growth
in the sort of classic roma framework
and then in a semi endogenous growth
framework and then go to the uh
fertility results okay so let me start
off by illustrating
this empty planet result in a set of
traditional models so
make one change in traditional models
you know instead of having
positive population growth or zero
population growth have negative
population growth and see what happens
that's that's the name of the game for
the first half of the paper
okay to to to do that let me just remind
you what the traditional results
are in a really simplified version of
the romer the roamer model
um and you know i'm sure you all know
but this this this
the model this is based on and this
paper by romer won the nobel prize in
economics a couple of years ago so this
is a
a very well respected uh important
modeling in the growth literature
so the insight that got romer the nobel
prize
was the notion that ideas are not rival
ideas uh don't suffer the same kind of
inherent scarcity as good so um if
if there's an apple on the table you can
eat it or i can eat it apples are scarce
bottles of of olive oil are scarce um
you know coal is scarce uh you know a
surgeon's time is scarce
everything in economics that we're
traditionally studying
is a scarce factor of production and
economics is the study of how you
allocate those scarce factors
but ideas are different if we've also
got the fundamental theorem of calculus
one person can use it a million people
can use it a billion people can use it
and you don't run out of the fundamental
theorem of calculus the same way you'd
run out of
apples or computers and so
that means that production is
characterized by increasing returns to
scale there's constant returns to
objects here just people
and increasing returns to objects and
ideas taken together this parameter
sigma
being positive measures the degree of
increasing returns to scale
then where do ideas come from in the
roma model there's a basic assumption
that says that you know each person can
produce
a constant proportional improvement in
productivity so the growth rate of
knowledge
is proportional to the number of people
and then the roma model just assumed
that population
was constant i'll come this is the
assumption i'm going to come back and
relax in just a second
um so if you if you solve this model
income per person lower case y is just
you know gdp divided by the number of
people that's just proportional to the
number of ideas
right the amount of knowledge each
improvement in knowledge raises
everyone's income because the
non-rivalry that's the the deep roamer
point
and then the growth rate of income per
person
depends on the growth rate of knowledge
which is proportional to population
right so this is a model where you can
get constant exponential growth
in living standards with a constant
population
and if you look at this equation you
realize well if there's population
growth in this model
that gives us exploding growth in living
standards
um we don't see exploding growth and
living standards historically
and we do see population growth so
there's some tension there and that's
what the semi-endogenous growth models
are designed to fix that i'll come back
to in a second
okay in the meantime what i want to do
is change this assumption that
population is constant
and replace it with an assumption that
the population itself is declining at a
constant exponential rate
so let ada denote this rate of
population decline
so think of eight as one percent per
year half a percent per year the
population's falling
at half a percent per year and then what
happens in this model
well if you combine the second and third
equations you get this this law of
motion for knowledge and
this differential equation is easy to
integrate right it says the growth rate
of knowledge
is itself falling at a constant
exponential rate
and not surprisingly if the growth rate
is falling exponentially
then the level is bounded that's what
happens when you integrate this
differential equation
you get the result that the stock of
knowledge
converges to some finite upper bound a
star
and since knowledge converges to some
finite upper bound
income per person does as well and you
can calculate these as functions of the
parameter values and it's interesting to
do that and i do a little bit of that
in the paper but let me let me leave it
for now by just saying
what we did is just by by changing this
assumption that population was constant
making it you know population growth
negative
you get this empty planet result you get
that living standards
asymptote they stagnate at some value y
star
as the population vanishes that's the
empty planet result
now we're going to look at this other
class of models the simian dodgers
growth class of models
and what was interesting about these
models is that
in the original framework the
the sort of roamer style models in the
simian dodgers growth models
led to very different results in the
presence of positive population growth
these models yield
very different outcomes and what's kind
of interesting is that with negative
population growth they yield very
similar outcomes
okay so again let me go through the same
kind of order as before
let me present the traditional result
with positive population growth
and then change that assumption and show
you what happens when population growth
is negative
so same goods production function we're
taking advantage of roamer's non-rivalry
here
and i'm making basically one change if
you want set lambda equal to one that
doesn't really matter
i'm introducing what ben described in
in the earlier paper as as you know this
sort of ideas are getting harder to find
force
right the fishing out force right and
beta kind of measures the rate at which
ideas are getting harder to find it says
the growth rate of knowledge is
proportional to the population
but the more ideas you discover the
harder it is
to to find the next one right and beta
measures the degree to which it's
getting harder
so beta to think of beta some positive
number
and then let's put in population growth
at some positive rate
n that's exogenous okay same equation
income per person is proportional to the
stock of ideas raised to some power
the stock of ideas is itself
proportional to the number of
people and that's an interesting finding
here which is
the more people you have the more ideas
you produce
and the more total stock of knowledge
you have and therefore the richer the
economy is
right people correspond to the economy
being rich in the long run by having
lots of ideas
not to the economy growing rapidly
that's what what happens
uh versus the earlier models and then if
you take this equation
and you know take logs and derivatives
of it it says that the growth rate of
income per person
depends on the growth rate of knowledge
which in turn depends on the growth rate
of people
the growth rate of income per person is
proportional to the rate of population
growth
where the factor of proportionality is
the degree of increasing returns to
scale in the economy essentially
okay and so this model you can have
positive population growth
being consistent with constant
exponential growth in living standards
so this is the expanding cosmos result
right we get exponential growth in
living standards
for a population that itself grows
exponentially maybe it fills the earth
maybe it fills the solar system maybe it
fills the cosmos right that's the kind
of take it to the
implausible extreme maybe a result of
this model
okay now let's do the same thing suppose
we change that assumption that
population growth is positive
to one of population growth being
negative again that kind of
remarkably i would say looks like the
future of the of the world that we live
in
right uh based on the evidence that i
presented earlier
okay so once again we've got this
differential equation you substitute
from the the
negative population growth equation in
and you see that
you know not only does the growth rate
of knowledge decline exponentially
because of this term but it falls even
faster so the growth rate of knowledge
falls even faster than exponentially
so of course the stock of knowledge is
still going to be bounded this is
another differential equation that's
really easy to integrate
and you get that once again the stock of
knowledge
is bounded right and you can play around
with the parameter values and
do some calculations um in the interest
of time
uh let me not do that let me instead say
what you know what we see let me just
sort of summarize
um is so first as a historical statement
fertility's been trending downward we
went from five kids to four kids to
three kids to two kids and now even less
in rich countries and an interesting
thing about that
is from the microeconomic perspective
from the perspective of the individual
family
there's nothing at all special about
having more than two kids or fewer than
two kids it's an individual family's
decision and some families decide on
three some families excited two
one zero whatever but there's nothing
magic
about above two versus below two from an
individual family's perspective
but the macroeconomics of the problem
makes this distinction absolutely
critical
because obviously if on average women
choose to have slightly more than two
kids
we get positive population growth
whereas if women decide to have slightly
fewer than two kids
we get negative population growth and
what i've shown you on the previous
you know four or five slides is that
that
difference makes all the difference in
the world to how we think about
growth and living standards in the
future if there's negative population
growth
that could condemn us to this empty
planet result
where living standards stagnate as the
population disappears
instead of this world we thought we
lived in where living standards were
going to keep growing exponentially
along with the population and so this
this relatively small difference
matters enormously when you project
growth forward
and the the sort of fascinating thing
about it is it seems like as an
empirical matter
we're much closer to the below two view
of the world
than we are to the above to view the
world so maybe this empty planet result
is something we should take seriously
that's i would say that
the most important finding in the paper
okay let me go to the second half of the
paper just very briefly and i won't go
through the model in detail it's
admittedly subtle and complicated and
took me a long time to understand fully
but i do want to give you the intuition
for what's going on so i write down a
model
where people choose how many kids to
have
right and in the equilibrium of this
model
um the idea part of kids is an
externality so we have kids because we
love them
and in my simple model people ignore the
fact
that their kids might you know be the
next you know
einstein and marie curry or jennifer
dude i guess
um you know with the nobel prize for
christopher
um and that they might create ideas that
benefit everyone in the world
right the individual families ignore the
fact that their kids might be isaac
newton
and so the planner is going to recognize
that you know social welfare recognizes
that having kids creates ideas and so
the planner wants you to have more kids
than you and i want to have there's an
externality in the simple model along
those lines
and admittedly this is a modeling choice
you can you know people are writing down
these kind of fertility models for a
while and
there are lots of other forces and you
can get different results i don't want
to claim this is a general result
rather i see it as illustrating an
important possibility
okay um okay
so so how do as i mentioned the key
insight that you get out of studying
this this endogenous fertility model
is that the social planner can get
trapped in the empty planet even a
social planner who wants this expanding
cosmos
if they're not careful and i'll try to
say what i mean by if they're not
careful
they can get trapped in the empty planet
so how do i understand that
so in this model population growth
depends on a state variable x which you
can think of as
knowledge per person it's a to some
power divided by n to some power let me
just call it knowledge per person
and we can parameterize the model so
that the equilibrium
women have fewer than two kids and so
population growth is negative
and if population growth is negative
look at what happens to x
i've already told you that a converges
to some constant
and n is declining and so x is going off
to infinity
so in the equilibrium x is rising
forever
what about in in the optimal allocation
the allocation that maximizes some
social welfare function
well the planner's going to want us to
have kids not only because we love them
but because they produce ideas that
raise everyone's income
the key subtlety here is suppose we
start out in the equilibrium allocation
where x is rising and population growth
is negative and ask
when do we adopt the good policies that
raise fertility the planner wants you to
have more kids
do we do we adopt the the policies that
raise fertility immediately
do we wait a decade do we wait 50 years
do we wait 100 years
that's the if you're not sufficiently
careful the point is
if society waits too long to switch to
the optimal rate of fertility
well then x is going to keep rising
and the idea value of kids gets small
as x rises because remember x is
knowledge per person
as x rises we have tons of knowledge for
every person in the economy
so the marginal benefit of another piece
of knowledge is getting smaller and
smaller
so the idea value of kids is getting
smaller and smaller
and because we've already said the
loving your kids force still leads to
negative population growth
well even if you add a positive you know
idea value of kids
the planner might still want negative
population growth if you wait too long
if you wait for the idea value of kids
to shrink sufficiently low
then even the planner who ex-ante
preferred the expanding cosmos
gets trapped by the empty planet so what
this says is that
it's not enough to worry about fertility
policy we have to worry about it sooner
rather than later
and um here's just a diagram i think i'm
almost out of time let me just conclude
so what i take away from this paper is
that fertility considerations
are likely to be much more important
than we thought this distinction between
slightly above two and slightly below
two that from an individual family
standpoint
you know just barely seems to matter
from an aggregate standpoint from a
macroeconomic standpoint
is is a big deal it's the difference
between the expanding cosmos
and the empty planet um as i mentioned
when i started this is not a prediction
it's a study of one force but i think
it's it's much more likely than i would
have thought
you know before before i started this
project and
uh there are other possibilities of
course we've we've talked about one with
ai producing ideas so that people aren't
necessary
important in my production function is
that people are a necessary input
you don't get ideas without having
people and maybe id
maybe ai can change that that's
something we should discuss in the
in the open period um there are other
forces technology may affect fertility
and mortality
maybe we end up reducing the mortality
rate to zero so that even having one
kid per person is enough to keep
population growing for example
um maybe evolutionary forces favor
groups that
you know have high fertility for for for
some reason maybe it selects for those
genes and so maybe
this this you know below replacement
world we look like we're living in maybe
that's not going to happen in the long
run
but anyway i think i'm out of time let
me go ahead and stop there
thank you very much chad and um
let me remind everybody of the q and a
again
our first discussion of these ideas is
rachel
nay rachel is a professor of economics
at the london school of economics and
the research associate at the center for
economic performance
as well as a research affiliate at the
center for economic policy research
her interests include macroeconomic
topics
such as growth and development
structural transformation
as well as labor markets and housing
markets
rachel the floor is yours thank you
anton
um thank you very much for having me
discuss this
two very interesting paper there's a lot
of interesting contents in both paper
but because of time what i will focus on
is the aspect related to the future of
economic growth
and and the role played by artificial
intelligence
declining population growth now when we
talk about artificial intelligence there
are many aspects of it
there were political as a bad
philosophical aspect which
i will not have time to talk about so
today i will purely focus on
their implication for future of economic
growth
okay so economic growth is about the
improvement in the living standard
when we think about the fundamental
source of growth as both ben
and chad point out is about technology
progress
technology progress can happen through
ind
or experience which is like we will be
doing something and we get better
than in doing something but the key
thing for technological progress
is that it require brain input so far
for the last 2000 years or so the main
brain input is the human brain
so here are some examples we already
heard mentioning that how their research
outputs
have improved the living standard for
for mankind over the last
two thousand years now chuck's paper
is very interesting and it is bringing
on some something that is really
important
so here is the figure that basically
repeat what chad has shown us
he's from the united nations about the
total
the life worker woman how many child
women have as you've seen in high income
country
has already fallen below the replacement
ratio which is about two
for the world as a whole is also falling
and in fact united nations predict that
but
in 80 years the population growth will
be stagnant
so there will be zero population growth
and that means going forward will
receive negative population growth
what chad has convincingly showed is
that when that is happening we might get
the empty planet result
which is the results that living
standard will be stagnant
and human race will start to disappear
now this is really an alarming result
and the reason for this is because the
private incentive of having children
we love children do not take into
account that
the children are producing ideas that
are useful for technology
technological progress so clearly
there's a role for
policy here and which chat mentioned
earlier as well
so we could try to you know introduce
some policy that help
to exterminate people to have more
children and the problem is
if wait too long then the empty panel
results
will not be able cannot be avoided so
that is something really really
worrying then it goes to ben's paper
which gives you the alternative scenario
which to say what if we have the
following situation
okay suppose we think of human brain or
man is
basically like machine so artificial
intelligence
can replicate human brain in fact in
chinese
we say the computer we translate as in
the
electrical brain so it's really saying
its electrical brain really can
replacing
human brain if it can
then what we will obtain is that we can
avoid stagnation
which is the empty planet itself and
even more
we might be able to move through to a
technology technology singularity
where the artificial intelligence can
self-improve
and the growth can explode now i think
we are all kind of convinced by then the
singularity results
seems quite impossible because one
simple thing
one can say is that many essential
activity
that cannot be done by ai and because of
that
which is sometimes we call the bonus
effect because of that you will not get
the situation where growth explode
so let me focus on the situation whether
ai
can solve the problem that chad
mentioned
which is the stagnation result so
how possible is it really that we can
have ai
completely replace human in generating
technological progress
meaning in that d production function
we do not need human anymore we can just
have
ai in it how is that possible
so here's a brief timeline of the
development of artificial intelligence
which
is quite remarkable which started in
1950
so over the last 70 years a lot of
progress has been made
okay there's lots of great discovery
but is it enough and what do we look for
into the future
so there's a report by the stanford
university called
the artificial intelligence index report
what this shows a few points i want to
highlight
one is that human brain itself
is actually still needed in improving
ai so for the last 20 years
since uh 2010 for the last 10 years
2010 to 2019 what we've seen is
published paper about artificial
intelligence
has increased by 300 percent
and the paper online before they were
published
has increased by two thousand percent so
there's a huge
increase in how researchers are trying
to improve
ai and at the same time i we also see a
lot of students choose to go to
university to study ai so it looks like
we still need quite a lot of human brain
to pour into make the artificial
intelligence
to replace the human brain so that
progress being made in many areas but
there is a lot of you know
question here ai is good for searching
pattern
using the observed observed data okay so
that is basically how artificial
intelligence work
with big data but can it really work
like human brains
on intuition and imagination now on the
right hand side
here i took one example from this annual
report
which is to show a video to the machine
and ask the machine to recognize what is
going on in that video
when you show the video of some high
activity link
for example like zumba dancing the
precision rate is very high
the machine really pick up the activity
very easily
but if you look at these hot uh other
activities for example here it showed
the hardest activity is drinking coffee
so presumably when people enjoy their
coffee they do not do much
special movement and there's no special
characteristic
for the machine to pick up very easily
so the precision rate is less than 10
and it has been very little progress
over the last 10 years
so my take from this is that it's still
quite a long time for the artificial
intelligence
to completely replace the human brain
and
it really matters a lot to see
if the world is going to have stagnant
population in 80 years
do we have enough time to make the
artificial
intelligence replace human growth so
when you think about the future of
growth
here's the question which is less costly
and more likely
producing human brain or producing human
brain like artificial intelligence
can we human with the help of artificial
intelligence
actually create an einstein-like
artificial intelligence
it to me i don't know it seems quite
difficult
but on the other hand if we go back to
check jones paper
we say that we need policy we need
policy to
increase fertility but it's not an easy
task on its own
agitated women today face a trade-off
between
career concern and having children so
just by giving child care subsidy on
maternity leave these are costly policy
and most of the time it might not work
so when we think about fertility of
course there's lots of theory about
volatility here i'm just going to focus
on a few things
what what is behind this so if you look
at historically
how can we have very high fertility in
the past which is like five children
per woman so because
there is a big role played by family
farm so family farm on the right hand
side here is some data from the iro
which shows you how the fraction of
women working on family farm has been
declining over time
now family plum is very special it
creates demand for children
good children can help on the farm and
they also allow women to combine
home production and work but
the process of urbanization and
structural transformation
have come along with the disappearance
of family farm
modern day when a woman have to go to
work it really means
leaving home so making it incompatible
to combine home production
and work so if you look at home
production here i show you a picture
of the home production time per day and
market production time per day
for women and for men so the first bar
is the woman the second bar is men
and these two bars represent the world
what we see here is something
really striking women's hotel home hour
and child care time is triple of mad
men's
so for every one hour man that's for
home production women have to do three
hours
now that itself this kind of picture
might give
young women especially of course when
choosing whether to get married
and to have children while we see that
women's education
is rising and there is rising concern
for gender equality
so let me just conclude with this on the
future fertility
so i i hope uh you know i solo convinced
myself
you know the artificial intelligence
will take some time but if we don't
change anything in 80 years
population growth will go negative so we
need to really think about how we can do
something about fertility
child care subsidy and maternity leave
will not be enough
one possibility maybe it will help women
to
choose to have more children is that if
there's more possibility of outsourcing
home production to the market
but that really rely on the development
of the service economy
now of course their social norm is
important as well
the social norm around the role of a
mother can play a crucial role for a
woman's decision to become a mother
but social norms itself are changing
over time
and they will change and it will respond
to technology and policy
so some hope there is if these things
are all working
perhaps we can revert the trend of
fertility
to bring it up above the replacement
level
before or you know together with the
artificial intelligence
and that will be the future of growth
hope thank you very much
thank you very much rachel our next
discussant is philip trammell phil is an
economist
at the global priorities institute at
the university of oxford
his research interests lie at the
intersection of
economic theory and moral philosophy
with specific focus on the long term
and as part of this focus he is also an
expert on long and growth
issues and incidentally he has also
written a recent paper on growth and a
transformative ai
together with me in which we synthesize
the literature
related to the theme of today's webinar
feel the floor is yours
uh phil we cannot hear you
is there perhaps a microphone that you
may have to plug in
um can you hear me now
yes all right sorry about that
um right so thank you chad ben and
rachel
and thank you anton for um giving me
this chance to
see if i can keep up with the joneses
some of what i uh say will overlap with
what's already been said but
yeah hopefully i have something you to
say
as anton said at the beginning when
thinking about growth
economists are typically content to
observe as
cal d'or first famously did that
growth has been roughly exponential at
two to three percent a year since the
industrial revolution
and so they'll assume that this will
continue
at least over the time scales that they
care about
sometimes they do this bluntly by just
stipulating an exogenous growth process
going on in the background and then
you know studying something else but
even when constructing endogenous or
semi-endogenous growth models
that is ones that model the inputs to
growth explicitly research and so on
a primary concern of these models is
usually to match
this stylized description of growth over
the past few centuries
for example the agion jones and jones
paper that ben presented
is unusually sympathetic to the
possibility of a growth regime shift
and acceleration but even so
it focuses less on scenarios in which
capital becomes highly substitutable for
labor
and tech production ones that you know
overcome that
balmol effect on the ground that
as long as that phi parameter been
mentioned is positive
which i think the authors believed at
the time
then capital accumulation is enough to
generate explosive growth
which is not what we've historically
observed and restrictions along these
lines
appear throughout the growth literature
as a result alternate growth regimes
currently seem to just be off most
people's radar
for example environmental economists
think have to think about longer time
scales than most economists
but they typically just assume
exponential growth
or a growth rate that falls to zero over
the next few centuries
a recent survey of economists and
environmental scientists
just asked when will growth end um
as if that you know roughly
characterized the the uncertainty
and of those with an opinion about half
said within this century
and about half said never
no one seems to have filled in a comment
saying they thought it would accelerate
or
anything like that plus when asked why
it might end
insufficient fertility wasn't explicitly
listed as
a reason and no one seems to have
commented on its absence
but on a longer time frame accelerating
growth wouldn't be a historical
the growth rate was far lower before the
industrial revolution
and before the agricultural revolution
it was lower still
so some forecasts on the basis of these
longer run trends
have predicted continual acceleration to
grow sometimes in the near future
if it multiplied by a factor of 20 again
you know it might be
we might see 40 growth a year or
something
um furthermore radically faster growth
doesn't seem
deeply theoretically impossible i don't
think you know lots of systems do
grow very quickly if you put mold in a
petri dish
it'll multiply a lot faster than two
percent a year right
so more formally the ben paper
finds that you can get permanent
acceleration
under this innocent seeming pair of
conditions
first you need capital that can start
doing research without human input
or can substitute well enough to
overcome that palm oil effect
and second you need phi at least zero
the fish fishing out effect uh not
not too strong um
yeah just to just to recap what here's
what phi elite zero means
when you have advanced tech on the one
hand
it gets easier to advance further
because you have the aid of
all the tech you've already developed
and on the other hand it gets harder
because you've already picked all the
low hanging fruit
phi less than zero means the second
effect wins out
okay so as you can see these two
conditions are basically a way of
formalizing the idea
of recursively self-improving ai
leading to a singularity and then
translating that into the language of
economics
that's a that's a great you know
contribution and formalization
in its own right but a really nice thing
about it
is that it lets us test these
requirements for the singularitarian
scenario
so as ben noted a recent paper estimates
phi to be substantially negative
or using chad's notation beta to be
positive implying that um
even reproducing and self-improving
robot researchers
couldn't bring about a real singularity
like a you know type one or type two
but they could still bring about a
one-time growth rate increase
as long as yeah they can perform all the
tasks involved in research
in any event um this is just one model
there
there are plenty others andrew sandberg
here put together a
summary of these back in 2013 uh of you
know
what what people had come up with at the
time
and uh anton and i did the same more
recently
to cover the past decade of economist
engagement with ai
but i i think the most significant
contribution on on this
front is just the paper that ben
presented
um it solidifies my own belief for
whatever little it's worth
that an ai growth the growth explosion
of one kind or another
even just a you know a growth rate
increase rather than a singularity
is not inevitable but not implausible
and it's at least a scenario we should
have on our radar
um so
yeah this is all very valuable for those
of us
interested in thinking about the range
of possibilities for
long-run growth
for those of us also interested in
trying to shape how the long-run future
might go though
what we especially want to keep an eye
out for are
opportunities for very long run path
dependence
right not just forecasting
in fact um i think almost a general
principle for those
interested in maximizing their long-term
impact would be to
look for systems with multiple stable
equilibria
which have very different social
welfares in them
and we're not yet locked into one and
then to look for opportunities to
steer toward a good a good stable
equilibrium
so we have to ask ourselves does the
development of ai
um offer us any opportunities like this
um if so i don't think the economics
literature has yet identified and
identified them actually
as ben garfinkel here has pointed out a
philanthropist who saw electric power
coming decades in advance
might not have found that insight to be
decision relevant
it just doesn't really help you do good
there could be long-term consequences of
the social disruption ai could weak or
of who first develops ai and like takes
over the world or something
um and most dramatically if we do
something to prevent ai from
wiping out the human species that would
certainly be a case
of avoiding a very bad and very stable
equilibrium
but scenarios like these aren't really
represented in the economics literature
on ai
by contrast path dependency is
a really clear implication of chad's
paper
um we may have this once in forever
opportunity to
steer civilization from the empty planet
equilibrium
to the um expanding
cosmos equilibrium by lobbying for
policies that maintain positive
population growth and thus maintain a
positive incentive
to uh to fund research and fertility
to my mind this is a really important
and
novel insight and it would be worth a
lot more papers to trace out
more fully under what conditions it
holds
but um i think it's pretty robust so
the key ingredient is just that if
there's too much tech per person
um the social planner can stop finding
it worthwhile to pay for further
research
for the reasons chad explained fertility
has proportional consumption costs right
you have to
to bring about a proportional population
increase people have to
give up a certain fraction of their time
um
to have the children but it would no
longer be producing proportional
research increases
because there's this mountain of ideas
you can hardly add much to in
proportional terms
so as long as this dynamic holds
you'll get that pair of equilibrium
so for example in the model people's
utility takes this quirky form you see
here
where c is average consumption of the
time
and n is how many descendants people
have alive at a time
but you might wonder you know what if
people are
more utilitarian what if they're perhaps
number dampened
time separable utilitarians like this
well uh if their utility function takes
this form as
chad points out in in the paper actually
we get the same results
and the utility functions are basically
just monotonic
transformations of one another so they
represent the same preference ordering
as how you can see that
anyway likewise in yeah in the model
people generate innovation just by
living
this is equivalent to exogenously
stipulating that
a constant fraction of the population
has to work
as researchers full time
but what if research has to be funded by
the social planner
at the cost of having fewer people
working in final good output
and thus lower consumption well
then at least if my own scratch work is
right
we still have our two stable equilibria
and in fact in this case the bad one
stagnates even more fully
um research can zero out even though
it's not like
everyone has died off because it's just
not worth allocating any of the
population to research as opposed to
final good production
um finally
sort of like rachel is saying i think
there's an important interaction between
the models
if we're headed for the empty planet
equilibrium the technology level
plateaus
but the plateau level can depend on
policy decisions at the margin
right like research funding or just a
little bit more fertility even if it
doesn't break us out of the equilibrium
and the empty planet result doesn't hold
if capital can accumulate
costlessly and do the research for us
so maybe all that matters is
well maybe it could happen that all that
ends up mattering is just making sure we
make it over the ai threshold
and letting the ai take it from there
all right well to wrap up if we care
about the long run
we should consider a wider spectrum of
ways long-run
growth might unfold not just those
matching the caldorf acts the last few
centuries
if we care about influencing the long
run we should also look for those rare
pivotal opportunities to change which
scenario plays out to
simplify a lot the bend paper helps us
with
the former showing how
a growth singularity via ai may or may
not
be compatible with reasonable economic
modeling
and the chad paper helps us with the
latter showing a counter-intuitive
channel through which
we could get locked into a low growth
equilibrium
sort of ironically via excessive tech
per person
and a policy channel that could avert it
he focuses on fertility subsidies
destroying technological ideas would do
the trick too because it would you know
shrink the number of ideas per person
but hopefully the future of civilization
doesn't
ultimately depend on long-termists
taking to book burning
and yeah hopefully all this paves the
way for future research on
how we can reach an expanding cosmos so
thank you
thank you phil and thank you all for
your contributions
and to everyone who has posted so many
interesting
questions in our q a now
luckily many of them have already been
answered in writing
because we are at the end of our
allocated time
so let me perhaps just uh
let both of our speakers have 30 seconds
to give us a quick reaction to the
discussion
uh then would you like to go first sure
iowa thanks everyone for all the great
questions in the q and a thanks rachel
and phil for very interesting uh
discussions i use very interesting to
pair these papers and i think
you know the distinction of whether you
know can you automate the ideas
production function
or not i mean that you know that's kind
of where what do we believe about that
um uh in terms of which which very
different trajectory do we end up on i
think it's a super interesting question
for a search
i guess you know just last comment you
know i think the singularity type people
you know they tell a story something
like you get a computer one algorithm
that's as good or better than a human
and because you can then
have huge increasing returns to scale
from that invention of that algorithm
that ai you then can just keep repeating
it over and over again as
instantiations on computing equipment
then you kind of can get to sort of
infinite input into the or very high
input growth into the
idea production function and i mean
that's where people get that's where you
get this
really get really strong externality i
think in a more micro statement of
what's going on
but i think the point that you know chad
and i are making that first paper with
philippe you know
another way to think about it is
actually you're not just going to repeat
the human
we're going to what we're going to do
it's sort of like we had a slide rule
and then we had a computer
and we have centrifuges we're gonna
we're gonna you know we have
automated pipetting we're gonna you know
it's again research just like production
is a whole set of different tasks
and probably what's gonna happen is
we're gonna slowly continue to automate
uh
some of those tasks and and you know the
more you automate you know the more you
leverage the
people who are left and can throw
capital at those automated tasks and i
think that is the way
that still away doesn't get you to
single areas necessarily but it's the
way potentially past uh the point chat's
making
uh and it's very who knows uh to me but
but i think it's really interesting i
think i think
again i think this work collectively
helps us really think about
where the where the rubber hits the road
in terms of what we have to believe and
where the action will be
in terms of the long run the long run uh
outcomes
think you've been chad yeah so let me uh
thank uh phil and rachel for excellent
discussions those were really
informative
and um i i think the one thing i took
away from your discussion and from
pairing these two papers together
is the point that you you both
identified so i'll just repeat it i
think it's important
you know an interesting question is does
the
ai revolution come soon enough to avoid
the empty planet and i think that's
really when you put these papers
together
the thing that jumps out of you the most
and as phil kind of mentioned in
ben was just referring to you know small
improvements
can help you get there and so maybe it's
possible to leverage our way into that
but
it's by no means obvious as it's been
pointed out if you've got this you know
fixed pool of ideas
then the ai improves the the fissures
but doesn't change the pool and so
you know i think a lot of these
questions deserve a lot more research
and so i think
anton thanks for putting this session
together i think it was really great and
very helpful
thank you everyone for joining us today
and i hope to see you again soon
at one of our future webinars on the
governance and economics of ai
bye
you |
ba1a7759-8258-4248-b2b3-718fbbe9dcfa | trentmkelly/LessWrong-43k | LessWrong | Why should I care about rationality?
|
606a4509-f533-48a1-aeae-e977be30ddb9 | trentmkelly/LessWrong-43k | LessWrong | Superposition Without Compression: Why Entangled Representations Are the Default
I have heard numerous claims recently that the underparameterisation of neural networks can be implied due to the polysemanticity of its neurons, which is prevalent in LLMs.
Whilst I have no doubt that polysemanticity is the only solution to an underparameterised model, I urge on the side of caution when using polysemanticity as proof of underparametarisation.
In this note I claim that: even when sufficient capacity is available, superposition may be the default due to its overwhelming prevalence in the solution space. Disentangled, monosemantic solutions occupy a tiny fraction of the total low-loss solutions.
This suggests that superposition arises not just as a necessity in underparametarised models, but also is an inevitability of the search space of neural networks.
In this note I show a comprehensible toy example where this is the case and hypothesise that this is also the case in larger networks.
These were very rough Sunday musings so I am very interested about what other people think about this claim :). |
d6e32b87-ce52-46c9-ac83-cc90d81e0285 | trentmkelly/LessWrong-43k | LessWrong | It's (not) how you use it
Crossposted from the EA Forum: https://forum.effectivealtruism.org/posts/LwhzE3scZTqxERtNn/it-s-not-how-you-use-it
The phrase "technology isn't bad in itself, it's just how you use it" is commonplace and contains some truth. But I think it's a mistake to go straight into judging the usage of technological products and not think about their design. Sure, it's intuitive to suppose that the choices humans make with how they interact with technologies play a decisive role in what purpose the technology ends up serving. My argument is that these choices are to be made earlier in the design and production of a certain technology; they're not choices humans find themselves making once they've acquired a technology. At that point, it's usually too late.
In History & Philosophy of Science (HPS) studies, this approach broadly falls into the camp of Marxist theories about the history of technology in the sense that the technological product has a "purpose", an "end" and it can have intrinsic risks. These risks, for this type of theorizing primarily concern the inscription of social norms and regularities that change the dynamics within society. Translated into the EA framework, these might be existential or suffering, and cost us the continuation of our species. It is, as a result, careless and irresponsible to create technologies without having clarity on what they'll be good for and how they could lead to catastrophic scenarios.
In the book Human Compatible, Stuart Russell shows how this irresponsibility applies to the development of ML. The analogy is simple: it's like preparing a mission to another planet without considering in advance how your crew is going to survive once they're on the new planet. If you expect them to deal with whatever risks and problems the environment of the new planet might have for humans after landing there, then you're not taking seriously the inherent dangers of your project, and quite frankly, the project itself. In other words, this is |
c588c50c-33a6-4cc1-9e39-57acfd8078f1 | trentmkelly/LessWrong-43k | LessWrong | Joy in the Merely Real
> ...Do not all charms fly
> At the mere touch of cold philosophy?
> There was an awful rainbow once in heaven:
> We know her woof, her texture; she is given
> In the dull catalogue of common things.
> —John Keats, Lamia
>
> "Nothing is 'mere'."
> —Richard Feynman
You've got to admire that phrase, "dull catalogue of common things". What is it, exactly, that goes in this catalogue? Besides rainbows, that is?
Why, things that are mundane, of course. Things that are normal; things that are unmagical; things that are known, or knowable; things that play by the rules (or that play by any rules, which makes them boring); things that are part of the ordinary universe; things that are, in a word, real.
Now that's what I call setting yourself up for a fall.
At that rate, sooner or later you're going to be disappointed in everything—either it will turn out not to exist, or even worse, it will turn out to be real.
If we cannot take joy in things that are merely real, our lives will always be empty.
For what sin are rainbows demoted to the dull catalogue of common things? For the sin of having a scientific explanation. "We know her woof, her texture", says Keats—an interesting use of the word "we", because I suspect that Keats didn't know the explanation himself. I suspect that just being told that someone else knew was too much for him to take. I suspect that just the notion of rainbows being scientifically explicable in principle would have been too much to take. And if Keats didn't think like that, well, I know plenty of people who do.
I have already remarked that nothing is inherently mysterious—nothing that actually exists, that is. If I am ignorant about a phenomenon, that is a fact about my state of mind, not a fact about the phenomenon; to worship a phenomenon because it seems so wonderfully mysterious, is to worship your own ignorance; a blank map does not correspond to a blank territory, it is just somewhere we h |
e3e2311c-d7cc-44f2-970e-08ce59141524 | trentmkelly/LessWrong-43k | LessWrong | Economics Roundup #2
Previously: Economics Roundup #1
Let’s take advantage of the normality while we have it. In all senses.
INSANE TAX PROPOSALS
There is Trump’s proposal to replace income taxes with tariffs, but he is not alone.
So here is your periodic reminder, since this is not actually new at core: Biden’s proposed budgets include completely insane tax regimes that would cripple our economic dynamism and growth if enacted. As in for high net worth individuals, taking unrealized capital gains at 25% and realized capital gains, such as those you are forced to take to pay your unrealized capital gains tax, at 44.6% plus state taxes.
Austen Allred explains how this plausibly destroys the entire startup ecosystem.
Which I know is confusing because in other contexts he also talks about how other laws (such as SB 1047) that would in no way apply to startups would also destroy the startup ecosystem. But in this case he is right.
> Austen Allred: It’s difficult to describe how insane a 25% tax on unrealized capital gains is.
>
> Not a one-time 25% hit. It’s compounding, annually taking 25% of every dollar of potential increase before it can grow.
>
> Not an exaggeration to say it could single-handedly crush the economy.
>
> An example to show how insane this is: You’re a founder and you start a company. You own… let’s say 30% of it.
>
> Everything is booming, you raise a round that values the company at at $500 million.
>
> You now personally owe $37.5 million in taxes.
>
> This year. In cash.
>
> Now there are investors who want to invest in the company, but you can’t just raise $37.5 million in cash overnight.
>
> So what happens?
>
> Well, you simply decide not to have a company worth a few hundred million dollars.
>
> Oh well, that’s only a handful of companies right?
>
> Well, as an investor, the only way the entire ecosystem works is if a few companies become worth hundreds of millions.
>
> Without that, venture capital no longer works. Investment is gone.
>
> Y C |
2959625c-c1b4-430d-8465-ac7c0b66a680 | trentmkelly/LessWrong-43k | LessWrong | Don’t waste your time meditating on meditation retreats!
Do "Emotional Work Retreats" instead!
TLDR;
If your goal is therapeutic progress - the kind of progress you also hope to get from going to a therapist, i.e. become happier long term, get rid of unhealthy emotional patterns, and improve your relationships with others and yourself - then spending 6+ hours meditating each day on meditation retreats is not the best use of your time. Instead, I recommend you meditate for 3-4 hours each day and spend the rest of your time doing various therapeutic practices like Internal Family Systems, Focusing, Ideal Parent Figures, CBT, Core Transformations, etc. If possible, include as many therapy sessions as possible during the retreat (potentially over Zoom), as well as some deep talk, Authentic Relating, cuddling, etc. with friends or even with romantic partners. I would still remain silent as much as possible for most of the retreat except for anything related to emotional work.
I offer 'Being your retreat buddy' as a service, i.e. sitting 'emotional work retreats' with you for 30€/day.
What I would NOT do on meditation retreats
I recommend not getting distracted by and not "wasting your time" trying to do any of the following:
* Become enlightened, reach the Jhanas, stream-entry, etc.
* Generally, "make progress" with meditation. Deepen your practice. Gain Insight into Buddhist concepts like Impermanence, No-Self, Emptiness, Arising & Passing, the nature of suffering, etc.
* Really figuring out Sam Harris’s “Look for the one who’s looking” or other so-called “pointing out instructions”, perfecting non-dual mindfulness
In my view, these endeavors are simply not the most effective/quickest way towards therapeutic progress. They are also somewhat "all or nothing": You gain almost nothing becoming almost enlightened or almost reaching the Jhanas, but you will have wasted a lot of time. Even if you reach the Jhanas, it's not clear at all how they help you with long-term therapeutic progress. And in my experience, insight |
ffc132ad-9b00-471e-8023-c186fd33d001 | trentmkelly/LessWrong-43k | LessWrong | Meetup : Melbourne Practical Rationality Meetup
Discussion article for the meetup : Melbourne Practical Rationality Meetup
WHEN: 04 October 2013 06:30:00PM (+1000)
WHERE: Level 2, 491 King St West Melbourne 3003
Practical rationality. This meetup repeats on the 1st Friday of each month and is distinct from our monthly social meetup. We aim to improve our thinking and decision making techniques, and 'level up'.
Topics for this month will be announced on our group's meetup page. http://www.meetup.com/Melbourne-Less-Wrong/
Discussion can be found on our mailing list: http://groups.google.com/group/melbourne-less-wrong
All are welcome from 6:30pm. If the door downstairs is locked, call the phone number on the door and we'll let you in. Any formal activity will likely kick off at 7:30pm sharp an g until 9pm. Following this there will be informal discussion and rationality life advice.
Please RSVP at the meetup page above so we can be aware of numbers.
Discussion article for the meetup : Melbourne Practical Rationality Meetup |
574c41e7-562c-42c2-a149-2ff0362b76d9 | trentmkelly/LessWrong-43k | LessWrong | how has this forum changed your life?
Hello! I am totally new here so please bear that in mind in the event I commit faux pas! I'm a writer who has written a LOT about rationality and when rationality eludes us. These are books like So You've Been Publicly Shamed, Them, The Psychopath Test and The Men Who Stare At Goats, among others.
I don't know if this question has been asked elsewhere, but I'd love to know: has learning to be more rational impacted your everyday lives, in small or (perhaps more interestingly) BIG ways? Have there been occasions when you've put these skills (perhaps learned at a workshop) into practice in your domestic or work lives and the ripples have been surprising/consequential in positive or negative ways?
I'm thinking of writing about what happens when a person learns to be more rational, hence my question.
Hope this is the start of a wonderful thread!
Jon Ronson |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.