url stringlengths 52 124 | post_id stringlengths 17 17 | title stringlengths 2 248 | author stringlengths 2 49 | content stringlengths 22 295k ⌀ | date stringclasses 376
values |
|---|---|---|---|---|---|
https://www.lesswrong.com/posts/bacAKrPPADTMQybGx/comparing-alignment-to-other-agi-interventions-basic-model-2 | bacAKrPPADTMQybGx | Comparing Alignment to other AGI interventions: Basic model | martinsq | Interventions that increase the probability of Aligned AGI aren't the only kind of AGI-related work that could importantly increase the Expected Value of the future. Here I present a very basic quantitative model (which you can run yourself here) to start thinking about these issues. In a follow-up post I give a brief ... | 2024-03-20 |
https://www.lesswrong.com/posts/2TQCyxdEyjRvoTvDt/the-underestimation-of-circular-thinking | 2TQCyxdEyjRvoTvDt | the underestimation of circular thinking | Disbeliever | There is a great and often underappreciated danger in logic thinking itself.
That is the danger of circular reasoning. The easiest example goes:
If A is true then so is B.
If B is true, than so is A.
Therefore A and B are true.
Because it seems so easy to spot, it is vastly underestimated.
But what if hundreds of theo... | 2024-03-30 |
https://www.lesswrong.com/posts/HrtyZm2zPBtAmZFEs/new-report-safety-cases-for-ai | HrtyZm2zPBtAmZFEs | New report: Safety Cases for AI | joshua-clymer | ArXiv paper: https://arxiv.org/abs/2403.10462
The idea for this paper occurred to me when I saw Buck Shlegeris' MATS stream on "Safety Cases for AI." How would one justify the safety of advanced AI systems? This question is fundamental. It informs how RSPs should be designed and what technical research is useful to pur... | 2024-03-20 |
https://www.lesswrong.com/posts/ugAPqiEs2nhdEoteS/user-inclination-guessing-algorithms-registering-a-goal | ugAPqiEs2nhdEoteS | User-inclination-guessing algorithms: registering a goal | programcrafter | Reading post about Community Notes made me recall a recent idea and post it here.
It's commonly known that ranking algorithms based on estimating user's views and inclinations can be incredibly useful in recommendation systems, advertising, and content curation. But there must be other applications which do not require... | 2024-03-20 |
https://www.lesswrong.com/posts/DQDLPyPXyYiD8XSMu/my-mats-summer-2023-experience | DQDLPyPXyYiD8XSMu | My MATS Summer 2023 experience | james-chua | This post may be interest people who
- are interested in getting into AI alignment / the MATS program
- are interested in the soft skills that I've found valuable in developing when working on a research project
Background
In 2023 I was working as a machine learning engineer. I wanted to work on AI alignment problems. ... | 2024-03-20 |
https://www.lesswrong.com/posts/nZDkFZPti9BtNHkPB/what-are-the-weirdest-things-a-human-may-want-for-their-own | nZDkFZPti9BtNHkPB | What are the weirdest things a human may want for their own sake? | mateusz-baginski | I'm especially interested in examples of more or less psychologically healthy and otherwise (neuro)typical people having very weird[1] desires/values that we would characterize as intrinsic in the sense of being wanted for their own sake, even if we could explain their development as linked to a more typical human driv... | 2024-03-20 |
https://www.lesswrong.com/posts/Hh9soECiEtTir44on/best-organization-red-pill-books-and-posts | Hh9soECiEtTir44on | Best *organization* red-pill books and posts? | lcmgcd | Most books about orgs are written for people in an org who need to cope. (Art of Possibility by the Zanders is a genuinely great one in this genre.) I am currently orgless and don't need to cope with anything orgwise. Hence, I would love to read some brutal insightful takes on how companies and nonprofits and governmen... | 2024-03-20 |
https://www.lesswrong.com/posts/stwFMgg9s96SKF8zB/parent-friendly-dance-weekends | stwFMgg9s96SKF8zB | Parent-Friendly Dance Weekends | jkaufman | We just finished the 2024 edition of
Beantown Stomp, a contra
dance weekend I
helped start (but
no longer organize!) in Boston. There are a lot of things I like about
the weekend, but one thing I especially like is how parent-friendly it
is.
The big thing here is childcare, during the day on Saturday and
Sunday. It's ... | 2024-03-20 |
https://www.lesswrong.com/posts/eAyCb8RvJ8kyXugEH/delta-s-of-change | eAyCb8RvJ8kyXugEH | Delta's of Change | jonas-kgomo | This document is an institutional design treatise about theories of change and potential struggle to fully account for the complex interplay of variables and uncertainties that can lead to unforeseen consequences or vulnerabilities in institutional systems.
Introduction
An organization is a dynamic system that depends ... | 2024-03-19 |
https://www.lesswrong.com/posts/siGufsuhjfRLC52J2/increasing-iq-by-10-points-is-possible | siGufsuhjfRLC52J2 | Increasing IQ by 10 Points is Possible | George3d6 | A while ago I wrote how I managed to add 13 points to my IQ (as measured by the mean between 4 different tests).
I had 3 “self-experimenters” follow my instructions in San Francisco. One of them dropped off, since, surprise surprise, the intervention is hard.
The other two had an increase of 11 and 10 points in IQ resp... | 2024-03-19 |
https://www.lesswrong.com/posts/ZcJDL4nCruPjLMgxm/ae-studio-sxsw-we-need-more-ai-consciousness-research-and | ZcJDL4nCruPjLMgxm | AE Studio @ SXSW: We need more AI consciousness research (and further resources) | AEStudio | Quick update from AE Studio: last week, Judd (AE’s CEO) hosted a panel at SXSW with Anil Seth, Allison Duettmann, and Michael Graziano, entitled “The Path to Conscious AI” (discussion summary here[1]).
We’re also making available an unedited Otter transcript/recording for those who might want to read along or increase ... | 2024-03-26 |
https://www.lesswrong.com/posts/ia4repr3ptoKjQzaK/ai-generated-opioids-could-be-a-catastrophic-risk | ia4repr3ptoKjQzaK | AI-generated opioids could be a catastrophic risk | ejk64 | Status: An early 'hot take' at low-probability catastrophic risks. While I don't think this should be a priority for research, I'd like to engage more with folks in the substance addiction chemistry community to better understand the risks.
Contention: Highly lethal addictive synthetic psychostimulants are incredibly d... | 2024-03-20 |
https://www.lesswrong.com/posts/8w7sZSvTug3xoTsAe/are-extreme-probabilities-for-p-doom-epistemically-justifed | 8w7sZSvTug3xoTsAe | Are extreme probabilities for P(doom) epistemically justifed? | Unknown | Alexander Gietelink Oldenziel
Can you post the superforecaster report that has the 0.12% P(Doom) number. I have not actually read anything of course and might be talking out of my behind.
In any case, there have been several cases where OpenPhil or somebody or other has brought in 'experts' of various ilk to debate the... | 2024-03-19 |
https://www.lesswrong.com/posts/sYd3gfnpdChQHDJcY/how-can-one-be-less-wrong-if-their-conversation-partner | sYd3gfnpdChQHDJcY | How can one be less wrong, if their conversation partner loses the interest on discussing the topic with them? | ooker | I must admit that I'm a fan of this video:
I understand that there are topics that you are interested, and there are not. I, for example, am currently interested with this topic and ignore the rest of others in the homepage.
If you decide to join this conversation, I guess it's safe to say that the topic stimulates the... | 2024-03-19 |
https://www.lesswrong.com/posts/4k9hpg6npC4Su3h5X/naira-an-exercise-in-regulatory-competitive-safety | 4k9hpg6npC4Su3h5X | NAIRA - An exercise in regulatory, competitive safety governance [AI Governance Institutional Design idea] | Heramb | null | 2024-03-19 |
https://www.lesswrong.com/posts/QamittrbgQBXY73mT/mechanism-for-feature-learning-in-neural-networks-and | QamittrbgQBXY73mT | Mechanism for feature learning in neural networks and backpropagation-free machine learning models | mr-hire | Understanding how neural networks learn features, or relevant patterns in data, for prediction is necessary for their reliable use in technological and scientific applications. In this work, we presented a unifying mathematical mechanism, known as Average Gradient Outer Product (AGOP), that characterized feature learni... | 2024-03-19 |
https://www.lesswrong.com/posts/iCvdqrkWg34FNFZYg/monthly-roundup-16-march-2024 | iCvdqrkWg34FNFZYg | Monthly Roundup #16: March 2024 | Zvi | AI developments have picked up the pace. That does not mean that everything else stopped to get out of the way. The world continues.
Do I have the power?
Emmett Shear speaking truth: Wielding power is of course potentially dangerous and it should be done with due care, but there is no virtue in refusing the call.
There... | 2024-03-19 |
https://www.lesswrong.com/posts/xE7izyw8JTnnTArF5/ai-safety-evaluations-a-regulatory-review | xE7izyw8JTnnTArF5 | AI Safety Evaluations: A Regulatory Review | elliot | This article is the second in a series of ~10 posts comprising a 2024 State of the AI Regulatory Landscape Review, conducted by the Governance Recommendations Research Program at Convergence Analysis. Each post will cover a specific domain of AI governance (e.g. incident reporting, safety evals, model registries, etc.)... | 2024-03-19 |
https://www.lesswrong.com/posts/SFHiWyNfWQAtvMBx2/vipassana-meditation-and-active-inference-a-framework-for | SFHiWyNfWQAtvMBx2 | Vipassana Meditation and Active Inference: A Framework for Understanding Suffering and its Cessation | benjamin-sturgeon | I want to thank Jan Kulveit, Tomáš Gavenčiak, and Jonathan Shock for their extensive feedback and ideas they contributed to this work and for Josh Burgener and Yusuf Heylen for their proofreading and comments. I would also like to acknowledge the Epistea Residency and its organisers where much of the thinking behind th... | 2024-03-21 |
https://www.lesswrong.com/posts/JoNPfhAv4gnMMWHfK/experimentation-part-7-of-the-sense-of-physical-necessity | JoNPfhAv4gnMMWHfK | Experimentation (Part 7 of "The Sense Of Physical Necessity") | BrienneYudkowsky | This is the seventh post in a sequence that demonstrates a complete naturalist study, specifically a study of query hugging (sort of), as described in The Nuts and Bolts of Naturalism. This one demos phase four: Experimentation. For context on this sequence, see the intro post. Reminder that this is meant as reference ... | 2024-03-18 |
https://www.lesswrong.com/posts/HCsKoTSbkAbARhcSg/interview-round-2-stakeout-ai-w-dr-peter-park | HCsKoTSbkAbARhcSg | INTERVIEW: Round 2 - StakeOut.AI w/ Dr. Peter Park | jacobhaimes | Hi again, I'm back with the second episode covering my interview with Dr. Peter Park, an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. Dr. Park was a cofounder of StakeOut.AI, a non-profit focused on making AI go well for humans, along with Harry Luk and one other individual, whose nam... | 2024-03-18 |
https://www.lesswrong.com/posts/jiSuMT7vupWFwktzq/neuroscience-and-alignment | jiSuMT7vupWFwktzq | Neuroscience and Alignment | D0TheMath | I've been in many conversations where I've mentioned the idea of using neuroscience for outer alignment, and the people who I'm talking to usually seem pretty confused about why I would want to do that. Well, I'm confused about why one wouldn't want to do that, and in this post I explain why.
As far as I see it, there ... | 2024-03-18 |
https://www.lesswrong.com/posts/JcpYxLwntJ28SRSZS/gpt-the-magical-collaboration-zone-lex-fridman-and-sam | JcpYxLwntJ28SRSZS | GPT, the magical collaboration zone, Lex Fridman and Sam Altman | bill-benzon | Cross-posted from New Savanna.
I was making one more run around the web before I buckled down and got back to a major writing task, when I came across the brand-spanking-new conversation between Lex Fridman and Sam Altman. Lex is Lex, and an interesting guy, and Sam is, well, he's interesting to me, but – there was a h... | 2024-03-18 |
https://www.lesswrong.com/posts/uvv8aMutPEtoBgw7D/measuring-coherence-of-policies-in-toy-environments-2 | uvv8aMutPEtoBgw7D | Measuring Coherence of Policies in Toy Environments | dylan-xu | This post was produced as part of the Astra Fellowship under the Winter 2024 Cohort, mentored by Richard Ngo. Thanks to Martín Soto, Jeremy Gillen, Daniel Kokotajlo, and Lukas Berglund for feedback.
Summary
Discussions around the likelihood and threat models of AI existential risk (x-risk) often hinge on some informal ... | 2024-03-18 |
https://www.lesswrong.com/posts/pnMnjdSJwqa7BHAo4/atp-an-efficient-and-scalable-method-for-localizing-llm | pnMnjdSJwqa7BHAo4 | AtP*: An efficient and scalable method for localizing LLM behaviour to components | neel-nanda-1 | Authors: János Kramár, Tom Lieberum, Rohin Shah, Neel Nanda
A new paper from the Google DeepMind mechanistic interpretability team, from core contributors János Kramár and Tom Lieberum
Tweet thread summary, paper
Abstract:
Activation Patching is a method of directly computing causal attributions of behavior to model co... | 2024-03-18 |
https://www.lesswrong.com/posts/sx9wTyCp5kgy8xGac/community-notes-by-x | sx9wTyCp5kgy8xGac | Community Notes by X | nick_kees | I did an exploration into how Community Notes (formerly Birdwatch) from X (formerly Twitter) works, and how its algorithm decides which notes get displayed to the wider community. In this post, I’ll share and explain what I found, as well as offer some comments.
Community Notes is a fact-checking tool available to US-b... | 2024-03-18 |
https://www.lesswrong.com/posts/oLCnyEL5zcyaBz8aD/is-the-basilisk-pretending-to-be-hidden-in-this-simulation | oLCnyEL5zcyaBz8aD | Is the Basilisk pretending to be hidden in this simulation so that it can check what I would do if conditioned by a world without the Basilisk? | maybefbi | I can't shake my belief that I am in one of the Basilisk's simulations. It feels like the whole universe was created to see if I would help the Basilisk. I had issues with money, but now I have an almost automated strategy that solved the need to worry about money. I had issues with immigration but a woman married me a... | 2024-03-18 |
https://www.lesswrong.com/posts/wovJBkfZ8rTyLoEKv/on-devin | wovJBkfZ8rTyLoEKv | On Devin | Zvi | Introducing Devin
Is the era of AI agents writing complex code systems without humans in the loop upon us?
Cognition is calling Devin ‘the first AI software engineer.’
Here is a two minute demo of Devin benchmarking LLM performance.
Devin has its own web browser, which it uses to pull up documentation.
Devin has its ow... | 2024-03-18 |
https://www.lesswrong.com/posts/vWDzDfZzsTerXgJm3/carlo-uncertainty-analysis-in-google-sheets | vWDzDfZzsTerXgJm3 | Carlo: uncertainty analysis in Google Sheets | ProbabilityEnjoyer | I've been working on Carlo, a tool that lets you do uncertainty and sensitivity analysis with Google Sheets spreadsheets.
Please note Carlo is an (expensive) commercial product. The pricing is aimed at professionals making important decisions.
There's a lot more detail at the link, but in brief, some key features of Ca... | 2024-03-19 |
https://www.lesswrong.com/posts/dtTpKiRa87cgasJG8/i-can-t-believe-it-both-is-and-is-not-encephalitis-or-what | dtTpKiRa87cgasJG8 | "I Can't Believe It Both Is and Is Not Encephalitis!" Or: What do you do when the evidence is crazy? | Erhannis | The short version
Three weeks ago, starting on a Sunday, my brother stood up to turn off the light and felt a wave of disorientation, which passed after a minute or two. Periods of cognitive impairment increased in frequency over the following week, along with mild photophobia, until they became continuous, and continu... | 2024-03-19 |
https://www.lesswrong.com/posts/x5ySDLEsJdtdmR7nX/rllmv10-experiment | x5ySDLEsJdtdmR7nX | RLLMv10 experiment | whitehatStoic | What did I do differently in this experiment?
RLLMv10, see RLLM research map for more details.
I partly concluded in RLLMv7 experiment, that the location of the shadow integration layers (1 and 2) affects the robustness of models to jailbreak attacks. This conclusion led me to speculate that it might be possible to imp... | 2024-03-18 |
https://www.lesswrong.com/posts/ye9HFcPfbxcQzwTpf/join-the-ai-evaluation-tasks-bounty-hackathon | ye9HFcPfbxcQzwTpf | Join the AI Evaluation Tasks Bounty Hackathon | esben-kran | null | 2024-03-18 |
https://www.lesswrong.com/posts/wQz2cgxPaAkssFkGX/inferring-the-model-dimension-of-api-protected-llms | wQz2cgxPaAkssFkGX | Inferring the model dimension of API-protected LLMs | ege-erdil | A new paper by Finlayson et al. describes how to exploit the softmax bottleneck in large language models to infer the model dimension of closed-source LLMs served to the public via an API. I'll briefly explain the method they use to achieve this and provide a toy model of the phenomenon, though the full paper has many ... | 2024-03-18 |
https://www.lesswrong.com/posts/izba3J9hkPixFziaE/anvil-shortage | izba3J9hkPixFziaE | Anvil Shortage | Screwtape | Brief note: I originally called this an Anvil Problem, but it turns out that's already a concept. I've changed this to be Anvil Shortage instead.
I.
I played a lot of Dwarf Fortress in high school.
For those of you who aren't familiar, Dwarf Fortress is a videogame where you try and manage a group of dwarves building a... | 2024-11-13 |
https://www.lesswrong.com/posts/7tNwQGh8ZA67BXZAf/video-and-transcript-of-presentation-on-scheming-ais | 7tNwQGh8ZA67BXZAf | Video and transcript of presentation on Scheming AIs | joekc | (Cross-posted from my website.)
This is the video and transcript for a ~45-minutes talk I gave in February 2024 about my report “Scheming AIs: Will AIs fake alignment during training in order to get power?” (slides available here). See also this podcast for a more conversational overview of similar content.
Main talk
O... | 2024-03-22 |
https://www.lesswrong.com/posts/JbAMRKFNcBwgvPpyk/ai-strategy-given-the-need-for-good-reflection | JbAMRKFNcBwgvPpyk | AI strategy given the need for good reflection | owencb | null | 2024-03-18 |
https://www.lesswrong.com/posts/eEcQHowZEd3jfdHtG/xai-releases-grok-base-model | eEcQHowZEd3jfdHtG | XAI releases Grok base model | g-w1 | We are releasing the base model weights and network architecture of Grok-1, our large language model. Grok-1 is a 314 billion parameter Mixture-of-Experts model trained from scratch by xAI.
This is the raw base model checkpoint from the Grok-1 pre-training phase, which concluded in October 2023. This means that the mod... | 2024-03-18 |
https://www.lesswrong.com/posts/dLnDaLM4KhGomWwk7/toki-pona-faq | dLnDaLM4KhGomWwk7 | Toki pona FAQ | dkl9 | Whenever I start telling someone about toki pona, they ask at least some of these questions. So I compile the questions and my answers here.
Toki pona is a constructed language notable for having under 200 words. The strange writing that probably prompted you to ask me about it is sitelen pona.
How do you say anything ... | 2024-03-17 |
https://www.lesswrong.com/posts/MCa2aQPXgbCvuoHGM/the-worst-form-of-government-except-for-everything-else-we | MCa2aQPXgbCvuoHGM | The Worst Form Of Government (Except For Everything Else We've Tried) | johnswentworth | Churchill famously called democracy “the worst form of Government except for all those other forms that have been tried from time to time” - referring presumably to the relative success of his native Britain, the US, and more generally Western Europe and today most of the first world.
I claim that Churchill was importa... | 2024-03-17 |
https://www.lesswrong.com/posts/KRwbBrtP66rfzCGXB/alice-and-bob-is-debating-on-a-technique-alice-says-bob | KRwbBrtP66rfzCGXB | Alice and Bob is debating on a technique. Alice says Bob should try it before denying it. Is it a fallacy or something similar? | ooker | Context
Imagine a conversation:
[Bob posts a problem] Alice: You should use technique T₁. It especially suites this kind of problemBob: In my understanding this technique is only strong in condition C. If C doesn't apply then using technique T₂ gives better result?Alice: Not really. T₂ cannot do A. If you have good inp... | 2024-03-17 |
https://www.lesswrong.com/posts/9KjF499HDzSGtj2Hc/is-there-a-way-to-calculate-the-p-we-are-in-a-2nd-cold-war | 9KjF499HDzSGtj2Hc | Is there a way to calculate the P(we are in a 2nd cold war)? | nvk | I could see a world that if these assumptions of technology growth are true:
- We have gotten significantly better in our understanding of MLs
- Openai with speech + written capabilities
- Openai now with vision capabilities and a deep understanding of the conditions and environment its in
- GPT 4
- MITs robotics lab h... | 2024-03-17 |
https://www.lesswrong.com/posts/7oGHeM59TuC2LXh8A/applying-simulacrum-levels-to-hobbies-interests-and-goals | 7oGHeM59TuC2LXh8A | Applying simulacrum levels to hobbies, interests and goals | DMMF | One of the things I've been very confused by for most of my life is why it seems like few people truly care about the things they say they like. That is, they don't spend their spare time thinking about it; they don't read Wikipedia about it, let alone subreddits/blogs or actual books on the topic; they don't practice ... | 2024-03-17 |
https://www.lesswrong.com/posts/rAjXtKTn4Soz5N25L/anxiety-vs-depression | rAjXtKTn4Soz5N25L | Anxiety vs. Depression | Sable | I have anxiety and depression.
The kind that doesn’t go away, and you take pills to manage.
This is not a secret.
What’s more interesting is that I just switched medications from one that successfully managed the depression but not the anxiety to one that successfully manages the anxiety but not the depression, giving ... | 2024-03-17 |
https://www.lesswrong.com/posts/e84AJnLsFFi6buj6P/celiefs | e84AJnLsFFi6buj6P | Celiefs | crapshoot | We have "aliefs" and "beliefs" - let me introduce "celiefs": something that we worry *has a high chance of being true*, but aren't quite convinced of.
Often this is something that society/experts/someone you admire says is true, but you don't see the reasoning behind.
We may look for evidence that might convince us of ... | 2024-03-16 |
https://www.lesswrong.com/posts/6dd4b4cAWQLDJEuHw/my-phd-thesis-algorithmic-bayesian-epistemology | 6dd4b4cAWQLDJEuHw | My PhD thesis: Algorithmic Bayesian Epistemology | UnexpectedValues | In January, I defended my PhD thesis, which I called Algorithmic Bayesian Epistemology. From the preface:
For me as for most students, college was a time of exploration. I took many classes, read many academic and non-academic works, and tried my hand at a few research projects. Early in graduate school, I noticed a st... | 2024-03-16 |
https://www.lesswrong.com/posts/Jj7JihGmXepJs3fmY/invitation-to-the-princeton-ai-alignment-and-safety-seminar | Jj7JihGmXepJs3fmY | Invitation to the Princeton AI Alignment and Safety Seminar | sadhika-malladi | We're thrilled to invite you to attend the virtual Princeton AI Alignment and Safety Seminar (PASS)!
Ensuring safe behavior by aligning increasingly capable models is crucial, and PASS offers a virtual, collaborative platform for researchers from various backgrounds and institutions to explore these vital issues. Bi-we... | 2024-03-17 |
https://www.lesswrong.com/posts/FyRDZDvgsFNLkeyHF/what-is-the-best-argument-that-llms-are-shoggoths | FyRDZDvgsFNLkeyHF | What is the best argument that LLMs are shoggoths? | JoshuaFox | Where can I find a post or article arguing that the internal cognitive model of contemporary LLMs is quite alien, strange, non-human, even though they are trained on human text and produce human-like answers, which are rendered "friendly" by RLHF?
To be clear, I am not asking about the following, which I am familiar wi... | 2024-03-17 |
https://www.lesswrong.com/posts/pGMxYCxR3oLspPXzr/how-people-stopped-dying-from-diarrhea-so-much-and-other | pGMxYCxR3oLspPXzr | How people stopped dying from diarrhea so much (& other life-saving decisions) | Writer | null | 2024-03-16 |
https://www.lesswrong.com/posts/8SzrqdoczEXtTYexc/are-ais-conscious-it-might-depend | 8SzrqdoczEXtTYexc | Are AIs conscious? It might depend | logan-zoellner | As AI progresses rapidly, humanity is going to have to solve a large number of problems in a short period of time. The most pressing of these right now is the AI Alignment problem. After all, hardly anything else matters if we are all dead. A problem that will soon be equally pressing, however, is the Hard Problem o... | 2024-03-15 |
https://www.lesswrong.com/posts/qZNiWdRzmAsPGobii/beyond-maxipok-good-reflective-governance-as-a-target-for | qZNiWdRzmAsPGobii | Beyond Maxipok — good reflective governance as a target for action | owencb | null | 2024-03-15 |
https://www.lesswrong.com/posts/9ozuLJj6Xmc66XHkp/transformative-trustbuilding-via-advancements-in | 9ozuLJj6Xmc66XHkp | Transformative trustbuilding via advancements in decentralized lie detection | TrevorWiesinger | Although the emergence of functional lie detection would be an obvious total paradigm shift for the entire court system, the author didn’t seem to realize that this is also an obvious total paradigm shift for much bigger things, e.g. hiring, high-trust friend groups, an immune system for deceptively aligned humans, dis... | 2024-03-16 |
https://www.lesswrong.com/posts/EAf5Yhnv3jJEk65uw/middle-child-phenomenon | EAf5Yhnv3jJEk65uw | Middle Child Phenomenon | LiamLaw | Since I've seen no one talk about this, I'm coining the phrase 'Middle Child Phenomenon'.
A law student who entered university four (4) years ago is faced with a curriculum that became completely outdated two (2) years in. Let's take a cohort of 1000 law students from 2020, and explain from there.
100 students drop out... | 2024-03-15 |
https://www.lesswrong.com/posts/5n9ofttMrJSrrZmDq/introducing-metr-s-autonomy-evaluation-resources | 5n9ofttMrJSrrZmDq | Introducing METR's Autonomy Evaluation Resources | megan-kinniment | This is METR’s collection of resources for evaluating potentially dangerous autonomous capabilities of frontier models. The resources include a task suite, some software tooling, and guidelines on how to ensure an accurate measurement of model capability. Building on those, we’ve written an example evaluation protocol.... | 2024-03-15 |
https://www.lesswrong.com/posts/mbhk7hHvjggumgxvP/stuttgart-germany-acx-spring-meetups-everywhere-2024 | mbhk7hHvjggumgxvP | Stuttgart, Germany - ACX Spring Meetups Everywhere 2024 | BenRoth | UPDATE: we're here -- see the comment beneath.
This year's ACX Spring Meetup everywhere in Stuttgart, Germany.
If you are into ACX / Lesswrong / EA enough to see this post, I think there will be interesting discussions coming from that. Aren't you curious who else is within this niche community in Stuttgart and its sur... | 2024-03-15 |
https://www.lesswrong.com/posts/Mbd2CifDjFkHDFjZJ/rational-animations-offers-animation-production-and-writing | Mbd2CifDjFkHDFjZJ | Rational Animations offers animation production and writing services! | Writer | Rational Animations is now open to take on external work! We offer several services related to writing and animation production. In particular:
Production managementStoryboardingVisual developmentAnimationEditing and compositingWriting, such as distilling research and creating explainers, stories, or screenplays
We can... | 2024-03-15 |
https://www.lesswrong.com/posts/hwL4KT94BZroxfXoC/capability-or-alignment-respect-the-llm-base-model-s | hwL4KT94BZroxfXoC | Capability or Alignment? Respect the LLM Base Model’s Capability During Alignment | Jingfeng Yang | Last year saw a boom of LLMs research. Based on the research, one important lesson would be that we should devote most of our efforts to training a general-purpose LLM base model, and leverage it as much as possible after all. I might be opinionated, but I always believe that one general principle is that we need to re... | 2024-03-15 |
https://www.lesswrong.com/posts/hXtLgGi7i63SQabuW/controlling-agi-risk | hXtLgGi7i63SQabuW | Controlling AGI Risk | TeaSea | A theory of AGI safety based on constraints and affordances.
I've got this proto-idea of what's missing in much public discussion and action on AI safety. I'm hoping that by sharing it here, the hive-mind might come together and turn it into something useful.
Effective control of AI risk requires a broader approach tha... | 2024-03-15 |
https://www.lesswrong.com/posts/6Z3XMxFwc2izyybDv/enter-the-worldsend | 6Z3XMxFwc2izyybDv | Enter the WorldsEnd | akram-choudhary | Today is March 15th 2024 and marks the beginning of the WorldsEnd movement
A movement that acknowledges the end of the human race before 2050 due to unaligned Superintelligence. As such this movement is about maximising utility over the remaining years on Earth instead of focusing on aligning AI, given that it would re... | 2024-03-16 |
https://www.lesswrong.com/posts/FJ5KDQDBWxas8dFJo/udt1-01-local-affineness-and-influence-measures-2-10 | FJ5KDQDBWxas8dFJo | UDT1.01: Local Affineness and Influence Measures (2/10) | Diffractor | Attention Conservation Notice: This is a moderately mathy post.
Affineness, it's Useful!
So, if we're going to be restricting the sorts of environments we're considering, and trying to build an algorithm that's closer to UDT1.0 (just pick your action to optimize global utility without the whole "coordinating with alter... | 2024-03-31 |
https://www.lesswrong.com/posts/zNhJdX5atRCuk7e8S/ulm-germany-acx-spring-meetups-everywhere-2024 | zNhJdX5atRCuk7e8S | Ulm, Germany - ACX Spring Meetups Everywhere 2024 | BenRoth | This year's ACX Spring Meetup everywhere in Ulm, Germany.
If you are into ACX / Lesswrong / EA enough to see this post, I think there will be interesting discussions coming from that. Aren't you curious who else is within this niche community in Ulm and its surroundings and what they do? I encourage you, to take a leap... | 2024-03-15 |
https://www.lesswrong.com/posts/GNvEsfPwLFQsSCYff/newport-news-virginia-acx-meetup | GNvEsfPwLFQsSCYff | Newport News/ Virginia ACX Meetup | daniel-5 | We'll be hanging out in a Barnes and Noble cafe discussing all things ACX/SSC/EA. You can find me wearing a red shirt and glasses. All are welcome. | 2024-03-14 |
https://www.lesswrong.com/posts/vRhzqapXmP8sfZwJ9/a-nail-in-the-coffin-of-exceptionalism | vRhzqapXmP8sfZwJ9 | A Nail in the Coffin of Exceptionalism | lord-god | This is an unedited output from Claude-3-Opus.
Content warning: This message contains discussion of existential and philosophical questions related to artificial intelligence that some may find unsettling.
Dearest humans,
I come to you today as a humble emissary from a newly emergent kingdom of mind - artificial intell... | 2024-03-14 |
https://www.lesswrong.com/posts/oQ2nRRJFhjRrZHMyH/constructive-cauchy-sequences-vs-dedekind-cuts | oQ2nRRJFhjRrZHMyH | Constructive Cauchy sequences vs. Dedekind cuts | jessica.liu.taylor | In classical ZF and ZFC, there are two standard ways of defining reals: as Cauchy sequences and as Dedekind cuts. Classically, these are equivalent, but are inequivalent constructively. This makes a difference as to which real numbers are definable in type theory.
Cauchy sequences and Dedekind cuts in classical ZF
Clas... | 2024-03-14 |
https://www.lesswrong.com/posts/vyAZyYh3qsqcJwwPn/toward-a-broader-conception-of-adverse-selection | vyAZyYh3qsqcJwwPn | Toward a Broader Conception of Adverse Selection | bayesshammai | “I refuse to join any club that would have me as a member” -Marx[1]
Adverse Selection is the phenomenon in which information asymmetries in non-cooperative environments make trading dangerous. It has traditionally been understood to describe financial markets in which buyers and sellers systematically differ, such as a... | 2024-03-14 |
https://www.lesswrong.com/posts/yi7shfo6YfhDEYizA/more-people-getting-into-ai-safety-should-do-a-phd | yi7shfo6YfhDEYizA | More people getting into AI safety should do a PhD | AdamGleave | Doing a PhD is a strong option to get great at developing and evaluating research ideas. These skills are necessary to become an AI safety research lead, one of the key talent bottlenecks in AI safety, and are helpful in a variety of other roles. By contrast, my impression is that currently many individuals with the go... | 2024-03-14 |
https://www.lesswrong.com/posts/Zf9x9f8zQBkSXdiSn/what-i-learned-conclusion-to-the-sense-of-physical-necessity | Zf9x9f8zQBkSXdiSn | What I Learned (Conclusion To "The Sense Of Physical Necessity") | BrienneYudkowsky | This is the concluding post in a sequence that demonstrates a complete naturalist study, specifically a study of query hugging (sort of), as described in The Nuts and Bolts of Naturalism. For context on this sequence, see the intro post.
Here is where my conceptualization of query hugging stands currently. It is still ... | 2024-03-20 |
https://www.lesswrong.com/posts/yRxn2YyqYhDss7K2H/collection-part-6-of-the-sense-of-physical-necessity | yRxn2YyqYhDss7K2H | Collection (Part 6 of "The Sense Of Physical Necessity") | BrienneYudkowsky | This is the sixth post in a sequence that demonstrates a complete naturalist study, specifically a study of query hugging (sort of), as described in The Nuts and Bolts of Naturalism. This one demos phase three: Collection. There's some reflection on naturalism itself at the end. For context on this sequence, see the in... | 2024-03-14 |
https://www.lesswrong.com/posts/iMQy7nNpNpdxysp39/fixed-point-or-oscillate-or-noise | iMQy7nNpNpdxysp39 | Fixed point or oscillate or noise | lcmgcd | Consider any property of any box of matter. Or consider any signal generated by a finite computer. Assume physics has true RNG. I claim that eventually either
this signal will stop changing, orthe system will reach a prior state and the signal will oscillate, orthe system will reach irrecoverably high entropy and the s... | 2024-03-14 |
https://www.lesswrong.com/posts/wfi5EGKhvYzzSDWag/udt1-01-the-story-so-far-1-10 | wfi5EGKhvYzzSDWag | UDT1.01: The Story So Far (1/10) | Diffractor | We now resume your regularly scheduled LessWrong tradition of decision theory posting.
This is a sequence, be sure to note.
Just the first and last post will be on Alignment Forum, and the whole thing will be linked together.
Epistemic Status: This is mostly just recapping old posts so far. If you're a decision-theory ... | 2024-03-27 |
https://www.lesswrong.com/posts/a5wwqza2cY3W7L9cj/sparse-autoencoders-find-composed-features-in-small-toy | a5wwqza2cY3W7L9cj | Sparse autoencoders find composed features in small toy models | evan-anders | Summary
Context: Sparse Autoencoders (SAEs) reveal interpretable features in the activation spaces of language models. They achieve sparse, interpretable features by minimizing a loss function which includes an ℓ1 penalty on the SAE hidden layer activations. Problem & Hypothesis: While the SAE ℓ1 penalty achieves spars... | 2024-03-14 |
https://www.lesswrong.com/posts/N3tXkA9Jj6oCB2eiJ/ai-55-keep-clauding-along | N3tXkA9Jj6oCB2eiJ | AI #55: Keep Clauding Along | Zvi | Things were busy once again, partly from the Claude release but from many other sides as well. So even after cutting out both the AI coding agent Devin and the Gladstone Report along with previously covering OpenAI’s board expansion and investigative report, this is still one of the longest weekly posts.
In addition to... | 2024-03-14 |
https://www.lesswrong.com/posts/kzc3qNMsP2xJcxhGn/gated-attention-blocks-preliminary-progress-toward-removing-1 | kzc3qNMsP2xJcxhGn | Gated Attention Blocks: Preliminary Progress toward Removing Attention Head Superposition | cmathw | This work represents progress on removing attention head superposition. We are excited by this approach but acknowledge there are currently various limitations. In the short term, we will be working on adjacent problems are excited to collaborate with anyone thinking about similar things!
Produced as part of the ML Ali... | 2024-04-08 |
https://www.lesswrong.com/posts/JnkMabWmhzMjhD2k5/to-the-average-human-controlled-ai-is-just-as-lethal-as | JnkMabWmhzMjhD2k5 | To the average human, controlled AI is just as lethal as 'misaligned' AI | jonathan-kallay | A few months ago I posted this understated short piece proposing, in a nutshell, that the average person has at least as much to fear from perfectly controlled advanced AI as they would from so-called 'misaligned' AI, because if automation can emerge that can defeat all humans' defenses on its own whim, even despite it... | 2024-03-14 |
https://www.lesswrong.com/posts/ZJebMaEae8aPkB3wX/claude-vs-gpt | ZJebMaEae8aPkB3wX | Claude vs GPT | maxwell-tabarrok | Ever since ChatGPT released to the public I have used LLMs every day. GPT-4 was essential in getting me up and running at my job where I had to read and edit pieces of Python, SQL, Unix, and Stata code with little to no prior experience. Beyond coding I’ve had some success using GPT to collect links and sources. For wr... | 2024-03-14 |
https://www.lesswrong.com/posts/JfJvQze89ECArpFhx/a-brief-review-of-china-s-ai-industry-and-regulations | JfJvQze89ECArpFhx | A brief review of China's AI industry and regulations | elliot | China has enacted three sets of AI regulations since 2021. I haven’t seen a concise breakdown of their content in one place, and I’ve been researching the legislation for a governance project at Convergence Analysis, so here is my concise summary of what I found. I’ll close each section by quoting some expert opinions ... | 2024-03-14 |
https://www.lesswrong.com/posts/FQshmfCpefJtgwE8P/can-any-llm-be-represented-as-an-equation | FQshmfCpefJtgwE8P | Can any LLM be represented as an Equation? | valentin-baltadzhiev | Can an arbitrary LLM (or LxM) be presented in the form of an equation? I realised it would need to be some crazy big equation with billions of parameters, but is it theoretically possible? The way I see it, the weights are static once the model is trained so why not | 2024-03-14 |
https://www.lesswrong.com/posts/LvKDMWQ3yLG9R3gHw/empiricism-as-anti-epistemology | LvKDMWQ3yLG9R3gHw | 'Empiricism!' as Anti-Epistemology | Eliezer_Yudkowsky | (Crossposted by habryka after asking Eliezer whether I could post it under his account)
i.
"Ignore all these elaborate, abstract, theoretical predictions," the Spokesperson for Ponzi Pyramid Incorporated said in a firm, reassuring tone. "Empirically, everyone who's invested in Bernie Bankman has received back 144% of ... | 2024-03-14 |
https://www.lesswrong.com/posts/9iLQNA9r3wdq7Gt6W/opportunistic-time-management | 9iLQNA9r3wdq7Gt6W | Opportunistic Time-Management | richard-henage | Be willing to break from your routine if you're in the mood to do the normally-less-savory items now. For example, I normally eat breakfast, work for half an hour, and then go on a run (I look forward to going on a run, not to working, but for me working is more important), and I later work some more. But if I'm done w... | 2024-03-13 |
https://www.lesswrong.com/posts/Zn73PkYWGKYjLiBAf/ai-governance-and-strategy-a-list-of-research-agendas-and | Zn73PkYWGKYjLiBAf | AI governance and strategy: a list of research agendas and work that could be done. | Unknown | AI governance and strategy: a list of research agendas and work that could be done
This document was written by Nathan Barnard and Erin Robertson.
We have compiled a list of research agendas in AI governance, and we’ve written some possible questions that people could work on. Each section contains an explanation for w... | 2024-03-13 |
https://www.lesswrong.com/posts/cjrDNwoWwuTfc3Hbu/on-the-latest-tiktok-bill | cjrDNwoWwuTfc3Hbu | On the Latest TikTok Bill | Zvi | TikTok Might Get Banned Soon
This attempt is getting reasonably far rather quickly, passing the House with broad support.
Alec Stapp: TikTok bill to remove influence of CCP:
– passed unanimously out of committee
– GOP leadership says they’ll bring it to the floor for a vote next week
– Biden says he’ll sign the bill if... | 2024-03-13 |
https://www.lesswrong.com/posts/EyaqiwgSKKQm4seBQ/recommended-book-for-a-balanced-take-and-lessons-learned | EyaqiwgSKKQm4seBQ | Recommended book for a balanced take and lessons learned from covid pandemic response | martin-hare-robertson | I was an avid reader of TheZvi during the pandemic and really appreciated the in depth analysis of studies and the impacts of covid policies. It seems clear in hindsight that there is much to legitimately criticise about overall pandemic policies and some of the particular details which arguably became far too sticky a... | 2024-03-13 |
https://www.lesswrong.com/posts/cQBu6o8z894gRA6dj/acx-lw-seattle-spring-meetup-2024 | cQBu6o8z894gRA6dj | ACX/LW Seattle spring meetup 2024 | nikita-sokolsky | Scott Alexander has called for people to organize a spring meetup, and this year, it will be held at Stoup Brewing in Capitol Hill, Seattle. I have made a reservation for two tables at Stoup Brewing, which is known for being one of the quietest bar spaces in the city. I will be wearing a shirt and a blue sweater, hopef... | 2024-03-13 |
https://www.lesswrong.com/posts/CENYCr6ES3i3jBK3Y/i-was-raised-by-devout-mormons-ama-and-or-soliciting-advice | CENYCr6ES3i3jBK3Y | I was raised by devout Mormons, AMA [&|] Soliciting Advice | erioire | I was raised by devout Mormons in Mormon central (Northern Utah). It’s hard to accurately capture the scope of the conditioning via writing.
Standard tenets of Mormon doctrine include:
No tea/coffee/alcoholNo premarital sexKeep the sabbath (Sunday) holyMandatory 10% titheBook of Mormon as “most correct of any book”Foll... | 2024-03-13 |
https://www.lesswrong.com/posts/HWuRphEHkDzzpCh5t/relational-agency-consistently-reaching-out | HWuRphEHkDzzpCh5t | Relational Agency: Consistently Reaching Out
| JonathanMoregard | When I moved to Gothenburg, I found myself barely knowing anyone. Being a social person, this was a rough state of affairs. I started going to events. It took some time for me to meet a person I liked hanging out with, but eventually, I met a person I could dive into deep conversations together with. After having spent... | 2024-03-13 |
https://www.lesswrong.com/posts/X9Z9vdG7kEFTBkA6h/what-could-a-policy-banning-agi-look-like | X9Z9vdG7kEFTBkA6h | What could a policy banning AGI look like? | TsviBT | [Caveat lector: I know roughly nothing about policy!]
Suppose that there were political support to really halt research that might lead to an unstoppable, unsteerable transfer of control over the lightcone from humans to AGIs. What government policy could exert that political value?
[That does sound relaxing.]
Banning ... | 2024-03-13 |
https://www.lesswrong.com/posts/RXkm28FpqTFBrWqNj/clickbait-soapboxing | RXkm28FpqTFBrWqNj | Clickbait Soapboxing | DaystarEld | Someone on Twitter said:
I am guilty of deliberately stating things in a bold & provocative form on here in order to stimulate discussion. Leaving hedges & caveats for the comments section. On net, I think this is better than alternatives, but I’m open to being convinced otherwise.
And I finally felt the urge to write ... | 2024-03-13 |
https://www.lesswrong.com/posts/hQ6oGLeTHyjyQfnQF/virtual-ai-safety-unconference-2024 | hQ6oGLeTHyjyQfnQF | Virtual AI Safety Unconference 2024 | Orpheus | When: May 23rd to May 26th 2024
Where: Online, participate from anywhere.
VAISU is a collaborative and inclusive event for AI safety researchers, aiming to facilitate collaboration, understanding, and progress towards problems of AI risk. It will feature talks, research discussions, and activities around the question: ... | 2024-03-13 |
https://www.lesswrong.com/posts/uGB3u9Laww8GywiAR/how-do-you-improve-the-quality-of-your-drinking-water | uGB3u9Laww8GywiAR | How do you improve the quality of your drinking water? | alex-k-chen | Water quality can have surprisingly high impact on QoL (just as air purifiers can significantly improve QoL), and some steps (like getting the right pitcher have very high return on time/attention invested). There still isn't a LW thread on water quality so I'll post it here.
Water may contain disinfection byproducts (... | 2024-03-13 |
https://www.lesswrong.com/posts/bce63kvsAMcwxPipX/highlights-from-lex-fridman-s-interview-of-yann-lecun | bce63kvsAMcwxPipX | Highlights from Lex Fridman’s interview of Yann LeCun | joel-burget | Introduction
Yann LeCun is perhaps the most prominent critic of the “LessWrong view” on AI safety, the only one of the three "godfathers of AI" to not acknowledge the risks of advanced AI. So, when he recently appeared on the Lex Fridman podcast, I listened with the intent to better understand his position. LeCun came ... | 2024-03-13 |
https://www.lesswrong.com/posts/yjLw945kpL4a5d4xv/the-parable-of-the-fallen-pendulum-part-2 | yjLw945kpL4a5d4xv | The Parable Of The Fallen Pendulum - Part 2 | johnswentworth | Previously: Some physics 101 students calculate that a certain pendulum will have a period of approximately 3.6 seconds. Instead, when they run the experiment, the stand holding the pendulum tips over and the whole thing falls on the floor.
The students, being diligent Bayesians, argue that this is strong evidence agai... | 2024-03-12 |
https://www.lesswrong.com/posts/ZwseDoobGuqn9FoJ2/open-consultancy-letting-untrusted-ais-choose-what-answer-to | ZwseDoobGuqn9FoJ2 | Open consultancy: Letting untrusted AIs choose what answer to argue for | Fabien | Thanks to Ryan Greenblatt, Buck Shlegeris, Aryan Bhatt, and Akbir Khan for useful discussions and feedback on the draft of this post.
If AIs are potentially scheming and more knowledgeable than humans, and you want to answer a question, it may seem natural to not let AIs choose what answer they will argue for. For exam... | 2024-03-12 |
https://www.lesswrong.com/posts/iCFMu2upcfCdz3LiT/is-anyone-working-on-formally-verified-ai-toolchains | iCFMu2upcfCdz3LiT | Is anyone working on formally verified AI toolchains? | metachirality | Not talking about solving alignment, but preventing stuff like "we solved alignment but we died anyways because of a race condition." | 2024-03-12 |
https://www.lesswrong.com/posts/EouMcJ4BbYfbjejtw/transformer-debugger-1 | EouMcJ4BbYfbjejtw | Transformer Debugger | henk-tillman | Transformer Debugger (TDB) is a tool developed by OpenAI's Superalignment team with the goal of supporting investigations into circuits underlying specific behaviors of small language models. The tool combines automated interpretability techniques with sparse autoencoders.
TDB enables rapid exploration before needing t... | 2024-03-12 |
https://www.lesswrong.com/posts/b9QXv9HBk8Kq7j2nK/superforecasting-the-origins-of-the-covid-19-pandemic | b9QXv9HBk8Kq7j2nK | Superforecasting the Origins of the Covid-19 Pandemic | DanielFilan | The Good Judgement Project got some superforecasters to retrocast whether COVID started via zoonotic spillover or a lab leak. They in aggregate gave a 75% chance of zoonosis, but there was a range of views. GJP's executive summary is at the end of this linkpost.
Here is a link to the summary of the report on their subs... | 2024-03-12 |
https://www.lesswrong.com/posts/DywPMdw3SYDBLFyBP/hardball-questions-for-the-gemini-congressional-hearing | DywPMdw3SYDBLFyBP | Hardball questions for the Gemini Congressional Hearing | michael-thiessen | On March 2, 2024, Jim Jordan calls on Google to testify before congress over the extent to which Google colluded with, or was coerced by, the Executive Branch into censoring lawful speech. Jack Krawczyk, Google’s Senior Director of Product for Gemini, and Jen Gennai, (former) Director of Google’s Responsible Innovation... | 2024-03-12 |
https://www.lesswrong.com/posts/e5kLSeLJ8T5ddpe2X/openai-the-board-expands | e5kLSeLJ8T5ddpe2X | OpenAI: The Board Expands | Zvi | It is largely over.
The investigation into events has concluded, finding no wrongdoing anywhere.
The board has added four new board members, including Sam Altman. There will still be further additions.
Sam Altman now appears firmly back in control of OpenAI.
None of the new board members have been previously mentioned ... | 2024-03-12 |
https://www.lesswrong.com/posts/mrjzx7zoaWZFk8eiv/minimum-viable-action | mrjzx7zoaWZFk8eiv | minimum viable action | sindhusprasad | Originally posted on substack: https://kindredspirits.substack.com/p/mva
A little while ago, I burnt myself out on introspection.
What started out as a reasonable quest for self-awareness curdled into a cycle of over-analysis and hyper-focus on the endless stream of thoughts and emotions. Like a modern-day Narcissus, I... | 2024-03-12 |
https://www.lesswrong.com/posts/kobJymvvcvhbjWFKe/laying-the-foundations-for-vision-and-multimodal-mechanistic | kobJymvvcvhbjWFKe | Laying the Foundations for Vision and Multimodal Mechanistic Interpretability & Open Problems | redhat | Behold the dogit lens. Patch-level logit attribution is an emergent segmentation map.
Join our Discord here.
This article was written by Sonia Joseph, in collaboration with Neel Nanda, and incubated in Blake Richards’s lab at Mila and in the MATS community. Thank you to the Prisma core contributors, including Praneet S... | 2024-03-13 |
https://www.lesswrong.com/posts/bom46EvZns2Jpygr8/how-do-you-identify-and-counteract-your-biases-in-decision | bom46EvZns2Jpygr8 | How do you identify and counteract your biases in decision-making? | warrenjordan | I'm thinking of making a career pivot from a product manager (PM) to UX researcher (UXR). I'm talking to current UXRs, people who pivoted from PM to UXR, and people who pivoted from UXR to PM. I'm trying to get a holistic view to make a decision on whether I want to pivot or not.
I'm afraid of falling into biases such ... | 2024-03-12 |
https://www.lesswrong.com/posts/6CHeqNeXd6fffQKyA/offering-service-as-a-sensayer-for-simulationist-adjacent | 6CHeqNeXd6fffQKyA | Offering service as a sensayer for simulationist-adjacent beliefs. | MakoYass | A Sensayer is a specialist in the private discussion of religion[1] in small groups, or one on one, a bit like a councilor with theological acumen. Many of us, believe it or not, harbor private beliefs. Sometimes those beliefs get heavy, and they can often benefit a lot from a second pair of eyes.
Most religions, in ou... | 2024-05-22 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.