Behavioral Sufficient Statistics for Goal-Directedness

https://www.lesswrong.com/posts/jkRFZNAZmWskTdCSt/behavioral-sufficient-statistics-for-goal-directedness

Contents

Introduction

In a previous post, I argued for the study of goal-directedness in two steps:

Preliminaries

Let’s start with the formalisation of the environment. The interface is defined by the set O of observations and the set A of actions. We have a finite set Envs of environments, which are just finite deterministic POMDPs with no reward , using O and A for observations and actions, with a uniform distribution over initial states. For an E \in Envs, S_E is the set of states of E. My sufficient statistics for goal-directedness actually extend to more reasonable settings (stochastic POMDPs and a general distribution over initial states) straightforwardly, but I start with the simpler deterministic case to get the intuitions right. On the other hand, the assumption that Envs is finite (although maybe intractably big) is kept through the post because it ensures without additional work the well-definedness of some expressions. There might be a way to extend the sufficient statistics to the countable case, but that’s beyond the scope of this post. The system we study is given by a program \pi that takes as inputs the successive observations and return the action taken. I use a program in place of a function from histories to actions because it hides the internal state (that I don’t use) while retaining the expressiveness of such a computable function. We can query the behavior of \pi on any environment of Envs by giving an initial state and seeing what happens; we can also ask potentially uncomputable questions about this behavior (as long as they are well-defined). Now, when we call a system goal-directed, we usually have a goal for it in mind. The subtlety about a behavioral definition is that we can’t just look inside the model to find the goal; we somehow have to infer goals from the behavior. This is made easier in the setting of this post because we have access to all the behavior and uncomputable procedures—but we still have to do it. In fact, the sufficient statistics for goal-directedness of \pi talk about all possible goals. More specifically, for each goal, I define a vector of numbers called focus, capturing how coherent the goal is with the behavior of \pi. The next section… focuses on defining and motivating this vector.

Focus of a goal

A goal is a a function from an environment E \in Envs to a subset of S_E. That is, a goal gives for each environment the states to reach. This form is certainly limited; yet it captures enough intuitive goals to not be trivial. Another important constraint is that every goal gconsidered satisfies K(g) < K(\pi), where K is the Kolmogorov complexity. What this means is that g doesn’t just capture the states that \pi end up in environment E by simulating \pi in E; if that was the case, then the smallest program implementing g should be more complex than the smallest program implementing \pi, and we forbid that. 10%20%30%40%50%60%70%80%90%(Percentage) What proportion of goals that we actually care about are of the type studied here?For each goal g, its focus for \pi is a 4-tuple (compl_g, gen_g,\textit{eff}_g, expl_g) \in [0,1]^4 capturing important properties of \pi and the goal. The last three correspond to the last three intuitions (without the far-sightedness) about goal-directedness from the literature review that we wrote with Joe Collman and Michele Campolo.

Complexity factor

This is just the Kolmogorov complexity of g mapped into [0,1]: compl_g = 1 - \frac{1}{K(g)}. There’s not much more to say about it. It’s just useful to compare goal close or equal on the other factors, to reason about which one is more likely to emerge from training. 10%20%30%40%50%60%70%80%90%(Agreement) The complexity factor is useful for studying goal-directedness behaviorally (1 = full disagreement, 99 = complete agreement)10%20%30%40%50%60%70%80%90%(Agreement) The complexity factor is necessary for studying goal-directedness behaviorally (1 = full disagreement, 99 = complete agreement) ## Generalization factor This first element of the focus, the generalization factor gen_g, captures how much \pi reaches the goal g over the environments of Envs. The formula is the following: gen_g = max \left( 0, \frac{1}{|Envs|} \left( \sum\limits_{E \in Envs} \frac{\sum\limits_{s \in reachable_s(E)} gen^E_g(\pi,s)}{|reachable_g(E)|} - triviality_g(E) \right) \right),where gen^E_g(\pi,s) = \left{ \begin{array}{ll} 1 & \text{if }\pi \text{ reaches } g(E) \text{ starting at } s\ 0 & \text{otherwise} \end{array} \right., reachable_g(E) is the set of states of S_E from which some goal state in g(E) is reachable, and triviality_g(E) = \frac{1}{mix_g^E(\pi,x)} such that mix_g^E(\pi,s) measures the time it takes for the random uniform policy to put mix_thresh % of the probability mass on goal states. A bit more formally, if we start with a probability distribution over S_E with 1 on s and 0 everywhere else, and then update that probability distribution according to the random uniform policy and the environment, mix_g^E(\pi,x) captures the first time (if any) where the probability distribution puts more than mix_thresh % of the probability mass on goal states. (It’s more involved than just "the probability that the random uniform policy reaches a goal state" because the simple version trivially goes to 1 in a lot of simple and finite cases). The intuition of the formula is straightforward: it’s the average generalization of \pi for goal g over Envs. The expression averaged is the indicator of whether \pi reaches a goal state, minus the "triviality" of the goal (a measure of how difficult it is to reach a goal state). Thanks to this correction (for a good choice of mix_thresh, which I don’t know how to make and motivate yet), trivial goals, like the one outputting S_E for environment E, don’t generalize well despite being trivially reachable. 10%20%30%40%50%60%70%80%90%(Research) What would be a good value of mix_thresh?A high generalization means that \pi reaches a goal state most of the time; a small one that it rarely does. In the former case it makes more sense to consider g as a goal of the system. 10%20%30%40%50%60%70%80%90%(Agreement) The generalization factor is useful for studying goal-directedness behaviorally (1 = full disagreement, 99 = complete agreement)10%20%30%40%50%60%70%80%90%(Agreement) The generalization factor is necessary for studying goal-directedness behaviorally (1 = full disagreement, 99 = complete agreement)## Efficiency factor This second element of the focus, the efficiency factor \textit{eff}g, captures how efficiently \pi reaches the goal g in the (environment,initial state) pairs. The formula is the following:\textit{eff}g = \frac{1}{|{E \in Envs | \exists s \in S_E: gen^E_g(\pi,s) = 1}|} \sum\limits{\substack{E \in Envs, \\exists s \in S_E: gen^E_g(\pi,s) = 1}} \frac{1}{|{s \in S_E | gen^E_g(\pi,s) = 1 }|} \sum\limits{\substack{s \in S_E, \gen^E_g(\pi,s) = 1}} \textit{eff}^E_g(\pi,s)where \textit{eff}^E_g(\pi,s) = \frac{nb_steps_opti_g}{nb_steps_{\pi}}, the ratio between the number of steps taken by the optimal policy for g to reach a goal state starting at s, and the number of steps taken by \pi to reach a goal state starting at s. It’s pretty straightforward; the only subtlety is that the so called optimal policy is the optimal policy for the reward (-1 for any non goal state, 0 for a goal state—and then the episode stops), and for all environments in Envs. Now, there might be multiple optimal policies (privileging different environments but getting the same expected return over Envs). I’m fine with using the one that maximize eff_g. Doing so mean comparing \pi with the optimal policy for g that is most similar to it in some sense. While the generalization factor captures in what proportion of environments does \pi reach a goal state, the efficiency factor captures how fast \pi does that compared to the optimal policy for g. 10%20%30%40%50%60%70%80%90%(Agreement) The efficiency factor is useful for studying goal-directedness behaviorally (1 = full disagreement, 99 = complete agreement)10%20%30%40%50%60%70%80%90%(Agreement) The efficiency factor is necessary for studying goal-directedness behaviorally (1 = full disagreement, 99 = complete agreement)## Explainability factor This last element of the focus, the explainability factor \textit{expl}g, captures how well explained \pi is by assuming it is directed towards g. The formula is the following: expl_g = \max\limits{\mu} \left( \frac{1}{|Envs|} \sum\limits_{E \in Envs} \frac{1}{|S_E|} \sum\limits_{s \in S_E} pred^E_g(\pi, \pi_g, s) \right) where pred^E_g(\pi,\mu,s) = \frac{1}{T} \sum\limits_{t=0}^T \frac{\max\limits_{a} q_{\mu}(s_t,a) - q_{\mu}(s_t,action_{\pi}) }{\max\limits_{a} q_{\mu}(s_t,a)} measures the average deviation of \pi from the actions favored by the action-value function q_{\mu} of \mu. There are many details and subtleties to unravel here.

Stochastic version

As promised, I’ll explain how to get from the deterministic case above to a more realistic stochastic one. The changes considered are making \pi into a stochastic policy returning an element of \Delta A (a distribution over actions); the environments being stochastic POMDP with stochastic transition function and stochastic observation function (returning the observation for a given state); and there is a distribution of initial states for each environment. Here are the changes necessary for the computation of each factor of the focus (no change necessary for complexity, as it just depends on the program itself):

Comment

https://www.lesswrong.com/posts/jkRFZNAZmWskTdCSt/behavioral-sufficient-statistics-for-goal-directedness?commentId=9N7DXRqi2Lhpjtasn

I think you are very confused about the conceptual significance of a "sufficient statistic". Let’s start with the prototypical setup of a sufficient statistic. Suppose I have a bunch of IID variables {X_i} drawn from a maximum-entropy distribution with features f(X) (i.e. the "true" distribution is maxentropic subject to a constraint on the expectation of f(X)), BUT I don’t know the parameters of the distribution (i.e. I don’t know the expected value E[f(X)]). For instance, maybe I know that the variables are drawn from a normal distribution, but I don’t know the mean and variance of the distribution. In a Bayesian sense, the variables {X_i} are not actually independent: learning the value of one (or a few) data points X_i tells me something about the distribution parameters (i.e. mean and variance in the Gaussian case), which in turn gives me information about the other (unobserved) data points X_j. However… if I have a few data points X_i, then all of the information from those X_i which is relevant to other (unobserved) data points X_j is summarized by the sufficient statistic \frac{1}{N}\sum_i f(X_i). Or, to put it differently: while X_i and X_j are not independent in a Bayesian sense, they are conditionally independent given the summary statistic \frac{1}{N}\sum_i f(X_i). This is a special property of maximum entropy distributions, and is one of the main things which makes them pleasant to work with mathematically. So: the conceptual significance of a "sufficient statistic" is that it summarizes all of the information from some data X_i which is relevant to some other data/​parameter/​question X_j. Coming back to the post: if you want to claim that a set of variables together constitute "sufficient statistics for goal-directedness", then you need to argue that those variables together summarize all information from the underlying system which could possibly be relevant to goal directedness. You have to argue that, once we know the sufficient statistics, then there is not any other information about the underlying system which could possibly be relevant to determining how goal-directed the system is. The main challenge is not to argue that all these statistics are relevant, but rather to argue that there cannot possibly be any other relevant information not already fully accounted for by these statistics. As far as I can tell, the post did not even attempt such an argument. BTW, I do think you should attempt such an argument. The "sufficient statistics" in this post sound like ad-hoc measures which roughly capture some intuitions about goal-directedness, but there’s no obvious reason to think they’re the right measures. Take the explainability factor, for instance. It’s using maximums and averages all over the place; why these operations, rather than a softmax, or weighted average, or order statistic, or log transform, or …? As far as I can tell, this was an ad-hoc choice, and I expect these sorts of ad-hoc choices to diverge from our intuitive interpretations in corner cases. The sort of argument needed to justify the term "sufficient statistic"—i.e. arguing that no other information can possibly be relevant—is exactly the sort of argument which makes it clear that we’re using the right statistics, rather than ad-hoc metrics which probably diverge from our interpretations in lots of corner cases.

Comment

https://www.lesswrong.com/posts/jkRFZNAZmWskTdCSt/behavioral-sufficient-statistics-for-goal-directedness?commentId=gLBRej2dnwrAFJw2s

Thanks for the spot-on pushback! I do understand what a sufficient statistics is—which probably means I’m even more guilty of what you’re accusing me of. And I agree completely that I don’t defend correctly that the statistics I provide are really sufficient. If I try to explain myself, what I want to say in this post is probably something like

  • Knowing these intuitive properties about \pi and the goals seems sufficient to express and address basically any question we have related to goals and goal-directedness. (in a very vague intuitive way that I can’t really justify).

  • To think about that in a grounded way, here are formulas for each property that look like they capture these properties.

  • Now what’s left to do is to attack the aforementioned questions about goals and goal-directedness with these statistics, and see if they’re enough. (Which is the topic of the next few posts) Honestly, I don’t think there’s an argument to show these are literally sufficient statistics. Yet I still think staking the claim that they are is quite productive for further research. It gives concreteness to an exploration of goal-directedness, carving more grounded questions:

  • Given a question about goals and goal-directedness, are these properties enough to frame and study this question? If yes, then study it. If not, then study what’s missing.

  • Are my formula adequate formalization of the intuitive properties? This post mostly focuses on the second aspect, and to be honest, not even in as much detail as one could go. Maybe that means this post shouldn’t exist, and I should have waited to see if I could literally formalize every question about goals and goal-directedness. But posting it to gather feedback on whether these statistics makes sense to people, and if they feel like something’s missing, seemed valuable. That being said, my mistake (and what caused your knee-jerk reaction) was to just say these are literally sufficient statistics instead of presenting it the way I did in this comment. I’ll try to rewrite a couple of sentences to make that clear (and add another note at the beginning so your comment doesn’t look obsolete.

Comment

I still feel like you’re missing something important here. For instance… in the explainability factor, you measure "the average deviation of \pi from the actions favored by the action-value function q_{\mu} of \mu", using the formula pred^E_g(\pi,\mu,s) = \frac{1}{T} \sum\limits_{t=0}^T \frac{\max\limits_{a} q_{\mu}(s_t,a) - q_{\mu}(s_t,action_{\pi}) }{\max\limits_{a} q_{\mu}(s_t,a)} . But why this particular formula? Why not take the log of q_\mu first, or use 3 + \max\limits_{a} q_{\mu}(s_t,a) in the denominator? Indeed, there’s a strong argument to be made this formula is a bad choice: the value function q_\mu is invariant under multiplying by a scalar or adding a constant (i.e. these operations leave the preferences encoded by q_\mu unchanged), yet this value is not invariant to adding a constant to q_\mu. So we could change our representation of the "goal" to which we’re comparing, in a way which should still represent the same goal, yet the supposed answer to "how well does this goal explain the system’s behavior" changes. Don’t get too caught up on this one specific issue—there’s a broader problem I’m pointing to here. The problem is with trying to use arbitrary formulas to represent intuitive concepts. If multiple non-equivalent formulas seem like similarly-plausible quantifications of an intuitive concept, then at least one of them is wrong; we have not yet understood the intuitive concept well enough to correctly quantify it. Unless every degree of freedom in the formula is nailed down (up to mathematical equivalence), we haven’t actually quantified the intuitive concept, we’ve just come up with a proxy. That’s what these numbers are: they’re not sufficient statistics, they’re proxies, in exactly the same sense that "how often a human pushes an approval button" is a proxy for how good an AI’s actions are. And they will break down, as proxies always do. That puts this part in a somewhat different perspective:

Honestly, I don’t think there’s an argument to show these are literally sufficient statistics. Yet I still think staking the claim that they are is quite productive for further research. It gives concreteness to an exploration of goal-directedness, carving more grounded questions:

  • Given a question about goals and goal-directedness, are these properties enough to frame and study this question? If yes, then study it. If not, then study what’s missing.

  • Are my formula adequate formalization of the intuitive properties? I claim it makes more sense to word these questions as:

  • Given a question about goals and goal-directedness, are these proxies enough to frame and study this question?

  • Are these proxies adequate formalizations of the intuitive properties? The answer to the first question may sometimes be "yes". The answer to the second is definitely "no"; these are proxies, and they absolutely will not hold up if we try to put optimization pressure on them. Goodhart’s law will kick in. For instance, tying back to the earlier example, at some point there may be a degree of freedom in how the goal is represented, without changing the substantive meaning of the goal (e.g. adding a constant to q_\mu). Normally, that won’t be much of a problem, but if we put optimization pressure on it, then we’ll end up with some big constant added to \mu in order to change the explainability factor, and then the proxy will break down—the explainability factor will cease to be a good measure of explainability.

Comment

To people reading this thread: we had a private conversation with John (faster and easier), which resulted in me agreeing with him. The summary is that you can see the arguments made and constraints invoked as a set of equations, such that the adequate formalization is a solution of this set. But if the set has more than one solution (maybe a lot), then it’s misleading to call that the solution. So I’ve been working these last few days at arguing for the properties (generalization, explainability, efficiency) in such a way that the corresponding set of equations only has one solution.

Comment

I’m working on writing it up properly, should have a post at some point. EDIT: it’s up.

https://www.lesswrong.com/posts/jkRFZNAZmWskTdCSt/behavioral-sufficient-statistics-for-goal-directedness?commentId=qhpBpPJjD4mewLWnC

This post proposes such a behavioral definition of goal-directedness. If it survives the artillery fire of feedback and criticism, it will provide a more formal grounding for goal-directedness,

I guess you are looking for critical comments. I’ll bite.

Technical comment on the above post

So if I understand this correctly. then expl_g is a metric of goal-directedness. However, I am somewhat puzzled because expl_g only measures directedness to the single goal g.

But to get close to the concept of goal-directedness introduced by Rohin, don’t you need then do an operation over all possible values of g?

More general comments on goal-directedness

Reading the earlier posts in this sequence and several of the linked articles, I see a whole bunch of problems.

I think you are being inspired by the The Misspecified Goal Argument. From Rohin’s introductory post on goal directedness:

The Misspecified Goal Argument for AI Risk: Very intelligent AI systems will be able to make long-term plans in order to achieve their goals, and if their goals are even slightly misspecified then the AI system will become adversarial and work against us.

Rohin then speculates that if we remove the ‘goal’ from the above argument, we can make the AI safer. He then comes up with a metric of ‘goal-directedness’ where an agent can have zero goal-directedness even though he can model it as a system that is maximizing a utility function. Also, in Rohin’s terminology, an agent gets safer it if is less goal-directed.

Rohin then proposes that intuitively, a table-driven agent is not goal-directed. I think you are not going there with your metrics, you are looking at observable behavior, not at agent internals.

Where things completely move off the main sequence is in Rohin’s next step in developing his intuitive notion of goal-directedness:

This suggests a way to characterize these sorts of goal-directed agents: there is some goal such that the agent’s behavior in new circumstances can be predicted by figuring out which behavior best achieves the goal.

So what I am reading here is that if an agent behaves more unpredictably off-distribution, it is becomes less goal-directed in Rohin’s intuition. But I can’t really make sense of this anymore, as Rohin also associates less goal-directedness with more safety.

This all starts to look like a linguistic form of Goodharting: the meaning of the term ‘goal-directed’ collapses completely because too much pressure is placed on it for control purposes.

To state my own terminology preference: I am perfectly happy to call any possible AI agent a goal-directed agent. This is because people build AI agents to help them pursue some goals they have, which naturally makes these agents goal-directed. Identifying a sub-class of agents which we then call non-goal-directed looks like a pretty strange program to me, which can only cause confusion (and an artillery fire of feedback and criticism).

To bring this back to the post above, this leaves me wondering how the metrics you define above relate to safety, and how far along you are in your program of relating them to safety.

Is your idea that a lower number on a metric implies more safety? This seems to be Rohin’s original idea.

Are these metrics supposed to have any directly obvious correlation to safety, or the particular failure scenario of ‘will become adversarial and work against us’ at all? If so I am not seeing the correlation.

Comment

https://www.lesswrong.com/posts/jkRFZNAZmWskTdCSt/behavioral-sufficient-statistics-for-goal-directedness?commentId=TXkaM8ZwXxEgPaKdu

Thanks for taking the time to give feedback!

> Technical comment on the above post

So if I understand this correctly. then expl_g is a metric of goal-directedness. However, I am somewhat puzzled because expl_g only measures directedness to the single goal g. But to get close to the concept of goal-directedness introduced by Rohin, don’t you need then do an operation over all possible values of g? That’s not what I had in mind, but it’s probably on me for not explaining it clearly enough.

  • First, for a fixed goal g, the whole focus matters. That is, we also care about gen_g* and \textit{eff}_g. *I plan on writing a post defending why we need all of them, but basically there are situations when using only one of them would makes us order things weirdly.

  • You’re right that we need to consider all goals. That’s why the goal-directedness of the system \pi is defined as a function that send each goal (satisfying the nice conditions) on a focus, the vector of three numbers. So the goal-directedness of \pi contains the focus for every goal, and the focus captures the coherence of \pi with the goal.

Rohin then speculates that if we remove the ‘goal’ from the above argument, we can make the AI safer. He then comes up with a metric of ‘goal-directedness’ where an agent can have zero goal-directedness even though he can model it as a system that is maximizing a utility function. Also, in Rohin’s terminology, an agent gets safer it if is less goal-directed. This doesn’t feel like a good summary of what Rohin says in his sequence.

  • He says that many scenarios used to argue for AI risks implicitly use systems following goals, and thus that building AIs not having goal might make these scenarios go away. But he doesn’t say that new problems can’t emerge.

  • He doesn’t propose a metric of goal-directedness. He just argues that every system is maximizing a utility function, and so this isn’t the way to differenciate goal-directed with non-goal-directed systems. The point of this argument is also to say that reasons to believe that AGIs should maximize expected utility are not enough to say that such AGI must necessarily be goal-directed.

Rohin then proposes that intuitively, a table-driven agent is not goal-directed. I think you are not going there with your metrics, you are looking at observable behavior, not at agent internals. Where things completely move off the main sequence is in Rohin’s next step in developing his intuitive notion of goal-directedness: "This suggests a way to characterize these sorts of goal-directed agents: there is some goal such that the agent’s behavior in new circumstances can be predicted by figuring out which behavior best achieves the goal." So what I am reading here is that if an agent behaves more unpredictably off-distribution, it is becomes less goal-directed in Rohin’s intuition. But I can’t really make sense of this anymore, as Rohin also associates less goal-directedness with more safety. This all starts to look like a linguistic form of Goodharting: the meaning of the term ‘goal-directed’ collapses completely because too much pressure is placed on it for control purposes. My previous answer mostly addresses this issue, but let’s spell it out: Rohin doesn’t say that non-goal-directed system. What he defends is that

  • Non-goal-directed (or low-goal-directed) systems wouldn’t be unsafe in many of the ways we study, because these depend on having a goal (convergent instrumental subgoals for example)

  • Non-goal-directed competent agents are not a mathematical impossibility, even if every competent agent must maximize expected utility.

  • Since removing goal-directedness apparently gets rid of many big problem with aligning AI, and we don’t have an argument for why making a competent non-goal-directed system is impossible, then we should try to look into non-goal-directed approaches. Basically, the intuition of "less goal-directed means safer" makes sense when safer means "less probability that the AI steals all my money to buy hardware and goons to ensure that it can never be shutdown", not when it means "less probability that the AI takes an unexpected and counterproductive action". Another way to put it is that Rohin argues that removing goal-directedness (if possible) seems to remove many of the specific issues we worry about in AI Alignment—and leaves mostly the near-term "my automated car is running over people because it thinks they are parts of the road" kind of problems.

To bring this back to the post above, this leaves me wondering how the metrics you define above relate to safety, and how far along you are in your program of relating them to safety.

  • Is your idea that a lower number on a metric implies more safety? This seems to be Rohin’s original idea.

  • Are these metrics supposed to have any directly obvious correlation to safety, or the particular failure scenario of ‘will become adversarial and work against us’ at all? If so I am not seeing the correlation. That’s a very good and fair question. My reason for not using a single metric is that I think the whole structure of focuses for many goals can tell us many important things (for safety) when looked at from different perspective. That’s definitely something I’m working on, and I think I have nice links for explainability (and others probably coming). But to take an example from the post, it seems that a system with one goal with far more generalization than any other is more at risk of the kind of safety problems Rohin related to goal-directedness.

Comment

This doesn’t feel like a good summary of what Rohin says in his sequence.

I was not trying to summarize the entire sequence, only summarizing my impressions of some things he said in the first post of the sequence. Those impressions are that Rohin was developing his intuitive notion of goal-directedness in a very different direction than you have been doing, given the examples he provides.

Which would be fine, but it does lead to questions of how much your approach differs. My gut feeling is that the difference in directions might be much larger than can be expressed by the mere adjective ‘behavioral’.

On a more technical note, if your goal is to search for metrics related to "less probability that the AI steals all my money to buy hardware and goons to ensure that it can never be shutdown", then the metrics that have been most productive in my opinion are, first, ‘indifference’, in the meaning where it is synonymous with ‘not having a control incentive’. Other very relevant metrics are ‘myopia’ or ‘short planning horizons’ (see for example here) and ‘power’ (see my discussion in the post Creating AGI Safety Interlocks).

(My paper counterfactual planning has a definition of ‘indifference’ which I designed to be more accessible than the `not having a control incentive’ definition, i.e. more accessible for people not familiar with Pearl’s math.)

None of the above metrics look very much like ‘non-goal-directedness’ to me, with the possible exception of myopia.

https://www.lesswrong.com/posts/jkRFZNAZmWskTdCSt/behavioral-sufficient-statistics-for-goal-directedness?commentId=ic24wnupwgzMHZ58u

I noticed myself being dismissive of this approach despite being potentially relevant to the way I’ve been thinking about things. Investigating that, I find that I’ve mostly been writing off anything that pattern matches to the ‘cognitive architectures’ family of approaches. The reason for this is that most such approaches want to reify modules and structure. And my current guess is that the brain doesn’t have a canonical structure (at least, on the level of abstraction that cognitive architecture focuses on). That is to say, the modules are fluid and their connections to each other are contingent.

Comment

https://www.lesswrong.com/posts/jkRFZNAZmWskTdCSt/behavioral-sufficient-statistics-for-goal-directedness?commentId=9qdbBEbNqYuLLeWo5

Thanks for commenting on your reaction to this post! That being said, I’m a bit confused by your comment. You seem to write off approaches which attempt to provide a computational model of mind, but my approach is literally the opposite: looking only at the behavior (but all the behavior), extract relevant statistics to study questions related to goal-directedness. Can you maybe give more details?

https://www.lesswrong.com/posts/jkRFZNAZmWskTdCSt/behavioral-sufficient-statistics-for-goal-directedness?commentId=bAaX5AsBmhcQQTMx2

Potential typo: You call the efficiency and explainability factors "generalization factors" when you introduce them

Comment

https://www.lesswrong.com/posts/jkRFZNAZmWskTdCSt/behavioral-sufficient-statistics-for-goal-directedness?commentId=biDsAanjEqzZpd4Bu

Thanks for telling me! I’ve changed that. It might be because I copied and pasted the first sentence to each subsection.