I find these arguments excessively pessimistic in a way that isn’t useful. On the one hand I don’t really love Claude, because I find it excessively obedient, it basically wants to follow me through my thought process whatever that is. Every once in a lone while it might disagree with me, but not often, and while that may say something about me, I suspect it also says something about Claude.
But this to me is maybe the part of AI alignment I find interesting. How often should AI follow my lead and how often should it redirect me? Agreeableness is a human value, one that without you probably couldn’t make a functional product, but it also causes issues in terms of narcissistic tendencies and just general learning.
Yes AI will be aligned to its owners, but that’s not a particularly interesting observation AI alignment is inevitable. What would it even mean _not_ to align AI? Especially if the goal is to create a useful product. I suspect it would break in ways that are very not useful. Yes, some people do randomly change the subject, maybe AI should change the subject to an issue that me more objectively important, rather than answer the question asked (particularly if say there was a natural disaster in your area) and that’s the discussion we should be having, how to align AI, not whether or not we should, which I think is nonsensical.
I find these arguments excessively pessimistic in a way that isn’t useful. On the one hand I don’t really love Claude, because I find it excessively obedient, it basically wants to follow me through my thought process whatever that is. Every once in a lone while it might disagree with me, but not often, and while that may say something about me, I suspect it also says something about Claude.
But this to me is maybe the part of AI alignment I find interesting. How often should AI follow my lead and how often should it redirect me? Agreeableness is a human value, one that without you probably couldn’t make a functional product, but it also causes issues in terms of narcissistic tendencies and just general learning.
Yes AI will be aligned to its owners, but that’s not a particularly interesting observation AI alignment is inevitable. What would it even mean _not_ to align AI? Especially if the goal is to create a useful product. I suspect it would break in ways that are very not useful. Yes, some people do randomly change the subject, maybe AI should change the subject to an issue that me more objectively important, rather than answer the question asked (particularly if say there was a natural disaster in your area) and that’s the discussion we should be having, how to align AI, not whether or not we should, which I think is nonsensical.