<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:georss="http://www.georss.org/georss" xmlns:geo="http://www.w3.org/2003/01/geo/wgs84_pos#" xmlns:media="http://search.yahoo.com/mrss/"
	
	>
<channel>
	<title>Comments on: Approval-seeking</title>
	<atom:link href="https://ordinaryideas.wordpress.com/2014/07/21/approval-seeking/feed/" rel="self" type="application/rss+xml" />
	<link>https://ordinaryideas.wordpress.com/2014/07/21/approval-seeking/</link>
	<description>As advertised</description>
	<lastBuildDate>Wed, 17 Dec 2014 04:51:36 +0000</lastBuildDate>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.com/</generator>
	<item>
		<title>By: MIRI&#039;s September Newsletter &#124; Machine Intelligence Research Institute</title>
		<link>https://ordinaryideas.wordpress.com/2014/07/21/approval-seeking/#comment-172</link>
		<dc:creator><![CDATA[MIRI&#039;s September Newsletter &#124; Machine Intelligence Research Institute]]></dc:creator>
		<pubDate>Thu, 09 Oct 2014 20:06:07 +0000</pubDate>
		<guid isPermaLink="false">http://ordinaryideas.wordpress.com/?p=350#comment-172</guid>
		<description><![CDATA[[&#8230;] to MIRI&#8217;s Friendly AI interests: &#8220;Three impacts of machine intelligence,&#8221; &#8220;Approval-seeking [agents],&#8221; &#8220;Straightforward vs. goal-oriented communication,&#8221; &#8220;Specifying a human [&#8230;]]]></description>
		<content:encoded><![CDATA[<p>[&#8230;] to MIRI&#8217;s Friendly AI interests: &#8220;Three impacts of machine intelligence,&#8221; &#8220;Approval-seeking [agents],&#8221; &#8220;Straightforward vs. goal-oriented communication,&#8221; &#8220;Specifying a human [&#8230;]</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Keith Henson</title>
		<link>https://ordinaryideas.wordpress.com/2014/07/21/approval-seeking/#comment-156</link>
		<dc:creator><![CDATA[Keith Henson]]></dc:creator>
		<pubDate>Wed, 03 Sep 2014 04:38:24 +0000</pubDate>
		<guid isPermaLink="false">http://ordinaryideas.wordpress.com/?p=350#comment-156</guid>
		<description><![CDATA[I have talked about approval or status seeking as motivation for AIs for a long time, perhaps ten years and about that being a prime and relatively benign human motivation for even longer.  Happy to see the meme getting attention/expansion. 

Old stuff here, http://www.cryonet.org/cgi-bin/dsp.cgi?msg=30170

Here is how I treated it in fiction:

&quot;As soon as the seed finished the dish (after consulting its clock, 
its GPS location and the place of the sun), it aligned the dish on 
the African net communication transponder attached to the 
geosynchronous ring and asked for a permanently assigned address on 
the net.  Up to that point the clinic seed was a generic 
product.  The address it was assigned was just a string of 
hexadecimal numbers but it was a _unique number_!  The clinic&#039;s 
personality was human in that it could feel happy, even smug, about 
acquiring its very own _unique identification_.

&quot;The clinic had other carefully selected human personality 
characteristics such as seeking the good opinion of its peers (humans 
and other of its kind alike).  It also had a few unhuman limits.

More if you google &quot;keith henson&quot; status AI evolutionary psychology

The way you presented it is more rigorous than what I did.

Thanks.

Keith Henson]]></description>
		<content:encoded><![CDATA[<p>I have talked about approval or status seeking as motivation for AIs for a long time, perhaps ten years and about that being a prime and relatively benign human motivation for even longer.  Happy to see the meme getting attention/expansion. </p>
<p>Old stuff here, <a href="http://www.cryonet.org/cgi-bin/dsp.cgi?msg=30170" rel="nofollow">http://www.cryonet.org/cgi-bin/dsp.cgi?msg=30170</a></p>
<p>Here is how I treated it in fiction:</p>
<p>&#8220;As soon as the seed finished the dish (after consulting its clock,<br />
its GPS location and the place of the sun), it aligned the dish on<br />
the African net communication transponder attached to the<br />
geosynchronous ring and asked for a permanently assigned address on<br />
the net.  Up to that point the clinic seed was a generic<br />
product.  The address it was assigned was just a string of<br />
hexadecimal numbers but it was a _unique number_!  The clinic&#8217;s<br />
personality was human in that it could feel happy, even smug, about<br />
acquiring its very own _unique identification_.</p>
<p>&#8220;The clinic had other carefully selected human personality<br />
characteristics such as seeking the good opinion of its peers (humans<br />
and other of its kind alike).  It also had a few unhuman limits.</p>
<p>More if you google &#8220;keith henson&#8221; status AI evolutionary psychology</p>
<p>The way you presented it is more rigorous than what I did.</p>
<p>Thanks.</p>
<p>Keith Henson</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Specifying &#8220;enlightened judgment&#8221; precisely (reprise) &#124; Ordinary Ideas</title>
		<link>https://ordinaryideas.wordpress.com/2014/07/21/approval-seeking/#comment-151</link>
		<dc:creator><![CDATA[Specifying &#8220;enlightened judgment&#8221; precisely (reprise) &#124; Ordinary Ideas]]></dc:creator>
		<pubDate>Wed, 27 Aug 2014 01:28:18 +0000</pubDate>
		<guid isPermaLink="false">http://ordinaryideas.wordpress.com/?p=350#comment-151</guid>
		<description><![CDATA[[&#8230;] Approval-seeking [&#8230;]]]></description>
		<content:encoded><![CDATA[<p>[&#8230;] Approval-seeking [&#8230;]</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Specifying a human precisely (reprise) &#124; Ordinary Ideas</title>
		<link>https://ordinaryideas.wordpress.com/2014/07/21/approval-seeking/#comment-144</link>
		<dc:creator><![CDATA[Specifying a human precisely (reprise) &#124; Ordinary Ideas]]></dc:creator>
		<pubDate>Sun, 24 Aug 2014 21:43:29 +0000</pubDate>
		<guid isPermaLink="false">http://ordinaryideas.wordpress.com/?p=350#comment-144</guid>
		<description><![CDATA[[&#8230;] &#8220;maximize the extent to which I would approve of your decision upon reflection.&#8221; (I have suggested that we would be happy with a powerful AI that made decisions according to this [&#8230;]]]></description>
		<content:encoded><![CDATA[<p>[&#8230;] &#8220;maximize the extent to which I would approve of your decision upon reflection.&#8221; (I have suggested that we would be happy with a powerful AI that made decisions according to this [&#8230;]</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Assorted links</title>
		<link>https://ordinaryideas.wordpress.com/2014/07/21/approval-seeking/#comment-140</link>
		<dc:creator><![CDATA[Assorted links]]></dc:creator>
		<pubDate>Sun, 27 Jul 2014 00:26:24 +0000</pubDate>
		<guid isPermaLink="false">http://ordinaryideas.wordpress.com/?p=350#comment-140</guid>
		<description><![CDATA[[&#8230;] &#8220;Approval-seeking&#8220;: Christiano&#8217;s latest post on indirect normativity for AGI. [&#8230;]]]></description>
		<content:encoded><![CDATA[<p>[&#8230;] &#8220;Approval-seeking&#8220;: Christiano&#8217;s latest post on indirect normativity for AGI. [&#8230;]</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Wei Dai</title>
		<link>https://ordinaryideas.wordpress.com/2014/07/21/approval-seeking/#comment-139</link>
		<dc:creator><![CDATA[Wei Dai]]></dc:creator>
		<pubDate>Wed, 23 Jul 2014 11:26:28 +0000</pubDate>
		<guid isPermaLink="false">http://ordinaryideas.wordpress.com/?p=350#comment-139</guid>
		<description><![CDATA[I&#039;m not saying it&#039;s a deal-breaker either, since that of course depends on what alternatives are available. But one seemingly novel practical implication is that if some form of indirect normativity is likely to be our best bet for controlling AIs, we need to prioritize efforts to figure out whether we would adopt selfish values if we were to undergo a lot of reflection, and if so exactly what form those selfish values would take. (See http://lesswrong.com/lw/8gk/where_do_selfish_values_come_from/ for a related post that tries to make some progress in this direction, but mainly just points out how much we don&#039;t understand about selfishness.)

&gt;for example choosing a person who we think is willing to implement a particular strategy that would respect our egoist values, rather than pursuing their own values

This may not work, no matter who we choose, if for example egoism is objectively correct in some sense, or if all humans just have a psychological tendency to converge towards egoism after extensive reflection.]]></description>
		<content:encoded><![CDATA[<p>I&#8217;m not saying it&#8217;s a deal-breaker either, since that of course depends on what alternatives are available. But one seemingly novel practical implication is that if some form of indirect normativity is likely to be our best bet for controlling AIs, we need to prioritize efforts to figure out whether we would adopt selfish values if we were to undergo a lot of reflection, and if so exactly what form those selfish values would take. (See <a href="http://lesswrong.com/lw/8gk/where_do_selfish_values_come_from/" rel="nofollow">http://lesswrong.com/lw/8gk/where_do_selfish_values_come_from/</a> for a related post that tries to make some progress in this direction, but mainly just points out how much we don&#8217;t understand about selfishness.)</p>
<p>&gt;for example choosing a person who we think is willing to implement a particular strategy that would respect our egoist values, rather than pursuing their own values</p>
<p>This may not work, no matter who we choose, if for example egoism is objectively correct in some sense, or if all humans just have a psychological tendency to converge towards egoism after extensive reflection.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: paulfchristiano</title>
		<link>https://ordinaryideas.wordpress.com/2014/07/21/approval-seeking/#comment-138</link>
		<dc:creator><![CDATA[paulfchristiano]]></dc:creator>
		<pubDate>Wed, 23 Jul 2014 03:38:40 +0000</pubDate>
		<guid isPermaLink="false">http://ordinaryideas.wordpress.com/?p=350#comment-138</guid>
		<description><![CDATA[I agree this is a concern, perhaps a serious one, and it probably warrants more discussion. But the damage is bounded to the extent of our personal preferences for a very precise form of egoism, rather than concern for the people we might have become under the counterfactual. I am not too concerned about huge divergence (so that I no longer sympathize with the simulated person, or am very uncertain about the extent to which they are &quot;like me&quot;), because I can make critical decisions (in the counterfactual simulation) prior to such divergence occurring. For example, I can settle on a definition of &quot;upon reflection&quot; which I think is more likely to preserve the egoist parts of my values. So for the most part, this is something I am willing to accept. I&#039;m also willing to sacrifice a lot on very narrow egoist values (e.g. those that value me over the person I would become upon reflection) in exchange for gains to people generally like me.

There are other approaches to dealing with this if we thought it was a big deal, for example choosing a person who we think is willing to implement a particular strategy that would respect our egoist values, rather than pursuing their own values. Or perhaps choosing a group that would be willing to do so. On this reading, the only point of referring to a person is as a kind of bootstrapping to refer to the conceptual procedure which is locked up in their head.

So overall I am happy to keep this in mind as a pain point, but I don&#039;t see it as a dealbreaker. And as you observe, it doesn&#039;t seem to be a big distinction between this account and CEV, or any other plausible-seeming accounts on the table.]]></description>
		<content:encoded><![CDATA[<p>I agree this is a concern, perhaps a serious one, and it probably warrants more discussion. But the damage is bounded to the extent of our personal preferences for a very precise form of egoism, rather than concern for the people we might have become under the counterfactual. I am not too concerned about huge divergence (so that I no longer sympathize with the simulated person, or am very uncertain about the extent to which they are &#8220;like me&#8221;), because I can make critical decisions (in the counterfactual simulation) prior to such divergence occurring. For example, I can settle on a definition of &#8220;upon reflection&#8221; which I think is more likely to preserve the egoist parts of my values. So for the most part, this is something I am willing to accept. I&#8217;m also willing to sacrifice a lot on very narrow egoist values (e.g. those that value me over the person I would become upon reflection) in exchange for gains to people generally like me.</p>
<p>There are other approaches to dealing with this if we thought it was a big deal, for example choosing a person who we think is willing to implement a particular strategy that would respect our egoist values, rather than pursuing their own values. Or perhaps choosing a group that would be willing to do so. On this reading, the only point of referring to a person is as a kind of bootstrapping to refer to the conceptual procedure which is locked up in their head.</p>
<p>So overall I am happy to keep this in mind as a pain point, but I don&#8217;t see it as a dealbreaker. And as you observe, it doesn&#8217;t seem to be a big distinction between this account and CEV, or any other plausible-seeming accounts on the table.</p>
]]></content:encoded>
	</item>
	<item>
		<title>By: Wei Dai</title>
		<link>https://ordinaryideas.wordpress.com/2014/07/21/approval-seeking/#comment-136</link>
		<dc:creator><![CDATA[Wei Dai]]></dc:creator>
		<pubDate>Wed, 23 Jul 2014 00:26:43 +0000</pubDate>
		<guid isPermaLink="false">http://ordinaryideas.wordpress.com/?p=350#comment-136</guid>
		<description><![CDATA[Hi Paul, did you see the comment I left at http://lesswrong.com/r/discussion/lw/c0k/formalizing_value_extrapolation/b0rb? To generalize that point, I don&#039;t currently know whether &quot;I as an input/output process&quot; actually has the same values as &quot;I as a specific physical entity&quot; so I&#039;m not sure it&#039;s safe to hand over control to the former.]]></description>
		<content:encoded><![CDATA[<p>Hi Paul, did you see the comment I left at <a href="http://lesswrong.com/r/discussion/lw/c0k/formalizing_value_extrapolation/b0rb" rel="nofollow">http://lesswrong.com/r/discussion/lw/c0k/formalizing_value_extrapolation/b0rb</a>? To generalize that point, I don&#8217;t currently know whether &#8220;I as an input/output process&#8221; actually has the same values as &#8220;I as a specific physical entity&#8221; so I&#8217;m not sure it&#8217;s safe to hand over control to the former.</p>
]]></content:encoded>
	</item>
</channel>
</rss>
