<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	xmlns:georss="http://www.georss.org/georss" xmlns:geo="http://www.w3.org/2003/01/geo/wgs84_pos#" xmlns:media="http://search.yahoo.com/mrss/"
	>

<channel>
	<title>Ordinary Ideas &#187; Mathematical Logic</title>
	<atom:link href="https://ordinaryideas.wordpress.com/category/mathematical-logic/feed/" rel="self" type="application/rss+xml" />
	<link>https://ordinaryideas.wordpress.com</link>
	<description>As advertised</description>
	<lastBuildDate>Wed, 31 Dec 2014 03:48:50 +0000</lastBuildDate>
	<language>en</language>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.com/</generator>
<cloud domain='ordinaryideas.wordpress.com' port='80' path='/?rsscloud=notify' registerProcedure='' protocol='http-post' />
<image>
		<url>https://s2.wp.com/i/buttonw-com.png</url>
		<title> &#187; Mathematical Logic</title>
		<link>https://ordinaryideas.wordpress.com</link>
	</image>
	<atom:link rel="search" type="application/opensearchdescription+xml" href="https://ordinaryideas.wordpress.com/osd.xml" title="Ordinary Ideas" />
	<atom:link rel='hub' href='https://ordinaryideas.wordpress.com/?pushpress=hub'/>
	<item>
		<title>Confronting Incompleteness</title>
		<link>https://ordinaryideas.wordpress.com/2011/12/25/the-truth-game-confronting-incompleteness/</link>
		<comments>https://ordinaryideas.wordpress.com/2011/12/25/the-truth-game-confronting-incompleteness/#comments</comments>
		<pubDate>Sun, 25 Dec 2011 07:58:06 +0000</pubDate>
		<dc:creator><![CDATA[paulfchristiano]]></dc:creator>
				<category><![CDATA[AI Safety]]></category>
		<category><![CDATA[Mathematical Logic]]></category>

		<guid isPermaLink="false">http://ordinaryideas.wordpress.com/?p=89</guid>
		<description><![CDATA[Consider the &#8220;Truth game&#8221;, played by an agent A: A outputs a sequence of mathematical assertions S1, S2, &#8230; For each statement S, A receives exp(-&#124;S&#124;) utilons. If A makes any false statements (regardless of how many it makes) it receives -infinity utilons (or just a constant larger than the largest possible reward). We can [&#8230;]<img alt="" border="0" src="https://pixel.wp.com/b.gif?host=ordinaryideas.wordpress.com&#038;blog=30479764&#038;post=89&#038;subd=ordinaryideas&#038;ref=&#038;feed=1" width="1" height="1" />]]></description>
				<content:encoded><![CDATA[<p>Consider the &#8220;Truth game&#8221;, played by an agent A:</p>
<blockquote><p>A outputs a sequence of mathematical assertions S1, S2, &#8230;</p>
<p>For each statement S, A receives exp(-|S|) utilons.</p>
<p>If A makes any false statements (regardless of how many it makes) it receives -infinity utilons (or just a constant larger than the largest possible reward).</p></blockquote>
<p><span id="more-89"></span></p>
<p>We can view A&#8217;s output in this game as an operationalization of &#8220;facts about which A is certain.&#8221; I believe that some intuitive judgments about &#8220;belief&#8221; are probably poorly grounded, and that thinking about a more concrete situation can patch some of these issues.</p>
<p>Naturally, it seems like A&#8217;s output should obey certain axioms. For example:</p>
<ul>
<li>If A outputs X and Y, then A outputs &#8220;X and Y.&#8221;</li>
<li>If A outputs X and &#8220;X implies Y&#8221;, then A outputs Y.  (&#8220;Modus Ponens&#8221;)</li>
<li>If A outputs X, then A outputs &#8220;X or Y.&#8221;</li>
</ul>
<p>And so on. For example, we could let A output some axioms in first order logic and then all of their logical consequences: then the set of statements output by A are precisely the theorems for a certain first order theory.</p>
<p>Humans have a strong intuitive feeling that they can &#8220;go beyond&#8221; any particular first order theory, in a sense formalized by Goedel&#8217;s incompleteness results. In particular, humans can feel pretty confident about &#8216;Anything that humans feel pretty confident about is true,&#8217; which is impossible if their confidence were only justified by provability in some formal system.</p>
<p>But let&#8217;s consider agent A playing the Truth game. Should A output a statement of the form &#8220;Everything A outputs is true&#8221;? Well, if the game is played in some reasonable language (say, first order statements about the integers) then A probably can&#8217;t articulate this sentence. But it has some good-enough analogs, like &#8220;If A outputs a statement S, A doesn&#8217;t output (not S)&#8221; which are fair game. Should A output these sentences?</p>
<p>If A ever outputs S and (not S), then it doesn&#8217;t matter what A does&#8211;it is getting -infinity utilons anyway. So A might as well assert that A never outputs both S and not(S). Let this statement be Con(A).</p>
<p>Another natural class of statements for A to output are the results of finite computations: if phi is a delta_0 formula (ie, if we can determine whether phi(x) is true using a deterministic computation) then any true statement of the form &#8220;There is some x such that phi(x)&#8221; should get output by A eventually&#8211;after all, eventually A might as well try every possible candidate x, and if any of them work A can promptly output &#8220;There is some x such that phi(x).&#8221; Call this statement Exhaustive(A). It seems clear that A might as well output Exhaustive(A).</p>
<p>Finally, it seems like A should be able to output &#8220;X implies Y&#8221; when there is a simple (say, strict finitist) proof that X implies Y. I&#8217;m not going to dwell on this because it doesn&#8217;t seem like it is either controversial nor actually problematic. Call this statement Prover(A).</p>
<p>Unfortunately, we are about to get into some trouble.</p>
<p>Consider the statement G = &#8220;A never outputs G.&#8221; It is easy for A to reason as follows:</p>
<blockquote><p>If A outputs G, then by simulating A, A will eventually output &#8216;A outputs G.&#8217; Thus A will have output G and (not G). Thus not Con(A).</p></blockquote>
<p>So if Prover(A), then A outputs &#8220;If A outputs G, then not Con(A)&#8221; which is the same statement as &#8220;If not G, then not Con(A)&#8221; which is in turn the same as &#8220;Con(A) implies G.&#8221; We&#8217;ve already argued that A might as well output &#8220;Con(A)&#8221;. So if we accept modus ponens we are saying that A might as well output &#8220;G&#8221;, in which case A will certainly receive infinite negative utility.</p>
<p>Which one of these legs should be dropped? Should A fail to output Con(A), should A fail to output Exhaustive(A), or should we abandon Prover(A) or modus ponens? If you were in A&#8217;s position you almost certainly wouldn&#8217;t output G. Which would you drop?</p>
<p>I would drop modus ponens, and I suspect that this is the path towards a satisfying theoretical resolution of the problem. When we view mathematical truth as fixed in the background, with A simply trying to discover it, modus ponens is easily justified (and decision-theoretic considerations make no difference). But A is itself made of math, and this seems like an extremely confused perspective, which we might seriously clarify mathematical logic by modifying (or we might not). A TDT agent in A&#8217;s shoes seems likely to output Con(A) and &#8220;Con(A) implies G&#8221; without outputting G, realizing that although G is true outputting it still isn&#8217;t advisable.</p>
<p>I will discuss this more in posts to come, but I should say that the real problem to me at this point seems to be: if you don&#8217;t have modus ponens, how do you do any reasoning at all?</p>
<p>(Also, let me say in passing that this analysis will carry over quite directly to the case of agents manipulating subjective uncertainties.)</p><br />  <a rel="nofollow" href="http://feeds.wordpress.com/1.0/gocomments/ordinaryideas.wordpress.com/89/"><img alt="" border="0" src="http://feeds.wordpress.com/1.0/comments/ordinaryideas.wordpress.com/89/" /></a> <img alt="" border="0" src="https://pixel.wp.com/b.gif?host=ordinaryideas.wordpress.com&#038;blog=30479764&#038;post=89&#038;subd=ordinaryideas&#038;ref=&#038;feed=1" width="1" height="1" />]]></content:encoded>
			<wfw:commentRss>https://ordinaryideas.wordpress.com/2011/12/25/the-truth-game-confronting-incompleteness/feed/</wfw:commentRss>
		<slash:comments>2</slash:comments>
	
		<media:content url="https://0.gravatar.com/avatar/0a58766009bacac271106a2f92289d90?s=96&#38;d=identicon&#38;r=G" medium="image">
			<media:title type="html">paulfchristiano</media:title>
		</media:content>
	</item>
		<item>
		<title>Epistemic Chicken</title>
		<link>https://ordinaryideas.wordpress.com/2011/12/21/epistemic-chicken/</link>
		<comments>https://ordinaryideas.wordpress.com/2011/12/21/epistemic-chicken/#comments</comments>
		<pubDate>Wed, 21 Dec 2011 00:28:52 +0000</pubDate>
		<dc:creator><![CDATA[paulfchristiano]]></dc:creator>
				<category><![CDATA[Mathematical Logic]]></category>

		<guid isPermaLink="false">http://ordinaryideas.wordpress.com/?p=62</guid>
		<description><![CDATA[Consider a fixed goal-seeking agent , who is told its own code and that its objective function is U = { T if A(&#60;A&#62;,&#60;U&#62;) halts after T steps, 0 otherwise }. Alternatively, consider a pair of agents A, B, running similar AIs, who are told their own code as well as their own utility function U [&#8230;]<img alt="" border="0" src="https://pixel.wp.com/b.gif?host=ordinaryideas.wordpress.com&#038;blog=30479764&#038;post=62&#038;subd=ordinaryideas&#038;ref=&#038;feed=1" width="1" height="1" />]]></description>
				<content:encoded><![CDATA[<p>Consider a fixed goal-seeking agent , who is told its own code and that its objective function is U = { T if A(&lt;A&gt;,&lt;U&gt;) halts after T steps, 0 otherwise }. Alternatively, consider a pair of agents A, B, running similar AIs, who are told their own code as well as their own utility function U = { -1 if you don&#8217;t halt, 0 if you halt but your opponent halts after at least as many steps, +1 otherwise }. What would you do as A, in either situation? (That is, what happens if A is an appropriate wrapper around an emulation of your brain, giving it access to arbitrarily powerful computational aids?)</p><br />  <a rel="nofollow" href="http://feeds.wordpress.com/1.0/gocomments/ordinaryideas.wordpress.com/62/"><img alt="" border="0" src="http://feeds.wordpress.com/1.0/comments/ordinaryideas.wordpress.com/62/" /></a> <img alt="" border="0" src="https://pixel.wp.com/b.gif?host=ordinaryideas.wordpress.com&#038;blog=30479764&#038;post=62&#038;subd=ordinaryideas&#038;ref=&#038;feed=1" width="1" height="1" />]]></content:encoded>
			<wfw:commentRss>https://ordinaryideas.wordpress.com/2011/12/21/epistemic-chicken/feed/</wfw:commentRss>
		<slash:comments>1</slash:comments>
	
		<media:content url="https://0.gravatar.com/avatar/0a58766009bacac271106a2f92289d90?s=96&#38;d=identicon&#38;r=G" medium="image">
			<media:title type="html">paulfchristiano</media:title>
		</media:content>
	</item>
	</channel>
</rss>
