<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Artificial Intelligence Blog &#187; Artificial Intelligence Blog &#187; carl</title>
	<atom:link href="http://artent.net/author/carl/feed/" rel="self" type="application/rss+xml" />
	<link>http://artent.net</link>
	<description>We&#039;re blogging machines!</description>
	<lastBuildDate>Sat, 14 Mar 2026 20:14:25 +0000</lastBuildDate>
	<language>en-US</language>
		<sy:updatePeriod>hourly</sy:updatePeriod>
		<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.org/?v=4.0</generator>
	<item>
		<title>DARPA: The future of computing is analog</title>
		<link>http://artent.net/2012/08/27/darpa-the-future-of-computing-is-analog/</link>
		<comments>http://artent.net/2012/08/27/darpa-the-future-of-computing-is-analog/#comments</comments>
		<pubDate>Tue, 28 Aug 2012 02:18:45 +0000</pubDate>
		<dc:creator><![CDATA[carl]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://162.243.213.31/?p=581</guid>
		<description><![CDATA[Wired has an interesting article &#8220;Darpa Has Seen the Future of Computing … And It’s Analog&#8221;. “One of the things that’s happened in the last 10 to 15 years is that power-scaling has stopped,” &#8230; Moore’s law — the maxim that processing power will double every 18 months or so — continues, but battery lives [&#8230;]]]></description>
				<content:encoded><![CDATA[<p>Wired has an interesting article <a href="http://www.wired.com/wiredenterprise/2012/08/upside/">&#8220;Darpa Has Seen the Future of Computing … And It’s Analog&#8221;</a>.</p>
<blockquote><p>
“One of the things that’s happened in the last 10 to 15 years is that power-scaling has stopped,” &#8230; Moore’s law — the maxim that processing power will double every 18 months or so — continues, but battery lives just haven’t kept up. “The efficiency of computation is not increasing very rapidly,” &#8230;.
</p></blockquote>
]]></content:encoded>
			<wfw:commentRss>http://artent.net/2012/08/27/darpa-the-future-of-computing-is-analog/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Automatic Differentiation</title>
		<link>http://artent.net/2012/08/19/automatic-differentiation/</link>
		<comments>http://artent.net/2012/08/19/automatic-differentiation/#comments</comments>
		<pubDate>Sun, 19 Aug 2012 14:09:28 +0000</pubDate>
		<dc:creator><![CDATA[carl]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://162.243.213.31/?p=480</guid>
		<description><![CDATA[I discovered automatic differentiation a few weeks ago, and I can&#8217;t believe I had never heard of it before. Although I believe everybody who has more than a passing knowledge of algorithms (especially numerical algorithms!) should know about it, apparently very few do. I will just give a very brief introduction here before pointing out [&#8230;]]]></description>
				<content:encoded><![CDATA[<p>I discovered <a href="http://en.wikipedia.org/wiki/Automatic_differentiation">automatic differentiation</a> a few weeks ago, and I can&#8217;t believe I had never heard of it before. Although I believe everybody who has more than a passing knowledge of algorithms (especially numerical algorithms!) should know about it, apparently very few do.</p>
<p>I will just give a very brief introduction here before pointing out a few good sources of information.</p>
<p>First of all, automatic differentiation &#8212; &#8220;autodiff&#8221; &#8212; is neither numerical differentiation nor symbolic differentiation, although it does calculate exact derivatives!</p>
<p>In the formulation that I find most astonishing, autodiff uses object-orientied programming techniques and operator overloading to simultaneously and transparently turn <em>all</em> (differentiable) function calculations into evaluations of the function&#8217;s first derivative.</p>
<p>More concretely, instead of calculating with a floating point variable <tt>x</tt>, we instead calculate with an <em>object</em> <tt>x</tt> that has two data components (<tt>x.value</tt> and <tt>x.deriv</tt>, say), the value and the derivative, and has methods that overload all of the mathematical functions and operators in the language. So that, for example, when one calculates <tt>y = cos x</tt> one is automatically calculating both $y = \cos x$ and $y&#8217; = &#8211; \sin x$ and with the results being stored in <tt>y.value</tt> and <tt>y.deriv</tt>! Operator overloading covers cases like <tt>x ^ 3</tt>, <tt>3 ^ x</tt>, or even <tt>x ^ x</tt> using the standard rules of differentiation. And once one realizes how that works, then expressions like <tt>x + y</tt>, <tt>x * y</tt>, and <tt>x / y</tt> become easy as well. In this way, <em>all</em> calculations can be handled.</p>
<p>It should be noted that this works for all numerical computations, not just calculations involving mathematical formulas, and that it can be easily generalized to calculating arbitrary higher-order derivatives.</p>
<p>By the way, the technique outlined above is &#8220;forward mode&#8221; autodiff; it is less obvious that there is also &#8220;reverse mode&#8221; autodiff. Forward mode is more efficient for functions of single variables; reverse mode is more efficient for functions with a single output value (i.e., real-valued as opposed to vector-valued functions). It turns out that reverse mode autodiff is a generalization of neural net back-propagation and was actually discovered before backprop!</p>
<p><a href="http://justindomke.wordpress.com/2009/02/17/automatic-differentiation-the-most-criminally-underused-tool-in-the-potential-machine-learning-toolbox/">Justin Domke made a great blog post on automatic differentiation</a> in 2009; and <a href="http://www.davidson.edu/math/neidinger/SIAMRev74362.pdf">Introduction to Automatic Differentiation and MATLAB Object-Oriented Programming</a> is a very accessible paper on actually implementing autodiff in Matlab.</p>
<p>Finally, <a href="http://www.autodiff.org/">www.autodiff.org/</a> seems to be the home of all things autodiff on the web.</p>
]]></content:encoded>
			<wfw:commentRss>http://artent.net/2012/08/19/automatic-differentiation/feed/</wfw:commentRss>
		<slash:comments>2</slash:comments>
		</item>
		<item>
		<title>Learning by generalization</title>
		<link>http://artent.net/2012/08/10/learning-by-generalization/</link>
		<comments>http://artent.net/2012/08/10/learning-by-generalization/#comments</comments>
		<pubDate>Fri, 10 Aug 2012 18:41:14 +0000</pubDate>
		<dc:creator><![CDATA[carl]]></dc:creator>
				<category><![CDATA[Abstraction for Learning]]></category>

		<guid isPermaLink="false">http://162.243.213.31/?p=270</guid>
		<description><![CDATA[Suppose some Intelligent Algorithm can learn to play tic-tac-toe. Tic-tac-toe is such a simple game that it can be exhaustively learned via memorization &#8211; but this is not what I mean here. I mean that the intelligent algorithm (IA) actually has the ability to generally learn, and that it has used this ability to learn [&#8230;]]]></description>
				<content:encoded><![CDATA[<p>Suppose some Intelligent Algorithm can learn to play tic-tac-toe. Tic-tac-toe is such a simple game that it can be exhaustively learned via memorization &#8211; but this is not what I mean here. I mean that the intelligent algorithm (IA) actually has the ability to generally learn, and that it has used this ability to learn tic-tac-toe. What would one like the IA to now know? Well, in addition to knowing the rules and optimal strategy of the game, the IA should also have figured out that tic-tac-toe is played on a planar grid, and that the winning positions correspond to horizontal, vertical, and diagonal lines on this grid.</p>
<p>Now after it has learned 3 x 3 tic-tac-toe, suppose that we want our IA to also learn to play 4 x 4 tic-tac-toe (perhaps this game should be called tic-toc-toe-tum). The IA should be able to use what it has learned in the 3 x 3 game to more easily learn the 4 x 4 game, and it should explicitly “understand” this relationship. Concretely, what does this mean?</p>
<p>I believe this is one of the couple fundamental open problems in machine learning.</p>
]]></content:encoded>
			<wfw:commentRss>http://artent.net/2012/08/10/learning-by-generalization/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Face Recognition Homepage</title>
		<link>http://artent.net/2012/07/21/face-recognition-homepage/</link>
		<comments>http://artent.net/2012/07/21/face-recognition-homepage/#comments</comments>
		<pubDate>Sun, 22 Jul 2012 01:33:23 +0000</pubDate>
		<dc:creator><![CDATA[carl]]></dc:creator>
				<category><![CDATA[Clustering]]></category>
		<category><![CDATA[General ML]]></category>

		<guid isPermaLink="false">http://162.243.213.31/?p=118</guid>
		<description><![CDATA[I just discover the Face Recognition Homepage while looking for a database of images of faces. But it also is a good place to find algorithms, source code, and other good stuff!]]></description>
				<content:encoded><![CDATA[<p>I just discover the <a href="http://www.face-rec.org/">Face Recognition Homepage</a> while looking for a database of images of faces. But it also is a good place to find algorithms, source code, and other good stuff!</p>
]]></content:encoded>
			<wfw:commentRss>http://artent.net/2012/07/21/face-recognition-homepage/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>Deep belief networks in the news</title>
		<link>http://artent.net/2012/07/13/deep-belief-network-article/</link>
		<comments>http://artent.net/2012/07/13/deep-belief-network-article/#comments</comments>
		<pubDate>Sat, 14 Jul 2012 01:54:24 +0000</pubDate>
		<dc:creator><![CDATA[carl]]></dc:creator>
				<category><![CDATA[Deep Belief Networks]]></category>

		<guid isPermaLink="false">http://162.243.213.31/?p=47</guid>
		<description><![CDATA[“How Many Computers to Identify a Cat? 16,000” is a New York Times article about Deep Belief Networks, “the hottest thing in the speech recognition field these days.”]]></description>
				<content:encoded><![CDATA[<p><a href="http://www.nytimes.com/2012/06/26/technology/in-a-big-network-of-computers-evidence-of-machine-learning.html">“How Many Computers to Identify a Cat? 16,000”</a> is a New York Times article about <a href="http://www.scholarpedia.org/article/Deep_belief_networks">Deep Belief Networks</a>, “the hottest thing in the speech recognition field these days.”</p>
]]></content:encoded>
			<wfw:commentRss>http://artent.net/2012/07/13/deep-belief-network-article/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
	</channel>
</rss>

<!-- Performance optimized by W3 Total Cache. Learn more: http://www.w3-edge.com/wordpress-plugins/

 Served from: artent.net @ 2026-04-11 05:35:40 by W3 Total Cache -->