<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>MTL Annual Research Report 2011 &#187; Berthold Horn</title>
	<atom:link href="http://www-mtl.mit.edu/wpmu/ar2011/tag/berthold-horn/feed/" rel="self" type="application/rss+xml" />
	<link>http://www-mtl.mit.edu/wpmu/ar2011</link>
	<description>Just another Microsystems Technology Laboratories Blogs site</description>
	<lastBuildDate>Tue, 14 Aug 2012 21:03:56 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.org/?v=3.5.1</generator>
		<item>
		<title>Computer Vision for Vehicles</title>
		<link>http://www-mtl.mit.edu/wpmu/ar2011/computer-vision-for-vehicles-2/</link>
		<comments>http://www-mtl.mit.edu/wpmu/ar2011/computer-vision-for-vehicles-2/#comments</comments>
		<pubDate>Thu, 30 Jun 2011 21:02:45 +0000</pubDate>
		<dc:creator>MTL WP admin</dc:creator>
				<category><![CDATA[Circuits & Systems]]></category>
		<category><![CDATA[Berthold Horn]]></category>
		<category><![CDATA[Ichiro Masaki]]></category>

		<guid isPermaLink="false">http://www-mtl.mit.edu/wpmu/ar2011/?p=3292</guid>
		<description><![CDATA[Under the pressure of increasing population, crowded traffic, the energy crisis, and environmental concerns, current transportation systems have run into...]]></description>
				<content:encoded><![CDATA[<div class="page-restrict-output"><div id="attachment_616" class="wp-caption alignright" style="width: 310px"><a href="https://www-mtl.mit.edu/wpmu/annualreport/files/2010/06/fang_vision_01.jpg" rel="lightbox[3292]"><img class="size-medium wp-image-616" title="fang_vision_01" src="https://www-mtl.mit.edu/wpmu/annualreport/files/2010/06/fang_vision_01-300x75.jpg" alt="Figure 1" width="300" height="75" /></a><p class="wp-caption-text">                              Figure 1: Segmentation result from single camera for urban day-time driving</p></div>
<p>Under the pressure of increasing population, crowded traffic, the energy crisis, and environmental concerns, current transportation systems have run into serious challenges in the following respects: safety, security, efficiency, mobile access, and the environment<sup> [<a href="http://www-mtl.mit.edu/wpmu/ar2011/computer-vision-for-vehicles-2/#footnote_0_3292" id="identifier_0_3292" class="footnote-link footnote-identifier-link" title=" I.~T.~S. of~America, &ldquo;National intelligent transportation system  program plan: A ten-year vision,&rdquo; the United States Department of  Transportation, Tech.  Rep., January 2002.">1</a>] </sup> . There have been over 200,000 pedestrian fatalities in the last 30 years in US. Eighty percent of police reports<sup> [<a href="http://www-mtl.mit.edu/wpmu/ar2011/computer-vision-for-vehicles-2/#footnote_0_3292" id="identifier_1_3292" class="footnote-link footnote-identifier-link" title=" I.~T.~S. of~America, &ldquo;National intelligent transportation system  program plan: A ten-year vision,&rdquo; the United States Department of  Transportation, Tech.  Rep., January 2002.">1</a>] </sup> cited driver errors as the primary cause of vehicle crashes. With the availability of faster computers, better sensor technology, and wider coverage of the wireless communication network, Intelligent Vehicles and Intelligent Transportation Systems (ITS) are gradually being seen as a crucial innovation to improve safety and to reduce damages. It is estimated that implementing collision-avoidance systems in vehicles could prevent 1.1 million accidents in the US each year &#8212; 17 percent of all traffic accidents, which could save 17,500 lives and $26 billion in accident-related costs<sup> [<a href="http://www-mtl.mit.edu/wpmu/ar2011/computer-vision-for-vehicles-2/#footnote_1_3292" id="identifier_2_3292" class="footnote-link footnote-identifier-link" title="The Intelligent Vehicle Initiative: Advancing  &ldquo;Human-Centered&rdquo; Smart Vehicles. Available:  http://www.tfhrc.gov/pubrds/pr97-10/p18.htm">2</a>] </sup> . The demand for in-car electronic products is increasing. Around 35 percent of the cost of car assembly comes from electronics<sup> [<a href="http://www-mtl.mit.edu/wpmu/ar2011/computer-vision-for-vehicles-2/#footnote_2_3292" id="identifier_3_3292" class="footnote-link footnote-identifier-link" title="&ldquo;Asia &ndash; New Hotbed for Consumer Automotive Electronics.&rdquo; Available:  http://www.technewsworld.com/story/52539.html">3</a>] </sup> .</p>
<div id="attachment_617" class="wp-caption alignright" style="width: 310px"><a href="https://www-mtl.mit.edu/wpmu/annualreport/files/2010/06/fang_vision_02.jpg" rel="lightbox[3292]"><img class="size-medium wp-image-617" title="fang_vision_02" src="https://www-mtl.mit.edu/wpmu/annualreport/files/2010/06/fang_vision_02-300x69.jpg" alt="Figure 2" width="300" height="69" /></a><p class="wp-caption-text">                                 Figure 2: Segmentation results for urban day-time driving, for night driving, and for pedestrian tracking.</p></div>
<p>Environment-understanding technology is very vital to provide Intelligent Vehicles with the ability to respond automatically to fast-changing environments and dangerous situations. To obtain perceptual abilities, it is expected to automatically detect static and dynamic obstacles and obtain their related information, such as locations, speed, collision/occlusion possibility, and other dynamic current/historic information. Conventional methods independently detect individual pieces of information, which are normally noisy and not very reliable. Instead we propose a fusion-based and layered-based information-retrieval methodology to systematically detect obstacles and obtain their location/timing information for visible and infrared sequences. The proposed obstacle-detection methodologies take advantage of connections between different kinds of information and increase the computational accuracy of obstacle information estimation, thus improving environment-understanding abilities and driving safety. Three examples are shown in Figures 1 and 2.<sup> [<a href="http://www-mtl.mit.edu/wpmu/ar2011/computer-vision-for-vehicles-2/#footnote_3_3292" id="identifier_4_3292" class="footnote-link footnote-identifier-link" title="Yajun Fang, Sumio Yokomitsu, Berthold Horn, Ichiro Masaki, &ldquo;A Layered-based Fusion-based Approach to Detect and Track the Movements of Pedestrians through Partially Occluded Situations.&rdquo; IEEE Intelligent Vehicles Symposium 2009 (IV2009).">4</a>] </sup><sup> [<a href="http://www-mtl.mit.edu/wpmu/ar2011/computer-vision-for-vehicles-2/#footnote_4_3292" id="identifier_5_3292" class="footnote-link footnote-identifier-link" title="Fang. Y., Horn. B.K.P., Masaki I., &ldquo;Systematic information fusion methodology for static and dynamic obstacle detection in ITS.&rdquo; 15th World Congress On ITS, 2008.">5</a>] </sup><sup> [<a href="http://www-mtl.mit.edu/wpmu/ar2011/computer-vision-for-vehicles-2/#footnote_5_3292" id="identifier_6_3292" class="footnote-link footnote-identifier-link" title="B.K.P.Horn, Y. Fang, I. Masaki, &ldquo;Time to Contact Relative to a Planar Surface.&rdquo; IEEE Intelligent Vehicles Symposium 2007.">6</a>] </sup><sup> [<a href="http://www-mtl.mit.edu/wpmu/ar2011/computer-vision-for-vehicles-2/#footnote_6_3292" id="identifier_7_3292" class="footnote-link footnote-identifier-link" title="Y. Fang, K. Yamada, Y. Ninomiya,   B.K.P. Horn, and I. Masaki, &ldquo;A Shape-Independent-Method for Pedestrian Detection with Far Infrared-images.&rdquo;  Special issue on &ldquo;In-Vehicle Computer Vision Systems&rdquo; of IEEE Transactions on Vehicular Technology, Vol.53, No.6, Nov. 2004, pp.1679-1697.">7</a>] </sup></p>
<ol class="footnotes"><li id="footnote_0_3292" class="footnote"> I.~T.~S. of~America, &#8220;National intelligent transportation system  program plan: A ten-year vision,&#8221; the United States Department of  Transportation, Tech.  Rep., January 2002.</li><li id="footnote_1_3292" class="footnote">The Intelligent Vehicle Initiative: Advancing  &#8220;Human-Centered&#8221; Smart Vehicles. Available:  http://www.tfhrc.gov/pubrds/pr97-10/p18.htm</li><li id="footnote_2_3292" class="footnote">“Asia &#8211; New Hotbed for Consumer Automotive Electronics.” Available:  http://www.technewsworld.com/story/52539.html</li><li id="footnote_3_3292" class="footnote">Yajun Fang, Sumio Yokomitsu, Berthold Horn, Ichiro Masaki, “A Layered-based Fusion-based Approach to Detect and Track the Movements of Pedestrians through Partially Occluded Situations.” IEEE Intelligent Vehicles Symposium 2009 (IV2009).</li><li id="footnote_4_3292" class="footnote">Fang. Y., Horn. B.K.P., Masaki I., “Systematic information fusion methodology for static and dynamic obstacle detection in ITS.” 15th World Congress On ITS, 2008.</li><li id="footnote_5_3292" class="footnote">B.K.P.Horn, Y. Fang, I. Masaki, “Time to Contact Relative to a Planar Surface.” IEEE Intelligent Vehicles Symposium 2007.</li><li id="footnote_6_3292" class="footnote">Y. Fang, K. Yamada, Y. Ninomiya,   B.K.P. Horn, and I. Masaki, “A Shape-Independent-Method for Pedestrian Detection with Far Infrared-images.”  Special issue on &#8220;In-Vehicle Computer Vision Systems&#8221; of IEEE Transactions on Vehicular Technology, Vol.53, No.6, Nov. 2004, pp.1679-1697.</li></ol></div>]]></content:encoded>
			<wfw:commentRss>http://www-mtl.mit.edu/wpmu/ar2011/computer-vision-for-vehicles-2/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
	</channel>
</rss>