<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>computer vision Archives - Artificial Intelligence</title>
	<atom:link href="https://www.aiuniverse.xyz/tag/computer-vision/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.aiuniverse.xyz/tag/computer-vision/</link>
	<description>Exploring the universe of Intelligence</description>
	<lastBuildDate>Sat, 17 Oct 2020 06:10:49 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>
	<item>
		<title>Investment Banking Practice Aimed at Robotics, Automation, AI Launched</title>
		<link>https://www.aiuniverse.xyz/investment-banking-practice-aimed-at-robotics-automation-ai-launched-2/</link>
					<comments>https://www.aiuniverse.xyz/investment-banking-practice-aimed-at-robotics-automation-ai-launched-2/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Sat, 17 Oct 2020 06:10:45 +0000</pubDate>
				<category><![CDATA[Robotics]]></category>
		<category><![CDATA[AI Launched]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[Automation]]></category>
		<category><![CDATA[computer vision]]></category>
		<category><![CDATA[Machine learning]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=12284</guid>

					<description><![CDATA[<p>Source: sme.org SEATTLE – Cascadia Capital said it is launching one of the nation’s first emerging growth investment banking practice groups dedicated to Robotics, Automation, and Artificial <a class="read-more-link" href="https://www.aiuniverse.xyz/investment-banking-practice-aimed-at-robotics-automation-ai-launched-2/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/investment-banking-practice-aimed-at-robotics-automation-ai-launched-2/">Investment Banking Practice Aimed at Robotics, Automation, AI Launched</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: sme.org</p>



<p>SEATTLE – Cascadia Capital said it is launching one of the nation’s first emerging growth investment banking practice groups dedicated to Robotics, Automation, and Artificial Intelligence (RAAI).</p>



<p>Cascadia said its new RAAI group is well-positioned to provide the nuanced M&amp;A and capital raising guidance business owners and entrepreneurs in this sector need as RAAI technology continues to upend critical industries, resulting in permanent market shifts.</p>



<p>The RAAI group is built on the firm’s history of advising companies in the RAAI space through its long-standing Industrials, Energy &amp; Applied Technology, Healthcare, Consumer, and Food &amp; Ag practice groups.</p>



<p>The practice is led by Cascadia Chairman &amp; CEO Michael Butler and Managing Directors Jamie Boyd and Firdaus Pohowalla, with support from Vice Presidents Yee Lee and Jason Lippenberger, Associate Tarek Elmasry, and Analysts Scott Whiting and Mikaela Slade.</p>



<p>“We have been active in the RAAI space for several years with growing excitement and felt that the time was right to formally pull together the firm’s expertise in a way that maximizes the service we deliver to our clients,” Butler said.</p>



<p>“Our bankers have established a deep understanding of the underlying technologies and end markets while simultaneously developing unparalleled relationships with strategic buyers and investors.”</p>



<p>The firm’s most recent RAAI clients include Lucidyne Technologies, Inc., a world-leading manufacturer of AI-enabled computer vision systems; Vexcel Imaging, a leader in digital mapping and drones; and Qi2, a developer of advanced robotic sensor systems for the oilfield. In addition, Cascadia has active engagements with several other current clients in the sector.</p>



<p>The core areas of focus under the RAAI banner include Robotics, Mobility Tech, Manufacturing Automation, Artificial Intelligence/Machine Learning, Computer Vision, Data Management, and Analytics, Augmented Reality/Virtual Reality, and Internet of Things.</p>
<p>The post <a href="https://www.aiuniverse.xyz/investment-banking-practice-aimed-at-robotics-automation-ai-launched-2/">Investment Banking Practice Aimed at Robotics, Automation, AI Launched</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/investment-banking-practice-aimed-at-robotics-automation-ai-launched-2/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>StradVision R&#038;D Director To Reveal Insights Into Optimizing DNNs At 2020 Embedded Vision Summit</title>
		<link>https://www.aiuniverse.xyz/stradvision-rd-director-to-reveal-insights-into-optimizing-dnns-at-2020-embedded-vision-summit/</link>
					<comments>https://www.aiuniverse.xyz/stradvision-rd-director-to-reveal-insights-into-optimizing-dnns-at-2020-embedded-vision-summit/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Fri, 18 Sep 2020 10:25:09 +0000</pubDate>
				<category><![CDATA[Deep Learning]]></category>
		<category><![CDATA[ADAS]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[computer vision]]></category>
		<category><![CDATA[deep learning]]></category>
		<category><![CDATA[StradVision]]></category>
		<category><![CDATA[Technology]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=11667</guid>

					<description><![CDATA[<p>Source: aithority.com StradVision, whose AI-based camera perception software is a leading innovator in Advanced Driver Assistance Systems (ADAS) and Autonomous Vehicles (AVs), will reveal insights into optimizing Deep <a class="read-more-link" href="https://www.aiuniverse.xyz/stradvision-rd-director-to-reveal-insights-into-optimizing-dnns-at-2020-embedded-vision-summit/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/stradvision-rd-director-to-reveal-insights-into-optimizing-dnns-at-2020-embedded-vision-summit/">StradVision R&#038;D Director To Reveal Insights Into Optimizing DNNs At 2020 Embedded Vision Summit</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: aithority.com</p>



<p>StradVision, whose AI-based camera perception software is a leading innovator in Advanced Driver Assistance Systems (ADAS) and Autonomous Vehicles (AVs), will reveal insights into optimizing Deep Neural Networks (DNNs) for multiple processors at the 2020 Embedded Vision Summit. During the “Designing Bespoke DNNs for Target Hardware” session, which will be held at 9:30am Pacific Time on September 17, R&amp;D Director and Co-founder Woonhyun Nam will also elaborate on StradVision’s patent know-how via its deep learning-based SVNet technology.</p>



<p>The presentation marks the sixth time that StradVision has participated in the pre-eminent conference and expo devoted to practical, deployed computer vision and visual AI.</p>



<p>This year’s conference includes speakers from Google, Intel, Samsung, Qualcomm, and LG Electronics, with a keynote speech from Google’s Distinguished Engineer David Patterson, who is also a professor at the University of California, Berkeley, and the Vice-Chair of the RISC-V Foundation.</p>



<p>During his session, Nam will touch on StradVision’s established patents, including their patented DNN-enabled SVNet software, as well as discuss the challenges and opportunities of Deep Neural Networks, including cost-effective techniques to optimize DNNs to better fit different processors while also reducing model size and power consumption. The effort required to transform the DNN for each processor can be prohibitive for DNN developers, and Nam will explain how quantization and structured sparsification techniques can help by reducing model size and computation significantly.</p>



<p>A Ph.D. holder of Pohang University of Science and Technology’s Computer Vision Lab, Nam will draw from his experience as the head of the algorithm engineering team at StradVision, whose pioneering SVNet software can quickly and accurately identify potentially hazardous objects and road conditions for ADAS assisted and autonomous vehicles, such as other vehicles, lanes, pedestrians, animals, free space, traffic signs, and lights, even in harsh weather conditions or poor lighting.</p>



<p>Nam’s team is responsible for defining SVNet’s technical architecture, deploying new algorithm pipelines, and delivering algorithm solutions best optimized for each hardware target.</p>



<p>SVNet&nbsp;relies on deep learning-based embedded perception algorithms, which compared with its competitors is more compact and requires dramatically less memory and electricity to run. It supports more than 14 hardware platforms thanks to StradVision’s patented and cutting-edge DNN-enabled software.</p>



<p>Attendees will also learn about StradVision’s experience in bringing AI-based camera perception software to the mass market through international collaboration as well as the view from&nbsp;Asia&nbsp;on the future of the ADAS and AV market.</p>



<p>StradVision’s software is currently being deployed in 8.8 million vehicles worldwide, such as SUVs, sedans, trucks, and self-driving buses, and maintains partnerships with leading global automotive Tier 1 suppliers and five of the world’s top auto OEMs. StradVision’s global partners also include NVIDIA, Aisin Group, Hyundai Motor Group, LG Electronics, Texas Instruments, Renesas, Qualcomm, Xilinx, Socionext, Ambarella, and BlackBerry QNX.</p>



<p>StradVision has obtained&nbsp;China’s&nbsp;Guobiao certification and the coveted ASPICE CL2 (Automotive Software Performance Improvement and Capability Determination Containment Level 2) certification and was recently awarded the Grand Prize in the Electric/Electronic Category at the 14th Korea Patent Excellence Awards.</p>



<p>Since 2012, the Embedded Vision Summit has been organized each May in&nbsp;California&nbsp;by Edge AI and Vision Alliance, an industry partnership operated by Berkeley Design Technology, Inc. Participants regularly include industry experts, business leaders, leading academics, investors, and entrepreneurs interested in visual AI.</p>



<p>Due to global health concerns, the 2020 Embedded Vision Summit will be held online from&nbsp;September 15 to 25.</p>
<p>The post <a href="https://www.aiuniverse.xyz/stradvision-rd-director-to-reveal-insights-into-optimizing-dnns-at-2020-embedded-vision-summit/">StradVision R&#038;D Director To Reveal Insights Into Optimizing DNNs At 2020 Embedded Vision Summit</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/stradvision-rd-director-to-reveal-insights-into-optimizing-dnns-at-2020-embedded-vision-summit/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>MACHINE LEARNING MODELS CAN REASON ABOUT DAILY TASKS AND ACTIONS</title>
		<link>https://www.aiuniverse.xyz/machine-learning-models-can-reason-about-daily-tasks-and-actions/</link>
					<comments>https://www.aiuniverse.xyz/machine-learning-models-can-reason-about-daily-tasks-and-actions/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Mon, 07 Sep 2020 07:38:50 +0000</pubDate>
				<category><![CDATA[Machine Learning]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[computer vision]]></category>
		<category><![CDATA[Machine learning]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=11409</guid>

					<description><![CDATA[<p>Source: analyticsinsight.net Late advancements in artificial intelligence have recharged interest in building frameworks that learn and think as individuals. Numerous advances have originated from utilizing deep neural networks trained <a class="read-more-link" href="https://www.aiuniverse.xyz/machine-learning-models-can-reason-about-daily-tasks-and-actions/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/machine-learning-models-can-reason-about-daily-tasks-and-actions/">MACHINE LEARNING MODELS CAN REASON ABOUT DAILY TASKS AND ACTIONS</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: analyticsinsight.net</p>



<p>Late advancements in artificial intelligence have recharged interest in building frameworks that learn and think as individuals. Numerous advances have originated from utilizing deep neural networks trained end-to-end in operations, for example, object recognition, video games, and board games, accomplishing tasks that are equal to or even beats people in certain regards. In spite of their biological inspiration and performance achievements, these frameworks are different from human intelligence in essential ways. Cognitive science is growing and proposing human-like learning and thinking machines should reach past current engineering trends in both what they learn, and how they learn it.</p>



<p>Machines should</p>



<p>• form causal models of the universe that help understanding and explanation, as opposed to just tackling pattern recognition problems;</p>



<p>• ground learning in intuitive theories of physics and psychology, to help and improve the information that is found out; and</p>



<p>• Leverage compositionality and figuring out how to figure out learning quickly and sum up knowledge to new assignments and circumstances.</p>



<p>Due to new computing advances, machine learning today isn’t like machine learning of the past. It was conceived from pattern recognition and the theory that PCs can learn without being programmed to perform explicit tasks; scientists intrigued by artificial intelligence needed to check whether computers could gain from data.</p>



<p>The iterative part of machine learning is significant in light of the fact that as models are presented to new information, they can independently adjust. They learn from previous computations to create solid, repeatable decisions and results. It’s a science that is not new – but rather one that has gained new momentum.</p>



<p>In another study at the European Conference on Computer Vision recently, scientists revealed a hybrid language-vision model that can contrast and compare a lot of dynamic occasions caught on video to coax out the elevated level ideas connecting them.</p>



<p>Their model showed improvement over people at two kinds of visual reasoning tasks — picking the video that adroitly best finishes the set, and picking the video that doesn’t fit. Demonstrated videos of a dog barking and a man yelling close to his dog, for instance, the model finished the set by picking the crying child from a set of five recordings. Researchers imitated their outcomes on two datasets for training AI frameworks in action recognition: MIT’s Multi-Moments in Time and DeepMind’s Kinetics.</p>



<p>According to Mathew Monfort, study co-creator and&nbsp; a research scientist at MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL), “Language representations permit us to incorporate contextual information learned from text databases into our visual models.”</p>



<p>Words like ‘running,’ ‘lifting,’ and ‘boxing’ share some common characteristics that make them all the more closely related with the idea ‘working out,’ for instance, than ‘driving.’</p>



<p>Utilizing WordNet, a database of word implications, the researchers planned the connection of each action-class label in Moments and Kinetics to different names in both datasets. Words like “sculpting,” “carving,” and “cutting,” for instance, were associated with more significant level ideas like “crafting,” “making workmanship,” and “cooking.” Now when the model perceives an action like sculpting, it can choose reasonably comparative exercises in the database.</p>



<p>To perceive how the model would be compared to people, the researchers requested human subjects play out similar arrangements of visual reasoning tasks online. Amazingly, the model performed as well as people in numerous situations, at times with startling outcomes. In a minor departure from the set completion task, subsequent to viewing a video of somebody wrapping a gift and covering an item in tape, the model recommended a video of somebody at the seashore covering another person in the sand.</p>



<p>Impediments of the model incorporate an inclination to overemphasize a few highlights. In one case, it recommended finishing a lot of sports videos with a video of a child and a ball, obviously connecting balls with exercise and competition.</p>



<p>A deep learning model that can be trained to “think” all the more abstractly might be fit for learning with less information, state analysts. Abstraction additionally prepares toward a more significant level, more human-like thinking.</p>
<p>The post <a href="https://www.aiuniverse.xyz/machine-learning-models-can-reason-about-daily-tasks-and-actions/">MACHINE LEARNING MODELS CAN REASON ABOUT DAILY TASKS AND ACTIONS</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/machine-learning-models-can-reason-about-daily-tasks-and-actions/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Counting on AI: The right time for researchers to embrace Artificial Intelligence</title>
		<link>https://www.aiuniverse.xyz/counting-on-ai-the-right-time-for-researchers-to-embrace-artificial-intelligence/</link>
					<comments>https://www.aiuniverse.xyz/counting-on-ai-the-right-time-for-researchers-to-embrace-artificial-intelligence/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Wed, 22 Jul 2020 09:02:40 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[computer vision]]></category>
		<category><![CDATA[COVID-19]]></category>
		<category><![CDATA[deep–learnin]]></category>
		<category><![CDATA[researchers]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=10388</guid>

					<description><![CDATA[<p>Source: dqindia.com While Artificial General Intelligence,&#160;or “singularity” as they call it,&#160;may be decades away, we have already reached a point where AI can significantly&#160;augment our intelligence and&#160;help <a class="read-more-link" href="https://www.aiuniverse.xyz/counting-on-ai-the-right-time-for-researchers-to-embrace-artificial-intelligence/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/counting-on-ai-the-right-time-for-researchers-to-embrace-artificial-intelligence/">Counting on AI: The right time for researchers to embrace Artificial Intelligence</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: dqindia.com</p>



<p>While Artificial General Intelligence,&nbsp;or “singularity” as they call it,&nbsp;may be decades away, we have already reached a point where AI can significantly&nbsp;augment our intelligence and&nbsp;help us achieve&nbsp;better&nbsp;outputs&nbsp;at a&nbsp;faster&nbsp;pace.</p>



<p>As of today, there is no area where AI has not been proven useful. From playing games to flying airplanes and from detecting cancers to automatically cleaning up selfie-portraits, AI has made its presence felt in all domains.</p>



<p>There are over&nbsp;8&nbsp;million&nbsp;active researchers&nbsp;who collectively&nbsp;spend&nbsp;over $1.5&nbsp;trillion&nbsp;on&nbsp;academic research,&nbsp;with&nbsp;the&nbsp;promise&nbsp;of&nbsp;advancing&nbsp;the&nbsp;world’s&nbsp;combined&nbsp;knowledge and&nbsp;intellect.</p>



<p>The&nbsp;right AI–powered tools and techniques can make&nbsp;a&nbsp;significant difference in how research is conducted and how fast results&nbsp;are&nbsp;obtained.</p>



<p>Until a year ago, the general public may not have understood or even paid much heed&nbsp;to&nbsp;the need for speed&nbsp;and&nbsp;accuracy&nbsp;when it comes to research.&nbsp;But&nbsp;because of&nbsp;the recent&nbsp;COVID-19 situation, many are recognizing&nbsp;and feeling the pain of the&nbsp;pace&nbsp;of research&nbsp;in&nbsp;the race&nbsp;to find&nbsp;an antiviral&nbsp;drug&nbsp;or a vaccine.</p>



<p>While some of the results of academic research&nbsp;are&nbsp;celebrated,&nbsp;it is easy to forget&nbsp;the&nbsp;countless&nbsp;steps and processes behind the scenes, which last&nbsp;many months before any results are achieved;&nbsp;more often than not, the results&nbsp;of research&nbsp;aren’t always revolutionary or directly useful.</p>



<p>One of the early stages&nbsp;of the research lifecycle is discovery.&nbsp;On average, researchers spend 4 hours every&nbsp;week searching&nbsp;through&nbsp;research and 5 hours reading&nbsp;articles, with only 50% of the&nbsp;articles&nbsp;being useful. Here, AI can come in to help&nbsp;researchers&nbsp;discover the right articles&nbsp;to read.</p>



<p>There&nbsp;are many tools out therethat are&nbsp;powered by natural language processing and search based on machine–learned concepts, which help researchers narrow down their reading and discover the relevant research&nbsp;much&nbsp;faster.</p>



<p>The next&nbsp;stage&nbsp;is the actual research,&nbsp;which consists of gathering&nbsp;data;&nbsp;running experiments based on various hypotheses;&nbsp;collecting, analyzing,&nbsp;and&nbsp;representing the research outputs;&nbsp;and arriving at&nbsp;the&nbsp;conclusions.</p>



<p>For the&nbsp;above&nbsp;steps, many AI open–source tools,&nbsp;such as Python, R, Pandas, Scikit,&nbsp;and&nbsp;Spark,&nbsp;as well as&nbsp;proprietary AI tools like Mathematica,&nbsp;Matlab,&nbsp;and&nbsp;SAS can be very useful, especially when&nbsp;directed toward&nbsp;statistical machine learning.</p>



<p>Many research labs are making use of advanced AI streams such as computer vision, robotic arms, IOT,&nbsp;and speech and audio to assist them in the research process. &nbsp;</p>



<p>Finally, the most important&nbsp;stage&nbsp;for researchers&nbsp;is&nbsp;the&nbsp;publication and dissemination of their research—the&nbsp;tedious&nbsp;and&nbsp;time–consuming albeit critical&nbsp;final&nbsp;step of the process.</p>



<p>While there are&nbsp;editing services that&nbsp;exist&nbsp;to&nbsp;help&nbsp;with&nbsp;manuscript&nbsp;preparation, formatting,&nbsp;and language&nbsp;correction, there are many AI tools out there that can be used by researchers,&nbsp;which help&nbsp;with&nbsp;writing&nbsp;manuscripts, correcting grammar and language,&nbsp;and&nbsp;formatting&nbsp;them&nbsp;as per target journal standards,&nbsp;in addition to&nbsp;automated solutions for styling&nbsp;figures, tables,&nbsp;captions,&nbsp;and citations.</p>



<p>Pub-sure.com is an online suite of assistive tools that helpsresearchers&nbsp;make&nbsp;their&nbsp;manuscripts&nbsp;publication–ready.</p>



<p>Since its inception, Cactus has been partnering with researchers to&nbsp;assist&nbsp;them in their&nbsp;research&nbsp;journey.&nbsp;It&nbsp;has been&nbsp;our constant endeavor to&nbsp;enable&nbsp;researchers and innovators to find analogous concepts and novel ideas from different industries and fields.</p>



<p>We are excited to have&nbsp;entered the AI and deep–learning space&nbsp;as well, as the need of the hour is&nbsp;to develop innovative products for publishers&nbsp;as well as&nbsp;business and tech solutions for stakeholders in&nbsp;the research landscape.</p>



<p>With powerful initiatives like&nbsp;researcher.life,&nbsp;our&nbsp;aim is to&nbsp;put the researcher at the center of research.&nbsp;We have already&nbsp;developedseveral AI–powered tools that help researchers focus on their main work, the research. As a community, however, we still have a&nbsp;long&nbsp;way to go&nbsp;before&nbsp;AI&nbsp;is&nbsp;fully integrated in the researcher’s ecosystem.</p>
<p>The post <a href="https://www.aiuniverse.xyz/counting-on-ai-the-right-time-for-researchers-to-embrace-artificial-intelligence/">Counting on AI: The right time for researchers to embrace Artificial Intelligence</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/counting-on-ai-the-right-time-for-researchers-to-embrace-artificial-intelligence/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Shiny objects foil robots, but RGB-D holds the key</title>
		<link>https://www.aiuniverse.xyz/shiny-objects-foil-robots-but-rgb-d-holds-the-key/</link>
					<comments>https://www.aiuniverse.xyz/shiny-objects-foil-robots-but-rgb-d-holds-the-key/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Wed, 22 Jul 2020 07:44:22 +0000</pubDate>
				<category><![CDATA[Robotics]]></category>
		<category><![CDATA[applications]]></category>
		<category><![CDATA[Automation]]></category>
		<category><![CDATA[computer vision]]></category>
		<category><![CDATA[machine vision]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=10379</guid>

					<description><![CDATA[<p>Source: zdnet.com Who doesn&#8217;t love shiny things? Well&#8230; robots for one. The same goes for transparent objects. At least, that&#8217;s long been the case. Machine vision has stumbled when it comes <a class="read-more-link" href="https://www.aiuniverse.xyz/shiny-objects-foil-robots-but-rgb-d-holds-the-key/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/shiny-objects-foil-robots-but-rgb-d-holds-the-key/">Shiny objects foil robots, but RGB-D holds the key</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: zdnet.com</p>



<p>Who doesn&#8217;t love shiny things? Well&#8230; robots for one. The same goes for transparent objects.</p>



<p>At least, that&#8217;s long been the case. Machine vision has stumbled when it comes to shiny or reflective surfaces, and that&#8217;s limited use cases for automation even as advances in the field push robots into more and more new spaces.</p>



<p>Now, researchers at robotics powerhouse Carnegie Mellon report success with a new technique to identify and grasp objects with troublesome surfaces. Rather than relying on expensive new sensor technology or intensive modeling and training via AI, the system instead goes back to basics, relying on a simple color camera.</p>



<p>To understand why it&#8217;s necessary to understand how robots currently sense objects prior to grasping. Cutting edge computer vision systems for pick-and-place applications often rely on infrared cameras, which are great for sensing and precisely measuring the depth of an object &#8212; useful data for a robot devising a grasping strategy &#8212; but fall short when it comes to visual quirks like transparency. Infrared light passes right through clear objects and is reflected and scattered by reflective surfaces.</p>



<p>Color cameras, however, can detect both. Just look at any color photo and you&#8217;ll clearly discern a glass on a table or a shiny metal railing, each with lots of rich detail. That was the vital clue. The CMU researchers built on this observation and developed a color camera system capable of recognizing shapes using color and, crucially, sensing transparent or reflective surfaces. </p>



<p>&#8220;We do sometimes miss,&#8221; David Held, an assistant professor in CMU&#8217;s Robotics Institute, acknowledged, &#8220;but for the most part it did a pretty good job, much better than any previous system for grasping transparent or reflective objects.&#8221;</p>



<p>That the solution is low-cost and the sensors battle-tested give it a tremendous leg up when it comes to the potential for adoption. The researchers point out that other attempts at robotic grasping of transparent objects have relied on training systems based on trial and error or on expensive human labeling of objects.</p>



<p>In the end, it&#8217;s the end, it&#8217;s not new sensors, but new strategies to use them that may give robots the powers they need to function in everyday life.&nbsp;</p>
<p>The post <a href="https://www.aiuniverse.xyz/shiny-objects-foil-robots-but-rgb-d-holds-the-key/">Shiny objects foil robots, but RGB-D holds the key</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/shiny-objects-foil-robots-but-rgb-d-holds-the-key/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Google Open-Sources Computer Vision Model Big Transfer</title>
		<link>https://www.aiuniverse.xyz/google-open-sources-computer-vision-model-big-transfer/</link>
					<comments>https://www.aiuniverse.xyz/google-open-sources-computer-vision-model-big-transfer/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Wed, 10 Jun 2020 07:27:35 +0000</pubDate>
				<category><![CDATA[Deep Learning]]></category>
		<category><![CDATA[computer vision]]></category>
		<category><![CDATA[deep-learning]]></category>
		<category><![CDATA[Google]]></category>
		<category><![CDATA[researchers]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=9422</guid>

					<description><![CDATA[<p>Source: infoq.com Google Brain has released the pre-trained models and fine-tuning code for Big Transfer (BiT), a deep-learning computer vision model. The models are pre-trained on publicly-available generic image datasets <a class="read-more-link" href="https://www.aiuniverse.xyz/google-open-sources-computer-vision-model-big-transfer/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/google-open-sources-computer-vision-model-big-transfer/">Google Open-Sources Computer Vision Model Big Transfer</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: infoq.com</p>



<p>Google Brain has released the pre-trained models and fine-tuning code for Big Transfer (BiT), a deep-learning computer vision model. The models are pre-trained on publicly-available generic image datasets and can meet or exceed state-of-the-art performance on several vision benchmarks after fine-tuning on just a few samples.</p>



<p>Paper co-authors Lucas Beyer and Alexander Kolesnikov gave an overview of their work in a recent blog post. To help advance the performance of deep-learning vision models, the team investigated large-scale pre-training and the effects of model size, dataset size, training duration, normalization strategy, and hyperparameter choice. As a result of this work, the team developed a &#8220;recipe&#8221; of components and training heuristics that achieves strong performance on a variety of benchmarks, including an &#8220;unprecedented top-5 accuracy of 80.0%&#8221; on the ObjectNet dataset. Beyer and Kolesnikov claim,</p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow"><p>[Big Transfer] will allow anyone to reach state-of-the-art performance on their task of interest, even with just a handful of labeled images per class.</p></blockquote>



<p>Deep-learning models have made great strides in computer vision, particularly in recognizing objects in images. One key to this success has been the availability of large-scale labelled datasets: collections of images with corresponding text descriptions of the objects they contain. These datasets must be created manually, with human workers applying a label to each of thousands of images: the popular ImageNet dataset, for example, contains over 14 million labeled images containing 21k different object classes. However, the images are usually generic, showing commonplace objects such as people, pets, or household items. Creating a dataset of similar scale for a specialized task, say for an industrial robot, might be prohibitively expensive or time-consuming.</p>



<p>In this situation, AI engineers often apply transfer learning, a strategy that has become popular with large-scale natural-language processing (NLP) models. A neural network is first <em>pre-trained</em> on a large generic dataset until it achieves a certain level of performance on a test dataset. Then the model is fine-tuned with a smaller task-specific dataset, sometimes with as few as a single example of the task-specific objects. Large NLP models routinely set new state-of-the-art performance levels using transfer learning.</p>



<p>For BiT, the Google researchers used a ResNet-v2 neural architecture. To investigate the effects of pre-training dataset size, the team replicated their experiments on three groups of models pre-trained with different datasets: BiT-S models pre-trained on 1.28M images from ILSVRC-2012, BiT-M models pre-trained on 14.2M images from ImageNet-21k, and BiT-L models pre-trained on 300M images from JFT-300M. The models were then fine-tuned and evaluated on several common benchmarks: ILSVRC-2012, CIFAR-10/100, Oxford-IIIT Pet, and Oxford Flowers-102.</p>



<p>The team noted several findings from their experiments. First, the benefits from increasing model size diminish on smaller datasets, and there is little benefit in pre-training smaller models on larger datasets. Second, the large models performed better using group normalization compared to batch normalization. Finally, to avoid an expensive hyperparameter search during fine-tuning, the team developed a heuristic called BiT-HyperRule, where all hyperparameters are fixed except &#8220;training schedule length, resolution, and whether to use MixUp regularization.&#8221;</p>



<p>Google has released the best-performing pre-trained models from the BiT-S and BiT-M groups. However, they have not released any of the BiT-L models based on the JFT-300M dataset. Commenters on Hacker News pointed out that no model trained on JFT-300M has ever been released. One commenter pointed to several models released by Facebook which were pre-trained on an even larger dataset. Another said:</p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow"><p>I&#8217;ve wondered if legal/copyright issues block any release: there&#8217;s always someone who tries to argue that a model is a derived work, and nothing in the JFT-300M papers mentions having licenses covering public redistribution.</p></blockquote>



<p>The code for fine-tuning and tutorials for using the released pre-trained models are available on GitHub.</p>
<p>The post <a href="https://www.aiuniverse.xyz/google-open-sources-computer-vision-model-big-transfer/">Google Open-Sources Computer Vision Model Big Transfer</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/google-open-sources-computer-vision-model-big-transfer/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Deep learning and AI drives ‘computer vision’ market</title>
		<link>https://www.aiuniverse.xyz/deep-learning-and-ai-drives-computer-vision-market/</link>
					<comments>https://www.aiuniverse.xyz/deep-learning-and-ai-drives-computer-vision-market/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Tue, 26 May 2020 06:48:46 +0000</pubDate>
				<category><![CDATA[Deep Learning]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[computer vision]]></category>
		<category><![CDATA[deep learning]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=9019</guid>

					<description><![CDATA[<p>Source: gadget.co.za Computer vision (CV) is aiding numerous promising applications and revolutionizing the way people do things. According to new research from Omdia, the proof-of-concepts that started <a class="read-more-link" href="https://www.aiuniverse.xyz/deep-learning-and-ai-drives-computer-vision-market/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/deep-learning-and-ai-drives-computer-vision-market/">Deep learning and AI drives ‘computer vision’ market</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: gadget.co.za</p>



<p>Computer vision (CV) is aiding numerous promising applications and revolutionizing the way people do things. According to new research from Omdia, the proof-of-concepts that started a few years ago are now going into production, with deployments across a wide range of applications.</p>



<p>Many businesses are realizing the value of vision and are starting to determine how to use CV technology for commercial benefit. Technology is evolving to address these needs. According to Omdia principal analyst Anand Joshi, “Deep learning has become the technology of choice for computer vision applications replacing classic computer vision techniques. This, in turn, is driving the need for new chipsets and software for CV applications.” Global CV market revenue is expected to grow from $2.9bn in 2018 to $33.5bn by 2025.</p>



<p>Omdia’s report, “Computer Vision Technologies and Market,” analyzes the market trends and technology issues surrounding CV. The study examines the industries and use cases and provides profiles of key industry players. Global market forecasts, segmented by region, industry, and use case, extend through 2025. An Executive Summary of the report is available for free download on the firm’s website.</p>
<p>The post <a href="https://www.aiuniverse.xyz/deep-learning-and-ai-drives-computer-vision-market/">Deep learning and AI drives ‘computer vision’ market</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/deep-learning-and-ai-drives-computer-vision-market/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>What’s the point: Google AI advances, Git security updates, and API gateway Gloo</title>
		<link>https://www.aiuniverse.xyz/whats-the-point-google-ai-advances-git-security-updates-and-api-gateway-gloo/</link>
					<comments>https://www.aiuniverse.xyz/whats-the-point-google-ai-advances-git-security-updates-and-api-gateway-gloo/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Sat, 18 Apr 2020 08:45:27 +0000</pubDate>
				<category><![CDATA[Google AI]]></category>
		<category><![CDATA[computer vision]]></category>
		<category><![CDATA[Git]]></category>
		<category><![CDATA[Google]]></category>
		<category><![CDATA[Reinforcement Learning]]></category>
		<category><![CDATA[Security]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=8255</guid>

					<description><![CDATA[<p>Source: devclass.com Google’s AI teams used the comparatively quiet post-easter days to get ML practitioners up to speed with their latest research in reinforcement learning, natural language <a class="read-more-link" href="https://www.aiuniverse.xyz/whats-the-point-google-ai-advances-git-security-updates-and-api-gateway-gloo/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/whats-the-point-google-ai-advances-git-security-updates-and-api-gateway-gloo/">What’s the point: Google AI advances, Git security updates, and API gateway Gloo</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: devclass.com</p>



<p>Google’s AI teams used the comparatively quiet post-easter days to get ML practitioners up to speed with their latest research in reinforcement learning, natural language processing, and computer vision.</p>



<p>In “An optimistic perspective on offline reinforcement learning”, a team of researchers has looked into ways to use a fixed offline dataset of logged interactions to teach agents how to handle themselves in real world situations. While agents normally learn by getting live feedback from their environment, this approach is meant to be useful in certain robotics use cases or autonomous driving, where enough recorded interaction data is available and other ways of collecting information either seem insufficient or are too expensive to realise.</p>



<p>These are usually seen as tricky to implement, since there’s no real way of knowing how an agent should be rewarded when it takes an action that differs from the dataset provided. To tackle that, Google’s AI team added some supervised learning methods into the mix which helps to improve generalisation and make the whole system more robust. The results are called Ensemble-DQN and Random Ensemble Mixture and can be investigated here.</p>



<p>Meanwhile another team has been busy improving the way objects are detected. The outcome has been dubbed EfficientDet and will be presented at the renowned computer vision conference CVPR in Seattle in June – if COVID refrains from putting a spoke in their wheel. It aims at introducing a “new family of scalable and efficient object detectors” to the computer vision community, building upon earlier work concerning the scaling of neural networks (EfficientNet).</p>



<p>In EfficientDet, EfficientNet is used as a backbone to more effectively extract features from images, while a new bi-directional feature network in combination with a fresh normalised fusion technique is meant to get to image characteristics faster at a lower computation cost.</p>



<p>If you’re more interested in NLP, Google has also been busy setting up a benchmark to make comparing multilingual representations easier. XTREME covers 40 languages from 12 language families and includes nine tasks ranging from sentence classification to question answering to evaluate methods making the most of the shared structures of languages. The project can be found at GitHub.</p>



<h4 class="wp-block-heading">Git pushes out security updates to stop tricksters</h4>



<p>This week, Git maintainer Junio C Hamano has unleashed versions v2.26.1, v2.25.3, v2.24.2, v2.23.2, v2.22.3, v2.21.2, v2.20.3, v2.19.4, v2.18.3, and v2.17.4 of the version control system onto the coding masses. </p>



<p>Updating is strongly advised, since the security fixes mediate an issue which “allowed a crafted URL to trick a Git client to send credential information for a wrong host to the attacker’s site”.&nbsp;</p>



<h4 class="wp-block-heading">Gloo lures admins with new dev portal</h4>



<p>Envoy-based API gateway Gloo hit version 1.3 earlier this week, focusing on performance, stability and extensibility improvements. However, the release also includes a developer portal, so that admins have an easier way of controlling who gets access to which APIs.</p>



<p>Once set up, they can select which interfaces should be shared at all and decide which users and groups get to see them once they’ve logged into the portal. The whole apparatus is designed for self-service, with the Gloo team promising easy integration into continuous delivery processes.</p>
<p>The post <a href="https://www.aiuniverse.xyz/whats-the-point-google-ai-advances-git-security-updates-and-api-gateway-gloo/">What’s the point: Google AI advances, Git security updates, and API gateway Gloo</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/whats-the-point-google-ai-advances-git-security-updates-and-api-gateway-gloo/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>I Wizards- A Complete Solution OF Artificial Intelligence &#038; Computer Vision</title>
		<link>https://www.aiuniverse.xyz/i-wizards-a-complete-solution-of-artificial-intelligence-computer-vision/</link>
					<comments>https://www.aiuniverse.xyz/i-wizards-a-complete-solution-of-artificial-intelligence-computer-vision/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Mon, 23 Mar 2020 07:48:07 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[CCTV cameras]]></category>
		<category><![CDATA[computer vision]]></category>
		<category><![CDATA[wizards solution]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=7654</guid>

					<description><![CDATA[<p>Source: inventiva.co.in There is a great importance of CCTV cameras in every place, be it a public place or a working place or home itself, every workplace <a class="read-more-link" href="https://www.aiuniverse.xyz/i-wizards-a-complete-solution-of-artificial-intelligence-computer-vision/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/i-wizards-a-complete-solution-of-artificial-intelligence-computer-vision/">I Wizards- A Complete Solution OF Artificial Intelligence &#038; Computer Vision</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: inventiva.co.in</p>



<p>There is a great importance of CCTV cameras in every place, be it a public place or a working place or home itself, every workplace and public place has CCTV cameras installed in it. CCTV cameras not only ensure safety but also provides peace of mind to the owner when he/she is absent from the office. There is no doubt in saying this that every place where CCTV camera is installed it acts as an severe deterrent to the criminals or anyone carting out illegal activity. In offices also it is always suggested to keep a record of your employees that at what time they are coming and leaving even, it helps to keep a record of visitors in the office.&nbsp;</p>



<p>It is essential to monitor each and everything happening in your office, as an entrepreneur one should always keep a check on what is going on in the office and what has to be done. Well talking of monitoring, records, and operating, it is not always possible for the owner and the employees to check CCTV footage every five to ten minutes, so to serve this purpose, one can have artificial intelligence to make their work effective and efficient.</p>



<p>Integration wizards solution has made it possible now this organization is a platform for artificial intelligence for CCTV cameras. The company is based in Bangalore, India that offers Computer Vision and Enterprise Mobility based solutions. It helps enterprises determine actionable intelligence from images, videos, and data captured live.&nbsp;</p>



<p><strong>What is unique about I wizards?</strong></p>



<p>Almost all the companies have a CCTV network, but only a few know its worth and importance. I wizards change this passive asset into an active solution by their flagship product, IRIS it is a smart solution that uses the existing CCTV network to provide customized solutions to enterprises. It provides insights into the areas that are otherwise opaque to organizations. From retail, logistics to warehouses or solar farms, it is easily installed anywhere to tackle various aspects of manufacturing, such as safety compliance, process optimization in warehouses, etc. It is also being deployed for outdoor security solutions and retail for multiple use cases.</p>



<p>Powered by artificial intelligence, the company’s real-time computer vision-based product IRIS analyzes and understands images on an advanced level. This involves everything from object recognition (think security systems that alert authorities after “seeing” a fire or an intruder) to navigation mapping (similar to GPS tracking with google maps), emotional recognition (now you don’t have to depend on manual customer feedback!), to posture recognition (useful in manual handling in industries), OCR and much more.</p>



<p><strong>About the CEO and co-founder:</strong></p>



<p>Co-founder and CEO, Kunal Kislay, is a young, determined entrepreneur who, along with two of his colleagues, Saquib Khan and Kumar Raman, bootstrapped a start-up that has been profitable from its inception, and now has an annual revenue of $3.2 million.</p>



<p>Kunal is a B.Tech IIT Mumbai alumnus with over a decade of experience in enterprise mobility, which inspired him to delve into the then futuristic domain further – Internet of Things, AI, Neural networks and Machine learning. He knew the pulse of technology and its changing paradigms. With the insight gained from the vast array of verticals he worked with and solutions he created, it was simple for him to realize that the future lies in the domain of Artificial Intelligence and Computer vision.&nbsp;&nbsp;</p>



<p>In the next few years, the aim of the company is expanding our installed base in warehouses, manufacturing premises and retail outlets at present.</p>
<p>The post <a href="https://www.aiuniverse.xyz/i-wizards-a-complete-solution-of-artificial-intelligence-computer-vision/">I Wizards- A Complete Solution OF Artificial Intelligence &#038; Computer Vision</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/i-wizards-a-complete-solution-of-artificial-intelligence-computer-vision/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Facebook AI Releases New Computer Vision Library Detectron2</title>
		<link>https://www.aiuniverse.xyz/facebook-ai-releases-new-computer-vision-library-detectron2/</link>
					<comments>https://www.aiuniverse.xyz/facebook-ai-releases-new-computer-vision-library-detectron2/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Tue, 29 Oct 2019 06:59:45 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[computer vision]]></category>
		<category><![CDATA[Facebook]]></category>
		<category><![CDATA[Research]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=4900</guid>

					<description><![CDATA[<p>Source: infoq.com Facebook AI Research (FAIR) has released Detectron2, a PyTorch-based computer vision library that brings a series of new research and production capabilities to the popular framework. Since its <a class="read-more-link" href="https://www.aiuniverse.xyz/facebook-ai-releases-new-computer-vision-library-detectron2/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/facebook-ai-releases-new-computer-vision-library-detectron2/">Facebook AI Releases New Computer Vision Library Detectron2</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: infoq.com</p>



<p>Facebook AI Research (FAIR) has released Detectron2, a PyTorch-based computer vision library that brings a series of new research and production capabilities to the popular framework.</p>



<p>Since its release in 2018, the original Detectron object detection platform has become one of FAIR’s most widely adopted open-source projects. While the first Detectron was written in Caffe2, Detectron2 represents a full rewrite of the original framework in PyTorch from the ground up, with several new object detection capabilities.</p>



<p>Detectron was, at the time of its initial release, a huge boost for the AI community. It enabled many to quickly and easily build state-of-the-art object detection models. Yet Detectron was stuck with a few limitations — limitations that quickly became deal-breakers for many AI practitioners.</p>



<ul class="wp-block-list"><li>Caffe2 was complicated, so implementing custom object detection models was a big challenge</li><li>From the beginning, Detectron was only designed for object detection and no other computer vision tasks such as semantic segmentation or pose estimation</li><li>With deploying machine learning models to production becoming a hot topic over the past couple of years, Detectron quickly fell behind as it lacked the capability for exporting inference models</li></ul>



<p>Detectron2 was built to tackle those deal-breakers, making for a more robust and modern library. From the Detectron2 team at FAIR:</p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow"><p>We built Detectron2 to meet the research needs of Facebook AI and to provide the foundation for object detection in production use cases at Facebook. We are now using Detectron2 to rapidly design and train the next-generation pose detection models that power Smart Camera, the AI camera system in Facebook’s Portal video-calling devices. By relying on Detectron2 as the unified library for object detection across research and production use cases, we are able to rapidly move research ideas into production models that are deployed at scale.</p></blockquote>



<p>The move to PyTorch aligns with the AI community’s growing need and desire for a flexible yet easy-to-use library. PyTorch itself is modular by design, making it far easier to extend than Caffe2. The vast majority of the AI community already uses just two libraries: TensorFlow and PyTorch.</p>



<p>Detectron2 has expanded to handle computer vision tasks beyond object detection including semantic segmentation, panoptic segmentation, pose estimation, and DensePose. The authors have made a noticeable effort to add pre-trained state-of-the-art models like Cascade R-CNN, Panoptic FPN, and TensorMask.</p>



<p>FAIR’s team hinted in their official blog post that they’re planning to release an additional component to the library, Detectron2go, to make it easier to deploy models to production. It’s said to include features like network quantization, model optimization, and formatting for mobile deployment.</p>
<p>The post <a href="https://www.aiuniverse.xyz/facebook-ai-releases-new-computer-vision-library-detectron2/">Facebook AI Releases New Computer Vision Library Detectron2</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/facebook-ai-releases-new-computer-vision-library-detectron2/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>
