<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Virtualized Archives - Artificial Intelligence</title>
	<atom:link href="https://www.aiuniverse.xyz/tag/virtualized/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.aiuniverse.xyz/tag/virtualized/</link>
	<description>Exploring the universe of Intelligence</description>
	<lastBuildDate>Fri, 08 May 2020 12:11:26 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>
	<item>
		<title>NEWS Virtualized GPUs Target Deep Learning Workloads on Kubernetes</title>
		<link>https://www.aiuniverse.xyz/news-virtualized-gpus-target-deep-learning-workloads-on-kubernetes/</link>
					<comments>https://www.aiuniverse.xyz/news-virtualized-gpus-target-deep-learning-workloads-on-kubernetes/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Fri, 08 May 2020 12:11:24 +0000</pubDate>
				<category><![CDATA[Deep Learning]]></category>
		<category><![CDATA[Artificial intelligence (AI)]]></category>
		<category><![CDATA[deep learning]]></category>
		<category><![CDATA[GPU]]></category>
		<category><![CDATA[Kubernetes]]></category>
		<category><![CDATA[Virtualized]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=8677</guid>

					<description><![CDATA[<p>Source: virtualizationreview.com Israel-based Run:AI, specializing in virtualizing artificial intelligence (AI) infrastructure, claimed an industry first in announcing a fractional GPU sharing system for deep learning workloads on <a class="read-more-link" href="https://www.aiuniverse.xyz/news-virtualized-gpus-target-deep-learning-workloads-on-kubernetes/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/news-virtualized-gpus-target-deep-learning-workloads-on-kubernetes/">NEWS Virtualized GPUs Target Deep Learning Workloads on Kubernetes</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: virtualizationreview.com</p>



<p>Israel-based Run:AI, specializing in virtualizing artificial intelligence (AI) infrastructure, claimed an industry first in announcing a fractional GPU sharing system for deep learning workloads on Kubernetes.</p>



<p>The company offers a namesake Run:AI platform built on top of Kubernetes to virtualize AI infrastructure in order to improve on the typical bare-metal approach that statically provisions AI workloads to data scientists. The firm says that approach comes with limits on experiment size and speed, low GPU utilization, and lack of IT controls.</p>



<p>Creating a virtual pool of GPU (graphics processing unit) resources, the company says, abstract data science workloads from infrastructure to simplify workflows.</p>



<p>In an announcement today (May 6), Run:AI said its fractional GPU system lets data science and AI engineering teams run multiple workloads simultaneously on a single GPU, helping organizations run more workloads such as computer vision, voice recognition and natural language processing on the same hardware, lowering costs.</p>



<p>To overcome some limitations on how Kubernetes handles GPUs, the company resorted to some tricky math, effectively marking them as floats that can be fractionalized for use in containers, rather that integers that either exist or don&#8217;t.</p>



<p>&#8220;Today’s de facto standard for deep learning workloads is to run them in containers orchestrated by Kubernetes,&#8221; the company said. &#8220;However, Kubernetes is only able to allocate whole physical GPUs to containers, lacking the isolation and virtualization capabilities needed to allow GPU resources to be shared without memory overflows or processing clashes.&#8221;</p>



<p>The result of the company&#8217;s work to overcome that limitation are virtualized logical GPUs &#8212; sporting their own memory and computing space &#8212; that appear as self-contained processors to containers.</p>



<p>Especially useful in lightweight workloads &#8212; including inference &#8212; eight or more container-run jobs can share the same physical chip, while typical use cases allow for only two to four jobs running on one GPU.</p>
<p>The post <a href="https://www.aiuniverse.xyz/news-virtualized-gpus-target-deep-learning-workloads-on-kubernetes/">NEWS Virtualized GPUs Target Deep Learning Workloads on Kubernetes</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/news-virtualized-gpus-target-deep-learning-workloads-on-kubernetes/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Virtualized AI: Deep Learning Needs More than Just More Compute Power</title>
		<link>https://www.aiuniverse.xyz/virtualized-ai-deep-learning-needs-more-than-just-more-compute-power/</link>
					<comments>https://www.aiuniverse.xyz/virtualized-ai-deep-learning-needs-more-than-just-more-compute-power/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Sat, 25 Apr 2020 12:14:31 +0000</pubDate>
				<category><![CDATA[Deep Learning]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[deep learning]]></category>
		<category><![CDATA[Google]]></category>
		<category><![CDATA[Virtualized]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=8366</guid>

					<description><![CDATA[<p>Source: enterpriseai.news Is the recent progress in deep learning true artificial intelligence? A widely-discussed article by Google’s Francois Chollet discusses the skill acquisition-based approach to gathering intelligence – the one currently in <a class="read-more-link" href="https://www.aiuniverse.xyz/virtualized-ai-deep-learning-needs-more-than-just-more-compute-power/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/virtualized-ai-deep-learning-needs-more-than-just-more-compute-power/">Virtualized AI: Deep Learning Needs More than Just More Compute Power</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: enterpriseai.news</p>



<p>Is the recent progress in deep learning true artificial intelligence? A widely-discussed article by Google’s Francois Chollet discusses the skill acquisition-based approach to gathering intelligence – the one currently in use in modern DL. He argues that with huge data sets available for training models, AI is mastering skill-acquisition but not necessarily the “scope, generalization difficulty, priors, and experience” that true AI should incorporate. Even with our progress in AI, and specifically DL, we are nowhere near the limits of what DL can achieve with bigger, better-trained more accurate models, those that take into account not only skill but experience, and generalization of that experience.</p>



<p>Understandably, this has put intense focus on computing power, particularly the hardware that enables data scientists to run complex training experiments. Nvidia increasing sees DL as a key market for its GPUs and bought Mellanox to speed communication inside a GPU cluster. With its recent acquisition of Habana, Intel is likely betting that custom AI accelerator hardware is a better match. Other AI-first hardware includes Cerebras’s massive chip in a custom box that’s designed for the specific types of intensive, long-running workloads that training DL models require. In the cloud, Google’s Tensor Processing Units offer another bespoke option.</p>



<p>For companies running their own DL workloads, more compute is generally better. Whether exotic AI accelerators or tried-and-tested GPUs, quicker model training means more iterations, faster innovation and reduced time-to-market. It may even mean we can achieve “strong” AI (i.e., AI than goes beyond “narrow AI,” which is the capability of doing a single, discrete task) quicker. In 2020, continuing the trend of recent years, companies will invest in ever-more AI hardware, in an effort to satisfy data scientists’ demands for compute to run bigger models to solve more complex business problems.</p>



<p>But hardware isn’t the whole picture. The conventional computing stack – from processor to firmware to virtualization, abstraction, orchestration and operating layers through to end-user software – was designed for traditional workloads, prioritizing high-availability, short-duration operations.</p>



<p>Training a DL model, though, is the opposite of this sort of workload. While running a model, an experiment may need 100 percent of all the computing power of one or multiple processors for hours or even days at a time.</p>



<p>Part of the challenge is that, while developing a DL algorithm, data scientists have two basic use-patterns for compute resources. The first phase of development is building a model, which includes writing new code and debugging it until the model is ready. During this phase, they tend to use a single GPU often but for short periods of time.</p>



<p>The second phase is training, where the model consumes all the training data and adjusts its parameters. This might consume more than a single GPU and could even take a whole cluster working for days. Sometimes data scientists want to try training a few variations of the same model in parallel to see which performs better.</p>



<p>In a large company, computing resources for DL are typically provided by the IT department. Perhaps each data scientist is statically allocated a fixed amount of physical resources, say a GPU or two for building and training models. Inevitably, this means that expensive processors are sitting idle. Alternatively, a data science team might share their processing power and have to squabble over who gets to tie up the Nvidia DGX AI supercomputer for three days and who has to wait their turn.</p>



<p>All of this also creates challenges for enterprise IT. The IT department has limited visibility into how data science teams are using their expensive compute resources. Meanwhile, the C-suite doesn’t really understand how their GPU resources are being used and whether that usage matches their business goals. Should they invest money in more hardware? Should they hire more data science teams? Or is the issue in the workflow, with&nbsp;<em>both&nbsp;</em>idle resources and data scientists, unable to utilize them, having to wait for compute time.</p>



<p>Every minute a GPU or AI accelerator is idle is an opportunity cost. IT departments face under-utilization of their GPUs while data science teams see their productivity damaged because, from their point of view, the hardware is ‘in use’ and can’t train a new model until it’s finished with its current job. If unused GPUs could be used at full capacity, it would allow faster model training and more iterations and faster time to market.</p>



<p>This is the challenge that companies are beginning to face. Better hardware and more of it might well be necessary, but it isn’t sufficient if the software stack isn’t set up to also make efficient and effective use of that hardware.</p>



<p>The fundamental question of how to share hardware efficiently isn’t new. Some of the challenges that data scientists face could be solved by looking again at how virtualization solved this problem in traditional computing.</p>



<p>Traditional computing uses virtualization to share a single physical resource between multiple workloads. But what if instead of sharing a single physical resource, virtualization was used to create a pool of resources, allowing DL projects to consume as much of the shared resources as they need in an elastic, dynamic way? A virtualized AI infrastructure for DL would run a single workload on multiple shared physical resources. Ideally these resources could be dynamically allocated to the experiments that need them the most, allowing IT administrators to manage resources efficiently, reducing idle GPU time and increasing cluster utilization.</p>



<p>The software stack for DL needs to evolve along with the chips, both to get the most out of individual training experiments and to better optimize running multiple experiments in parallel. Companies will need a full stack, AI-first solution that accounts for the needs of both DL work loads and, critically, DL organizations.</p>
<p>The post <a href="https://www.aiuniverse.xyz/virtualized-ai-deep-learning-needs-more-than-just-more-compute-power/">Virtualized AI: Deep Learning Needs More than Just More Compute Power</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/virtualized-ai-deep-learning-needs-more-than-just-more-compute-power/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>
