<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>datasets Archives - Artificial Intelligence</title>
	<atom:link href="https://www.aiuniverse.xyz/tag/datasets/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.aiuniverse.xyz/tag/datasets/</link>
	<description>Exploring the universe of Intelligence</description>
	<lastBuildDate>Tue, 17 Mar 2020 10:08:24 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>
	<item>
		<title>Call to Action to the Tech Community on New Machine Readable COVID-19 Dataset</title>
		<link>https://www.aiuniverse.xyz/call-to-action-to-the-tech-community-on-new-machine-readable-covid-19-dataset/</link>
					<comments>https://www.aiuniverse.xyz/call-to-action-to-the-tech-community-on-new-machine-readable-covid-19-dataset/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Tue, 17 Mar 2020 10:08:22 +0000</pubDate>
				<category><![CDATA[Data Mining]]></category>
		<category><![CDATA[data mining]]></category>
		<category><![CDATA[datasets]]></category>
		<category><![CDATA[Machine Readable]]></category>
		<category><![CDATA[researchers]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=7499</guid>

					<description><![CDATA[<p>Source: whitehouse.gov Today, researchers and leaders from the Allen Institute for AI, Chan Zuckerberg Initiative (CZI), Georgetown University’s Center for Security and Emerging Technology (CSET), Microsoft, and <a class="read-more-link" href="https://www.aiuniverse.xyz/call-to-action-to-the-tech-community-on-new-machine-readable-covid-19-dataset/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/call-to-action-to-the-tech-community-on-new-machine-readable-covid-19-dataset/">Call to Action to the Tech Community on New Machine Readable COVID-19 Dataset</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: whitehouse.gov</p>



<p>Today, researchers and leaders from the Allen Institute for AI, Chan Zuckerberg Initiative (CZI), Georgetown University’s Center for Security and Emerging Technology (CSET), Microsoft, and the National Library of Medicine (NLM) at the National Institutes of Health released the COVID-19 Open Research Dataset (CORD-19) of scholarly literature about COVID-19, SARS-CoV-2, and the Coronavirus group.</p>



<p>Requested by The White House Office of Science and Technology Policy, the dataset represents the most extensive machine-readable Coronavirus literature collection available for data and text mining to date, with over 29,000 articles, more than 13,000 of which have full text.</p>



<p>Now, The White House joins these institutions in issuing a call to action to the Nation’s artificial intelligence experts to develop new text and data mining techniques that can help the science community answer high-priority scientific questions related to COVID-19.</p>



<p>The collection was constructed via a unique collaboration between Microsoft, NLM, CZI, and the Allen Institute for AI, coordinated by Georgetown University. Microsoft’s web-scale literature curation tools were used to identify and bring together worldwide scientific efforts and results, CZI provided access to pre-publication content, NLM provided access to literature content, and the Allen AI team transformed the content into machine-readable form, making the corpus ready for analysis and study.</p>



<p>The CORD-19 resource is available on the Allen Institute’s SemanticScholar.org website and will continue to be updated as new research is published in archival services and peer-reviewed publications. Researchers should submit the text and data mining tools and insights they develop in response to this call to action via the <a href="https://www.kaggle.com/allen-institute-for-ai/CORD-19-research-challenge">Kagg</a>le platform. Through Kaggle, a machine learning and data science community owned by Google Cloud, these tools will be openly available for researchers around the world.</p>



<p>To inform the call to action, key scientific questions related to COVID-19 were developed in coordination with the National Academies of Sciences, Engineering, and Medicine’s Standing Committee on Emerging Infectious Diseases and 21<sup>st</sup>&nbsp;Century Health Threats and the World Health Organization. The call to action and key questions are both available on Kaggle.</p>



<p>“Decisive action from America’s science and technology enterprise is critical to prevent, detect, treat, and develop solutions to COVID-19. The White House will continue to be a strong partner in this all hands-on-deck approach. We thank each institution for voluntarily lending its expertise and innovation to this collaborative effort, and call on the United States research community to put artificial intelligence technologies to work in answering key scientific questions about the novel Coronavirus,” said Michael Kratsios, U.S. Chief Technology Officer, The White House.</p>



<p>“This valuable new resource is the fruit of unselfish collaboration and now offers the opportunity to find answers to important questions about COVID-19,” said Dr. Dewey Murdick, Director of Data Science at Georgetown University’s Center for Security and Emerging Technology (CSET), who coordinated the cross-team effort. “Once the crisis has passed, we hope this project will inspire new ways to use machine learning to advance scientific research.”</p>



<p>“It’s all-hands on deck as we face the COVID-19 pandemic,” said Dr. Eric Horvitz, Chief Scientific Officer at Microsoft. “We need to come together as companies, governments, and scientists and work to bring our best technologies to bear across biomedicine, epidemiology, AI, and other sciences. The COVID-19 literature resource and challenge will stimulate efforts that can accelerate the path to solutions on COVID-19.”</p>



<p>“Sharing vital information across scientific and medical communities is key to accelerating our ability to respond to the coronavirus pandemic,” said Dr. Cori Bargmann, Head of Science at the Chan Zuckerberg Initiative. “The new COVID-19 Open Research Dataset will help researchers worldwide to access important information faster.”</p>



<p>“We are excited to be part of this collaboration to aid in the COVID-19 response and that the group is making use of our open access subset on coronavirus literature,” said Dr. Patricia Flatley Brennan, Director of the National Library of Medicine at the National Institutes of Health. “Our current collection of more than 10,000 full-text scholarly articles related to coronavirus provides a critical resource for text mining efforts like this one.”</p>



<p><strong>“</strong>One of the most immediate&nbsp;and impactful applications of AI is in the ability to help scientists, academics, and technologists find the right information in a sea of scientific papers to move research faster. We applaud the OSTP, WHO, NIH and all organizations that are taking a proactive approach to use the most advanced technology in the fight against COVID-19,”&nbsp;said Dr. Oren Etzioni, Chief Executive Officer of the Allen Institute for AI.&nbsp;“The Allen Institute for AI, and particularly the Semantic Scholar team, is committed to updating and improving this important resource and the associated AI methods the community will be using to tackle this crucial problem.”</p>



<p>“It’s difficult for people to manually go through more than 20,000 articles and synthesize their findings. Recent advances in technology can be helpful here. We’re putting machine readable versions of these articles in front of our community of more than 4 million data scientists. Our hope is that AI can be used to help find answers to a key set of questions about COVID-19,” said Anthony Goldbloom, Co-Founder and Chief Executive Officer at Kaggle.</p>
<p>The post <a href="https://www.aiuniverse.xyz/call-to-action-to-the-tech-community-on-new-machine-readable-covid-19-dataset/">Call to Action to the Tech Community on New Machine Readable COVID-19 Dataset</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/call-to-action-to-the-tech-community-on-new-machine-readable-covid-19-dataset/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>The Decade of Data Science</title>
		<link>https://www.aiuniverse.xyz/the-decade-of-data-science/</link>
					<comments>https://www.aiuniverse.xyz/the-decade-of-data-science/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Thu, 02 Jan 2020 07:14:26 +0000</pubDate>
				<category><![CDATA[Data Science]]></category>
		<category><![CDATA[data science]]></category>
		<category><![CDATA[datasets]]></category>
		<category><![CDATA[Development]]></category>
		<category><![CDATA[GitHub]]></category>
		<category><![CDATA[Google]]></category>
		<category><![CDATA[Microsoft]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=5913</guid>

					<description><![CDATA[<p>Source: towardsdatascience.com It’s been a hell of a decade for data science — Watson dominated Jeopardy (2011), the White House announced the first Chief Data Scientist (2015), <a class="read-more-link" href="https://www.aiuniverse.xyz/the-decade-of-data-science/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/the-decade-of-data-science/">The Decade of Data Science</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: towardsdatascience.com</p>



<p>It’s been a hell of a decade for data science — Watson dominated Jeopardy (2011), the White House announced the first Chief Data Scientist (2015), and Deepfakes provided all the Nicolas Cage films we ever wanted (2017). Even our discipline’s name, data science, came out of the last 10 years. With 2019 coming to a close, there will inevitably be articles and lists with date-specific events like those above. However, there have also been several important trends that have developed slowly over the last 10 years. When I think of data science now compared to where it was in 2010, three items stand out to me.</p>



<h2 class="wp-block-heading" id="8476">1. A Language of Our Own</h2>



<p>Data scientists often come to the profession from a variety of backgrounds, Mathematics, Astronomy, or Medicine. This all lends perspective to our work, but one of the early drawbacks of having many backgrounds was that it resulted in a tower of Babel of programming languages. The early 2010s saw the beginning of a language war. Paid solutions like STATA, SPSS, SAS, Matlab, and Mathematica vied with open source tools like R, C++, Python, Julia, Octave and Gretl.</p>



<p>At first, R appeared like it might emerge victorious. Microsoft even made a play by acquiring Revolution Analytics (a company that provides an enhanced version of R). Despite this early lead, 2015 marked a shift towards Python. Its impressive library of packages, easy-to-learn syntax, and solid performance all played in its favor.</p>



<p>Fast forward to today. The war is over, new borders are drawn, and Python is the language of choice. This isn’t to say that all other languages are dead. R is still heavily used in statistics, and depending on your industry, you might be utilizing SAS or Matlab. However, 9 times out of 10, the answer is Python. The benefits of this shared language can’t be overstated. It makes it easier to read others’ work and code is more easily shared.</p>



<h2 class="wp-block-heading" id="4670">2. Jupyter Notebooks &amp; GitHub</h2>



<p>Search for a tutorial on any data science subject and whatever article you land on will almost certainly contain a link to a Jupyter Notebook hosted on GitHub. This practice is so common it’s essentially an unspoken standard. And rightfully so: it’s a powerful combo. Jupyter Notebooks combine your documentation and code into a single file that makes understanding the <em>what</em> and <em>why</em> much easier. Couple this with GitHub and you not only have a method for explaining your work but also a way of sharing it. GitHub has even taken steps to embrace Jupyter Notebooks by allowing them to be rendered right in the browser.</p>



<p>Where Python gives us a common language, Jupyter and GitHub have given us a common grammar and etiquette of sharing.</p>



<h2 class="wp-block-heading" id="ef50">3. Proliferation of Datasets</h2>



<p>10 years ago data was much harder to come by. Even when you did find a dataset on your desired topic, chances were it was small. There was data out there, but it was unorganized, formats were inconsistent, and people were less likely to share what they had collected. Before I had the MNIST or Titanic dataset, I had the “Cars” dataset from STATA. It was a collection of vehicles from 1978 that included features like headroom and fuel efficiency and came with a whopping 74 observations.</p>



<p>Compare that to now, where you literally have companies posting their data for the world to see. In a way, Kaggle has become the Tinder of corporate data. Instead of being hesitant, companies are now enthusiastically sharing data, hoping to get a swipe right from an eager young quant. Local governments have also stepped up their game. Cities like Chicago and New York have made it possible for anyone to get access to data on topics ranging from road conditions to rodent sightings. Resources for searching datasets have also improved with tools like the r/datasets subreddit and the Google Dataset Search. If data is the new oil, then every data scientist today probably feels a little like the Daniel Day-Lewis in <em>There Will Be Blood</em>.</p>



<h2 class="wp-block-heading" id="2085">The Common Thread: Open Science</h2>



<p>There’s a through line in the above topics, they each represent a move towards Open Science or the philosophy that tools, methods, and data should be made openly available. It’s the adoption of this ideology that I think has been one of the most profitable investments our community has made in the last decade, aiding our growth, expanding our capabilities, and strengthening the quality of our work.</p>



<p>In terms of growth, the gates to data science are now open to all, in contrast to less than a decade ago when they required a hefty toll of entry. Before Python, SAS was the closest thing to a standard language, and before Jupyter Notebooks there were Wolfram Notebooks. But these solutions weren’t open and free. Depending on the service, you would have been paying between $800 — $8000 just to run a regression with an “industry standard” tool. Even their websites don’t make it easy for beginners. Visit any of them and you’re presented with dozens of different versions and links asking if you’d like to “get a quote,” making the experience feel less like diving into a dynamic discipline and more like buying a used car.</p>



<p>Commercial software for data science has waned over the last decade. I won’t claim to know the exact reason, but I’ll give my perspective: nothing extinguishes curiosity like a price tag. When I was nearing the end of my undergrad, I remember thinking “I like doing numerical analysis, but there’s no way I can afford STATA. I guess I’d better learn this R thing.” I wasn’t alone, there has been a massive influx of amateurs who are curious and eager to test the waters of data science.</p>



<p>In addition to easy entry, open science has expanded the depth and breadth of everyone’s capabilities. A single data scientist isn’t necessarily a single resource. There’s a massive library of code from contributors around the world. This may take the form of Python packages, Medium tutorials, or Stack Overflow answers. This code corpus makes projects less about building something from scratch and more like mixing and matching Lego sets.</p>



<p>This openness also helps with keeping our work honest. Along with data science, this decade also brought us the phrase “replication crisis.” While there’s no cure-all for fraudulent work, transparency is a potent preventative measure. One of my favorite articles of the decade, The Scientific Paper is Obsolete, has a great quote from Wolfram talking about notebooks, saying, “there’s no bullshit there. It is what it is, it does what it does. You don’t get to fudge your data.”</p>



<h2 class="wp-block-heading" id="330f">Looking Forward, Looking Back</h2>



<p>In some ways this wasn’t just a good decade for data science, it was&nbsp;<em>the&nbsp;</em>decade of data science. We entered the zeitgeist of the 2010s. We were given the title of sexiest profession, we got a Brad Pitt and Jonah Hill movie, and millions saw AlphaGO become a world champion. In some ways, I feel like this decade was our coming of age. While the next decade will bring with it many improved tools and methodologies, the 2010s will be a time I look back on with a sense of fondness and nostalgia.</p>
<p>The post <a href="https://www.aiuniverse.xyz/the-decade-of-data-science/">The Decade of Data Science</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/the-decade-of-data-science/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>NumPy 1.17.0 is here, officially drops Python 2.7 support pushing forward Python 3 adoption</title>
		<link>https://www.aiuniverse.xyz/numpy-1-17-0-is-here-officially-drops-python-2-7-support-pushing-forward-python-3-adoption/</link>
					<comments>https://www.aiuniverse.xyz/numpy-1-17-0-is-here-officially-drops-python-2-7-support-pushing-forward-python-3-adoption/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Thu, 01 Aug 2019 05:40:00 +0000</pubDate>
				<category><![CDATA[Python]]></category>
		<category><![CDATA[datasets]]></category>
		<category><![CDATA[NumPy]]></category>
		<category><![CDATA[numpy.random]]></category>
		<category><![CDATA[transforms]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=4181</guid>

					<description><![CDATA[<p>Source: hub.packtpub.com Last week, the Python team released NumPy version 1.17.0. This version has many new features, improvements and changes to increase the performance of NumPy. The major highlight <a class="read-more-link" href="https://www.aiuniverse.xyz/numpy-1-17-0-is-here-officially-drops-python-2-7-support-pushing-forward-python-3-adoption/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/numpy-1-17-0-is-here-officially-drops-python-2-7-support-pushing-forward-python-3-adoption/">NumPy 1.17.0 is here, officially drops Python 2.7 support pushing forward Python 3 adoption</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: hub.packtpub.com</p>



<p>Last week, the Python team released NumPy version 1.17.0. This version has many new features, improvements and changes to increase the performance of NumPy.</p>



<p>The major highlight of this release includes a new extensible numpy.random module, new radix sort &amp; timsort sorting methods and a NumPy pocketfft FFT implementation for accurate transforms and better handling of datasets of prime length. Overriding of numpy functions has also been made possible by default.</p>



<p>NumPy 1.17.0 will support Python versions 3.5 – 3.7. Python 3.8b2 will work with the new release source packages, but may not find support in future releases. Python version 2.7 has been officially dropped.</p>



<h4 class="wp-block-heading">What’s new in NumPy 1.17.0?</h4>



<h4 class="wp-block-heading">New extensible numpy.random module with selectable random number generators</h4>



<p>NumPy 1.17.0 has a new extensible numpy.random module. It also includes four selectable random number generators and improved seeding designed for use in parallel processes.&nbsp;<em>PCG64</em>&nbsp;is the new default numpy.random module while&nbsp;<em>MT19937</em>&nbsp;is retained for backwards compatibility.</p>



<h4 class="wp-block-heading">Timsort and radix sort have replaced mergesort for stable sorting</h4>



<p>Both the radix sort and timsort have been implemented and can be used instead of mergesort. The sorting kind options ‘stable’ and ‘mergesort’ have been made aliases of each other with the actual sort implementation for maintaining backward compatibility. Radix sort is used for small integer types of 16 bits or less and timsort is used for all the remaining types of bits.</p>



<h4 class="wp-block-heading">empty_like and related functions now accept a shape argument</h4>



<p>Functions like empty_like, full_like, ones_like and zeros_like will now accept a shape keyword argument, which can be used to create a new array as the prototype and overriding its shape also. These functions become extremely useful when combined with&nbsp;<em>the __array_function__&nbsp;</em>protocol, as it allows the creation of new arbitrary-shape arrays from NumPy-like libraries.</p>



<h4 class="wp-block-heading">User-defined LAPACK detection order</h4>



<p><em>numpy.distutils</em>&nbsp;now uses an environment variable, comma-separated and case insensitive detection order to determine the detection order for LAPACK libraries. This aims to help users with MKL installation to try different implementations.</p>



<h4 class="wp-block-heading">.npy files support unicode field names</h4>



<p>A new format version of .npy files has been introduced. This enables structured types with non-latin1 field names. It can be used automatically when needed.</p>



<h4 class="wp-block-heading">New mode “empty” for pad</h4>



<p>The new mode “empty” pads an array to a desired shape without initializing any new entries.</p>



<h4 class="wp-block-heading">New Deprications in NumPy 1.17.0</h4>



<h4 class="wp-block-heading">numpy.polynomial functions warn when passed float in place of int</h4>



<p>Previously, functions in numpy.polynomial module used to accept float values. With the latest NumPy version 1.17.0, using float values is deprecated for consistency with the rest of NumPy. In future releases, it will cause a TypeError.</p>



<h4 class="wp-block-heading">Deprecate numpy.distutils.exec_command and temp_file_name</h4>



<p>The internal use of these functions has been refactored for better alternatives such as replace&nbsp;<em>exec_command</em>&nbsp;with subprocess. Also, replace Popen and temp_file_name&nbsp;with tempfile.mkstemp.</p>



<h4 class="wp-block-heading">Writeable flag of C-API wrapped arrays</h4>



<p>When an array is created from the C-API to wrap a pointer to data, the writeable flag set during creation indicates the read-write nature of the data. In the future releases, it will not be possible to convert the writeable flag to True from python as it is considered dangerous.</p>



<h2 class="wp-block-heading">Other improvements and changes</h2>



<h4 class="wp-block-heading">Replacement of the fftpack based fft module by the pocketfft library</h4>



<p><em>pocketfft</em>&nbsp;library contains additional modifications compared to&nbsp;<em>fftpack</em>&nbsp;which helps in improving accuracy and performance. If FFT lengths has large prime factors then&nbsp;<em>pocketfft</em>uses Bluestein’s algorithm, which maintains O(N log N) run time complexity instead of deteriorating towards O(N*N) for prime lengths.</p>



<h4 class="wp-block-heading">Array comparison assertions include maximum differences</h4>



<p>Error messages from array comparison tests such as testing.assert_allclos now include “max absolute difference” and “max relative difference” along with previous “mismatch” percentage. This makes it easier to update absolute and relative error tolerances.</p>



<h4 class="wp-block-heading">median and percentile family of functions no longer warn about nan</h4>



<p>Functions like numpy.median, numpy.percentile, and numpy.quantile are used to emit a&nbsp;<em>RuntimeWarning</em>&nbsp;when encountering a nan. Since these functions return the nan value, the warning is redundant and hence has been removed.</p>



<h4 class="wp-block-heading">timedelta64 % 0 behavior adjusted to return NaT</h4>



<p>The modulus operation with two&nbsp;<em>np.timedelta64</em>&nbsp;operands now returns&nbsp;<em>NaT</em>&nbsp;in case of division by zero, rather than returning zero.</p>



<p>Though users are happy with NumPy 1.17.0 features, some are upset over the Python version 2.7 being officially dropped.</p>
<p>The post <a href="https://www.aiuniverse.xyz/numpy-1-17-0-is-here-officially-drops-python-2-7-support-pushing-forward-python-3-adoption/">NumPy 1.17.0 is here, officially drops Python 2.7 support pushing forward Python 3 adoption</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/numpy-1-17-0-is-here-officially-drops-python-2-7-support-pushing-forward-python-3-adoption/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Big Blue opens up hub for machine learning datasets</title>
		<link>https://www.aiuniverse.xyz/big-blue-opens-up-hub-for-machine-learning-datasets/</link>
					<comments>https://www.aiuniverse.xyz/big-blue-opens-up-hub-for-machine-learning-datasets/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Thu, 18 Jul 2019 12:21:37 +0000</pubDate>
				<category><![CDATA[Deep Learning]]></category>
		<category><![CDATA[Big Blue]]></category>
		<category><![CDATA[datasets]]></category>
		<category><![CDATA[deep learning]]></category>
		<category><![CDATA[IBM]]></category>
		<category><![CDATA[Learning]]></category>
		<category><![CDATA[machine]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=4058</guid>

					<description><![CDATA[<p>Source: devclass.com IBM has launched a repository of datasets for training which data scientists can pick and mix to train their deep learning and machine learning models. <a class="read-more-link" href="https://www.aiuniverse.xyz/big-blue-opens-up-hub-for-machine-learning-datasets/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/big-blue-opens-up-hub-for-machine-learning-datasets/">Big Blue opens up hub for machine learning datasets</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: devclass.com</p>



<p>IBM has launched a repository of datasets for training which data scientists can pick and mix to train their deep learning and machine learning models.<br></p>



<p>The IBM Data Asset eXchange (DAX) is designed to complement the Model Asset eXchange it launched earlier this year, which offers researchers and developers models to deploy or train with their own data.<br></p>



<p>In a blog announcing the data exchange, a quartet of IBM luminaries, wrote “Developers adopting ML models need open data that they can use confidently under clearly defined open data licenses.”<br></p>



<p>The data sets in question will be covered by the Linux Foundation’s Community Data License Agreement (CDLA) open data licensing framework to enable data sharing and collaboration – “where possible”.</p>



<p>DAX will also provide “unique access to various IBM and IBM Research datasets.” Big Blue has pledged to publish further datasets, and said “The datasets on DAX will integrate with IBM Cloud and AI services as appropriate.”<br></p>



<p>There are other ways to source data and models, with IBM’s announcement referencing GitHub and Kaggle, while the PyTorch hub launched a model repository earlier this year.<br></p>



<p>IBM claimed DAX would be “unique in its high level of quality and curation”, as it would help developers build “end-to-end” deep learning workflows, and allow “developers to consume open data with confidence under clearly defined open data licenses.”<br></p>



<p>That might sound rather dull to developers used to skunkworks-like conditions, but as machine learning creeps across the enterprise, compliance and ethical practices become a bigger concern.<br></p>



<p>“The CODAIT team’s goal is to make it straightforward to use DAX and MAX assets in conjunction with IBM AI products as well as other hybrid, multicloud AI tooling,” the team said, which will presumably be a relief for those developers who don’t want to actually lock themselves into IBM’s way of machine learning.<br></p>



<p>As of today, there are eight datasets on the exchange, including IBM’s Contracts Proposition Bank, which features text from IBM’s contracts, the NOAA Weather Data set for JFK Airport,  and a set containing 100 randomly sampled discussion threads from Ubuntu Forums.</p>
<p>The post <a href="https://www.aiuniverse.xyz/big-blue-opens-up-hub-for-machine-learning-datasets/">Big Blue opens up hub for machine learning datasets</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/big-blue-opens-up-hub-for-machine-learning-datasets/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>
