<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Databases Archives - Artificial Intelligence</title>
	<atom:link href="https://www.aiuniverse.xyz/tag/databases/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.aiuniverse.xyz/tag/databases/</link>
	<description>Exploring the universe of Intelligence</description>
	<lastBuildDate>Thu, 11 Mar 2021 06:51:54 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>
	<item>
		<title>Yale study shows limitations of applying artificial intelligence to registry databases</title>
		<link>https://www.aiuniverse.xyz/yale-study-shows-limitations-of-applying-artificial-intelligence-to-registry-databases/</link>
					<comments>https://www.aiuniverse.xyz/yale-study-shows-limitations-of-applying-artificial-intelligence-to-registry-databases/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Thu, 11 Mar 2021 06:51:52 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[Applying]]></category>
		<category><![CDATA[Databases]]></category>
		<category><![CDATA[limitations]]></category>
		<category><![CDATA[registry]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=13393</guid>

					<description><![CDATA[<p>Source &#8211; https://medicine.yale.edu/ Artificial intelligence will play a pivotal role in the future of health care, medical experts say, but so far, the industry has been unable <a class="read-more-link" href="https://www.aiuniverse.xyz/yale-study-shows-limitations-of-applying-artificial-intelligence-to-registry-databases/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/yale-study-shows-limitations-of-applying-artificial-intelligence-to-registry-databases/">Yale study shows limitations of applying artificial intelligence to registry databases</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source &#8211; https://medicine.yale.edu/</p>



<p>Artificial intelligence will play a pivotal role in the future of health care, medical experts say, but so far, the industry has been unable to fully leverage this tool. A Yale study has illuminated the limitations of these analytics when applied to traditional medical databases — suggesting that the key to unlocking their value may be in the way datasets are prepared.</p>



<p>Machine learning techniques are well-suited for processing complex, high-dimensional data or identifying nonlinear patterns, which provide researchers and clinicians with a framework to generate new insights. Achieving the potential of artificial intelligence will require improving the data quality of electronic health records (EHR).</p>



<p>“Our study found that advanced methods that have revolutionized predictions outside healthcare did not meaningfully improve prediction of mortality in a large national registry. These registries that rely on manually abstracted data within a restricted number of fields may, therefore, not be capturing many patient features that have implications for their outcomes,” said Rohan Khera, MD, MS, the first author of the new study published in JAMA Cardiology. “We believe that the next frontier for improving clinical prediction may be the application of these methods to the high-dimensional granular data collected in the EHR.”</p>



<p>The authors used the American College of Cardiology’s (ACC) Chest Pain-MI Registry from 2011 to 2016, which includes nearly 1 million patients hospitalized for an acute myocardial infarction (AMI) or heart attack across more than 1,000 U.S. hospitals. The researchers applied three different machine learning models to predict death after hospitalization and only observed marginal gains over models using the traditional logistic regression, applied to these nationwide data.</p>



<p>Clinical registries such as the Chest Pain-MI Registry have been the mainstay for assessing patient outcomes across many hospitals through standardized data collection. These registries can advance clinical understanding and knowledge but are less suited at complex data collection and abstraction. To infer additional insights will require rethinking how to aggregate novel digital data streams that are being generated at most U.S. hospitals, the researchers said.</p>



<p>The study also underscores that while some methods are more efficient or transparent, the clinical value of machine learning will be determined by data collection and processing.</p>



<p>“The clinical adoption of machine learning will depend on whether it delivers better information – and that may importantly depend on the data that are used,” said Harlan Krumholz, MD, SM, director of the Center for Outcomes Research and Evaluation (CORE) at Yale and senior author of the study.</p>



<p>The research team included clinicians and scientists from across several Yale departments, including Robert McNamara MD, MHS, Nihar Desai, MD, MPH, Chenxi Huang, PhD, and Bobak Jack Mortazavi, PhD.</p>



<p>The research was supported by the American College of Cardiology Foundation.</p>
<p>The post <a href="https://www.aiuniverse.xyz/yale-study-shows-limitations-of-applying-artificial-intelligence-to-registry-databases/">Yale study shows limitations of applying artificial intelligence to registry databases</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/yale-study-shows-limitations-of-applying-artificial-intelligence-to-registry-databases/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>The CAP theorem, and how it applies to microservices</title>
		<link>https://www.aiuniverse.xyz/the-cap-theorem-and-how-it-applies-to-microservices/</link>
					<comments>https://www.aiuniverse.xyz/the-cap-theorem-and-how-it-applies-to-microservices/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Fri, 11 Dec 2020 05:12:26 +0000</pubDate>
				<category><![CDATA[Microservices]]></category>
		<category><![CDATA[application]]></category>
		<category><![CDATA[Databases]]></category>
		<category><![CDATA[Developers]]></category>
		<category><![CDATA[Google]]></category>
		<category><![CDATA[Microservice]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=12411</guid>

					<description><![CDATA[<p>Source: searchapparchitecture.techtarget.com It&#8217;s not unusual for developers and architects who jump into microservices for the first time to &#8220;want it all&#8221; in terms of performance, uptime and <a class="read-more-link" href="https://www.aiuniverse.xyz/the-cap-theorem-and-how-it-applies-to-microservices/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/the-cap-theorem-and-how-it-applies-to-microservices/">The CAP theorem, and how it applies to microservices</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: searchapparchitecture.techtarget.com</p>



<p>It&#8217;s not unusual for developers and architects who jump into microservices for the first time to &#8220;want it all&#8221; in terms of performance, uptime and resiliency. After all, these are the goals that drive a software team&#8217;s decision to pursue this type of architecture design. The unfortunate truth is that trying to create an application that perfectly embodies all of these traits will eventually steer them to failure.</p>



<p>This phenomenon is summed up in something called the CAP theorem, which states that a distributed system can deliver only two of the three overarching goals of microservices design: consistency, availability and partition tolerance. According to CAP, not only is it impossible to &#8220;have it all&#8221; &#8212; you may even struggle to deliver more than one of these qualities at a time.</p>



<p>When it comes to microservices, the CAP theorem seems to pose an unsolvable problem. Which of these three things can you afford to trade away? However, the essential point is that you don&#8217;t have a choice. You&#8217;ll have to face that fact when it comes to your design stage, and you&#8217;ll need to think carefully about the type of application you&#8217;re building, as well as its most essential needs.</p>



<p>In this article, we&#8217;ll review the basics of how the CAP theorem applies to microservices, and then examine the concepts and guidelines you can follow when it&#8217;s time to make a decision.</p>



<h3 class="wp-block-heading">CAP theory and microservices</h3>



<p>Let&#8217;s start by reviewing the three qualities CAP specifically refers to:</p>



<ul class="wp-block-list"><li><strong>Consistency</strong> means that all clients see the same data at the same time, no matter the path of their request. This is critical for applications that do frequent updates.</li><li><strong>Availability</strong> means that all functioning application components will return a valid response, even if they are down. This is particularly important if an application&#8217;s user population has a low tolerance for outages (such as a retail portal).</li><li><strong>Partition</strong> <strong>tolerance</strong> means that the application will operate even during a network failure that results in lost or delayed messages between services. This comes into play for applications that integrate with a large number of distributed, independent components.</li></ul>



<p>Databases often sit at the center of the CAP problem. Microservices often rely on NoSQL databases, since they&#8217;re designed to scale horizontally and support distributed application processes. And, partition tolerance is a &#8220;must have&#8221; in these types of systems because they are so sensitive to failure.</p>



<p>You can certainly design these kinds of databases for consistency and partition tolerance, or even for availability and partitioning. But designing for consistency and availability just isn&#8217;t an option.</p>



<h2 class="wp-block-heading">The PACELC theorem</h2>



<p>This prohibitive requirement for partition-tolerance in distributed systems gave rise to what is known as the PACELC theorem, a sibling to the CAP theorem. The acronym PACELC stands for &#8220;if partitioned, then availability and consistency; else, latency and consistency.&#8221; In other words: If there is a partition, the distributed system must trade availability for consistency; if not, the choice is between latency and consistency.</p>



<p>Designing your applications specifically to avoid partitioning problems in a distributed system will force you to sacrifice either availability or user experience to retain operational consistency. However, the key term here is &#8220;operational&#8221; &#8212; while latency is a primary concern during normal operations, a failure can quickly make availability the overall priority. So, why not create models for both scenarios?</p>



<p>It may help to frame CAP concepts in both &#8220;normal&#8221; and &#8220;fault&#8221; modes, provided that faults in a distributed system are essentially inevitable. This enables you to create two database and microservices implementation models: one that handles normal operation, and another that kicks in during failures. For example, you can design your database to optimize consistency during a partition failure, and then continue to focus on mitigating latency during normal operation.</p>



<h3 class="wp-block-heading">Applying PACELC to microservices</h3>



<p>If we use PACELC rather than &#8220;pure CAP&#8221; to define databases, we can classify them according to how they make the trades.</p>



<ul class="wp-block-list"><li>In PACELC terms, relational database management systems and NoSQL databases that implement ACID (atomicity, consistency, isolation, urability) are designed to assure consistency, classifying them as PC/EC. Typical business applications, like human resources apps and ticketing systems, will likely use this model, particularly if there are multiple users using different component instances. Google&#8217;s Bigtable database is a good example of this.</li><li>In-memory databases like MongoDB and Hazelcast fit into a PA/EC model, which is best suited for things like e-commerce apps, which need high availability even during network or component failures.</li><li>Real-time applications, such as IoT systems, fit into the PC/EL model that databases like PNUTS provide. This is the case in any application where consistency across replications is critical.</li><li>Database systems based on the PA/EL model, such as Dynamo and Cassandra, are best for real-time applications that don&#8217;t experience frequent updates, since consistency will be less of an issue.</li></ul>



<h3 class="wp-block-heading">Know the tradeoffs</h3>



<p>The bottom line is this: It&#8217;s critical to know exactly what you&#8217;re trading in a PACELC-guided application, and to know which scenarios call for which sacrifice. Here are three things to remember when making your decision:</p>



<ul class="wp-block-list"><li><strong>Consistency</strong>&nbsp;is most valuable where many users update the same data elements.</li><li><strong>Availability</strong>&nbsp;is critical for applications involving consumers (who get frustrated easily) and also for some IoT applications.</li><li><strong>Latency</strong>&nbsp;is most likely critical for real-time and&nbsp;<a href="https://internetofthingsagenda.techtarget.com/definition/Internet-of-Things-IoT">IoT</a>&nbsp;applications where processing delays must be kept to a minimum.</li></ul>



<p>Make your database choice wisely. Then, design your microservices workflows and framework to ensure you don&#8217;t compromise your goals.</p>
<p>The post <a href="https://www.aiuniverse.xyz/the-cap-theorem-and-how-it-applies-to-microservices/">The CAP theorem, and how it applies to microservices</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/the-cap-theorem-and-how-it-applies-to-microservices/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Bridging the gap between databases and data science</title>
		<link>https://www.aiuniverse.xyz/bridging-the-gap-between-databases-and-data-science/</link>
					<comments>https://www.aiuniverse.xyz/bridging-the-gap-between-databases-and-data-science/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Tue, 09 Jun 2020 06:44:46 +0000</pubDate>
				<category><![CDATA[Data Science]]></category>
		<category><![CDATA[Analytical Tools]]></category>
		<category><![CDATA[application]]></category>
		<category><![CDATA[data science]]></category>
		<category><![CDATA[Databases]]></category>
		<category><![CDATA[Developing]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=9394</guid>

					<description><![CDATA[<p>Source: universiteitleiden.nl Relational databases are used to store information or data in such a way that it preserves relations between the data. This property makes it a <a class="read-more-link" href="https://www.aiuniverse.xyz/bridging-the-gap-between-databases-and-data-science/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/bridging-the-gap-between-databases-and-data-science/">Bridging the gap between databases and data science</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: universiteitleiden.nl</p>



<p>Relational databases are used to store information or data in such a way that it preserves relations between the data. This property makes it a useful tool for data scientists. There is, however, a gap between the relational database research community and data scientists. This leads to inefficient use of databases in data science. PhD-student Mark Raasveldt tried to bridge the gap between the relational databases and data science. PhD defense 9 June 2020.</p>



<h4 class="wp-block-heading" id="integration-with-analytical-tools">Integration with analytical tools</h4>



<p>Most data scientists use analytical tools, such as R, Python and C/C++, for their research. These tools are difficult to integrate with current database systems, resulting in slow and cumbersome data analysis. ‘Data scientists have opted to reinvent database systems by developing a zoo of data management alternatives that perform similar tasks to classical database management systems, but have many of the problems that were solved in the database field decades ago,’ says Raasveldt.</p>



<p>‘The database research community has made tremendous strides in developing powerful database engines that allow for efficient analytical query processing.’ Raasveldt tried to combine these innovations in the database science with the analytical tools that are mostly used by data scientists. &nbsp;‘We investigate how we can facilitate efficient and painless integration of analytical tools and relational database management systems,’ says Raasveldt.</p>



<h4 class="wp-block-heading" id="large-datasets">Large datasets</h4>



<p>Another issue with the use of standard database systems in computer science is the size of the data that is handled. Most database systems are not optimized for large data sets and large-scale data analysis using remote servers. To optimize the database systems, there are three methods that can be considered.</p>



<p>‘We focus our investigation on the three primary methods for database-client integration: client-server connections, in-database processing and embedding the database inside the client application,’ Raasveldt explains. For every method, he studied the implementations in existing database systems and he evaluated how efficient they are for the large datasets and workloads that are common in data science.</p>



<h4 class="wp-block-heading" id="duckdb">DuckDB</h4>



<p>Raasveldts final result was a new data management system, called DuckDB, that was purpose-built for efficient and painless integration with R and Python (and other analytical tools). This management system is meant to be used as a mature database system that is not only used for research purposes.</p>



<p>‘In DuckDB we take all the lessons that we have learned investigating database-client integrations and create an easy-to-use and highly efficient embedded database.’ Raasveldt will continue his work as a postdoc at the CWI, where he will work on further developing DuckDB.</p>
<p>The post <a href="https://www.aiuniverse.xyz/bridging-the-gap-between-databases-and-data-science/">Bridging the gap between databases and data science</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/bridging-the-gap-between-databases-and-data-science/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>NLM Leverages Data, Text Mining to Sharpen COVID-19 Research Databases</title>
		<link>https://www.aiuniverse.xyz/nlm-leverages-data-text-mining-to-sharpen-covid-19-research-databases/</link>
					<comments>https://www.aiuniverse.xyz/nlm-leverages-data-text-mining-to-sharpen-covid-19-research-databases/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Wed, 13 May 2020 13:25:39 +0000</pubDate>
				<category><![CDATA[Data Mining]]></category>
		<category><![CDATA[data mining]]></category>
		<category><![CDATA[data science]]></category>
		<category><![CDATA[Databases]]></category>
		<category><![CDATA[Research]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=8745</guid>

					<description><![CDATA[<p>Sojurce: governmentciomedia.com The National Library of Medicine is leveraging its database resources and artificial intelligence capabilities to rapidly provide COVID-19 literature and resources to researchers and scientists <a class="read-more-link" href="https://www.aiuniverse.xyz/nlm-leverages-data-text-mining-to-sharpen-covid-19-research-databases/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/nlm-leverages-data-text-mining-to-sharpen-covid-19-research-databases/">NLM Leverages Data, Text Mining to Sharpen COVID-19 Research Databases</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Sojurce: governmentciomedia.com</p>



<p>The National Library of Medicine is leveraging its database resources and artificial intelligence capabilities to rapidly provide COVID-19 literature and resources to researchers and scientists as the world races to understand and respond to the pandemic.</p>



<p>The White House in March tapped NLM, under the National Institutes of Health, to join a public-private partnership&nbsp;called the COVID-19 Open Research Dataset (CORD-19) to develop data-mining techniques that could help the science community answer critical questions pertaining to COVID-19. Leveraging its existing infrastructure and establishing processes for content submission, NLM has quickly brought access to COVID-19 literature and clinical trial content on its PubMed Central (PMC) and ClinicalTrials.gov databases.</p>



<p>“As of May 1, about 46,000 articles had been deposited by publishers to PMC or updated in PMC to have a license that allows for text and data-mining, of which more than 5,600 articles specifically focus on the current novel coronavirus,” said NLM National Center for Biotechnology Information Acting Director Stephen Sherry. “Some 49 publishers are now included in the PMC COVID-19 initiative.”</p>



<p>Within the first few weeks since launching the project, PMC saw significant COVID-19 download and data-sharing rates, said PMC Program Manager Kathryn Funk in an NIH webinar. As part of the project, Funk&#8217;s team worked to standardize submission data in a machine-readable format.</p>



<p>“The early results have been encouraging,” Funk said. “Articles in the Public Health Emergency Collection and PMC were retrieved more than 2 million times in the first two to three weeks of the initiative, and the coordinating dataset has been downloaded more than 75,000 times at this time. It’s our hope that through expanded access and machine learning, NLM will be able to help accelerate scientific research on COVID-19.”</p>



<p>NLM has also leveraged ClinicalTrials.gov’s existing infrastructure to scale up and provide quick access to information about trials related to COVID-19. Teams&nbsp;conducting trials around the world can submit standardized and structured information about their&nbsp;trials directly through an online submission portal called the Protocol Registration and Results system, where trial information is then posted to ClinicalTrials.gov within a couple of days of initial submission,&nbsp;Sherry said.</p>



<p>The data standardization and structure are critical to enabling AI technologies like machine learning and natural-language processing, which can help users more effectively mine and analyze the databases’ resources and literature to generate knowledge and support research that assist in responding to COVID-19, Sherry said.</p>



<p>“ClinicalTrials.gov also leverages NLM resources such as the biomedical vocabularies and standards integrated in the unified Medical Language System (UMLS) to support its search capabilities,” Sherry said, citing the database’s complete list of registered COVID-19 studies. “Users can filter the search results further by different study design characteristics, recruitment status, location information and other factors to identify trials of interest. All of these search capabilities are also available through the ClinicalTrials.gov API.”</p>



<p>Sherry likened the ClinicalTrials.gov infrastructure as an “information scaffold” for discovering information about clinical trials, as the platform applies unique identifiers called National Clinical Trial (NCT) numbers to each trial so that individuals can label and identify trials.</p>



<p>“As a result, different resources with information about particular trials can be linked and discovered through the use of unique NCT numbers, [such as] ClinicalTrials.gov records, press releases, journal articles, protocol document[s], informed consent forms, systematic reviews, reports, regulatory documents, individual participant-level data,” Sherry said.</p>



<p>Creating an open data repository ecosystem like ClinicalTrials.gov requires integrating different data contributors in a way that enable interoperability and usability of data, said NIH Director of Data Science Strategy Susan Gregurick, who helped establish the agency&#8217;s data science office in 2018.</p>



<p>“NIH strongly encourages open-access, data-sharing repositories as your first go-to choice when you’re looking for a repository to share your data and your information,” Gregurick said during an agency webinar last month.</p>



<p>Although NLM had already pledged to modernize its databases, support data-driven science, collaborate with relevant stakeholders and build a future-ready workforce in its strategic plan, such as the multi-year effort to overall modernize ClinicalTrials.gov, COVID-19 has sparked a number of new data-backed initiatives and digital resources around COVID-19, said Sherry and Gregurick.</p>



<p>These are&nbsp;not just on PMC and ClinicalTrials.gov, but also on new platforms and resources, including:</p>



<ul class="wp-block-list"><li><strong>LitCovid</strong>, a COVID-19-specific open-resource literature hub that curates and disseminates a constantly growing comprehensive collection of international research papers relevant to public health. “This resource builds on NLM research to develop new approaches to locating and indexing the literature related to COVID-19, including a text classification algorithm for screening and ranking relevant documents, topic modeling for suggesting relevant research categories and information extraction for obtaining geographic locations found in the abstract,” Sherry said.</li><li>COVID-19 genetic sequence information additions to <strong>GenBank</strong>, the world’s largest genetic sequence database that released the first COVID-19 sequence to the public Jan. 12 and the first sequence collected in America in collaboration with the Centers for Disease Control and Prevention Jan. 25. “As of April 9, we have 579 SARS-CoV-2 sequences from 26 different countries publicly available,” Sherry said, adding that NLM has create a data hub on GenBank for individuals to search, retrieve and analyze COVID-19 sequences that have been submitted.</li><li>The <strong>Sequence Read Archive</strong>, an 14-petabyte archive of high-throughput genetic sequence data that as of February became available on commercial cloud-computing platforms, which Sherry said significantly expanded the discovery potential of the data to help identify mutational patterns and inform drug and vaccine development.</li><li><strong>PubChem</strong>, an open chemistry database that contains compounds used in COVID-19 clinical trials and found in COVID-19-related protein database structures.</li></ul>
<p>The post <a href="https://www.aiuniverse.xyz/nlm-leverages-data-text-mining-to-sharpen-covid-19-research-databases/">NLM Leverages Data, Text Mining to Sharpen COVID-19 Research Databases</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/nlm-leverages-data-text-mining-to-sharpen-covid-19-research-databases/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Object Stores Starting to Look Like Databases</title>
		<link>https://www.aiuniverse.xyz/object-stores-starting-to-look-like-databases/</link>
					<comments>https://www.aiuniverse.xyz/object-stores-starting-to-look-like-databases/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Fri, 17 Apr 2020 10:06:32 +0000</pubDate>
				<category><![CDATA[Google Cloud AutoML]]></category>
		<category><![CDATA[data analysts]]></category>
		<category><![CDATA[Databases]]></category>
		<category><![CDATA[Google Cloud]]></category>
		<category><![CDATA[Machine learning]]></category>
		<category><![CDATA[Microsoft]]></category>
		<category><![CDATA[software]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=8239</guid>

					<description><![CDATA[<p>Source: Don’t look now, but object stores – those vast repositories of data sitting behind an S3 API – are beginning to resemble databases. They’re obviously still <a class="read-more-link" href="https://www.aiuniverse.xyz/object-stores-starting-to-look-like-databases/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/object-stores-starting-to-look-like-databases/">Object Stores Starting to Look Like Databases</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: </p>



<p>Don’t look now, but object stores – those vast repositories of data sitting behind an S3 API – are beginning to resemble databases. They’re obviously still separate categories today, but as the next-generation data architecture takes shape to solve emerging real-time data processing and machine learning challenges, the lines separating things like object stores, databases, and streaming data frameworks will begin to blur.</p>



<p>Object stores have become the primary repository for the vast amounts of less-structured data that’s generated today. Organizations clearly are using object-based data lakes in the cloud and on premise to store unstructured data, like images and video. But they’re also using them to store many of the other types of data, like sensor and log data from mobile and IoT devices, that the world is generating.</p>



<p>The object store is becoming a general purpose data repository, and along the way it’s getting closer to the most popular data workloads, including SQL-based analytics and machine learning. The folks at object storage software vendor Cloudian are moving their wares in that direction too, according to Cloudian CTO Gary Ogasawara.</p>



<p>“We’re moving more and more to that,” Ogasawara tells Datanami. “If you can combine the best of both worlds – have the huge capacity of an object store and the advanced query capability of an SQL-type database – that would be the ideal. That’s what people are really asking for.”</p>



<h3 class="wp-block-heading">Past Is Prologue</h3>



<p>We’ve seen this film before. When Apache Hadoop was the hot storage repository for big data (really, less-structured data), the first big community efforts was to develop a relational database for it. That way, data analysts with existing SQL skills – as well as BI applications expecting SQL data – would be able to leverage it without extensive retraining. And besides, after running less-structured data through MapReduce jobs, you  needed a place to put the structured data. A database is that logical place.</p>



<p>This led to the creation of Apache Hive out of Facebook, and the community followed with a host of other SQL-on-Hadoop engines (or relational databases, if you like), including Apache Impala, Presto, and Spark SQL, among others. Of course, Hadoop’s momentum fizzled over the past few years, in part due to the rise of S3 from Amazon Web Services and other cloud-based object storage systems, notably Azure BLOB Storage from Microsoft and Google Cloud Storage, which are universally more user-friendly than Hadoop, if not always cheaper.</p>



<p>In the cloud, users are presented with a wide range of specialty storage repositories and processing engines for SQL and machine learning. On the SQL front, you have Amazon RedShift, Azure Data Warehouse, and Google BigQuery. On top of these “native” offerings, the big data community has adapted many existing and popular analytics databases, including Teradata, Vertica, and others, to work with S3 and other object stores with an S3-compatible API.</p>



<p>The same goes for machine learning workloads. Once the data is in S3 (or Blob Store or Google Cloud Storage), it’s a relatively simple manner to use that data to build and train machine learning models in SageMaker, Azure Machine Learning, or Google Cloud AutoML. With the rise of the cloud, every member of the big data and machine learning community has moved to support the cloud, and with it object storage systems.</p>



<p>As the cloud’s momentum grows, S3 has become the defacto data access standard for the next generation of applications, from SQL analytics and machine learning to more traditional apps too. For many new applications, data is simply expected to be stored in an object storage system, and developers expect to be able to access that data over the S3 API.</p>



<h3 class="wp-block-heading">A Hybrid Architecture</h3>



<p>But of course, not all new applications will live on the cloud with ready access to petabytes of data and gigaflops of computing power. In fact, with the rise of 5G networks and the explosion of smart devices on the Internet of Things (IoT), the physical world is the next frontier for computing, and that’s changing the dynamics for data architects who are trying to foresee new trends.</p>



<p>At Cloudian, Ogasawara and his team are working on adapting its HyperStore object storage architecture to fit into the emerging edge-and-hub computing model. One of the examples he uses is the case of an autonomous car. With cameras, LIDAR, and other sensors, each self-driving car generates terabytes worth of data every day, and petabytes per year.</p>



<p>“That is all being generated at the edge,” he says. “Even with a 5G network, you will never be able to transmit all that data to somewhere else for analyses. You have to push that storage and processing as close to the edge as possible.”</p>



<p>Cloudian is currently working on developing a version of HyperStore that sits on the edge. In the self-driving car example, the local version of HyperStore would run right on the car and assist with storing and processing data coming off the sensors in real time. This computing environment would constitute a fast “inner loop,” Ogasawara says.</p>



<p>“But then you have a slower outer loop that’s also collecting data, and that includes the hub where the large, vast data lake resides in object storage,” he continues. “Here you can do more extensively training of ML models, for example, and then push that kind of metadata out to the edge, where it’s essentially a compiled version of your model that can be used very quickly.”</p>



<p>In the old days, object stores resembled relatively simple (and nearly infinitely scalable) key-value stores. But to support future use cases — like self-driving cars as well as weather modeling and genomics — the object store needs to learn new tricks, like how to stream data in and intelligently filter it so that only a subset of the most important data is forwarded from the edge to the hub.</p>



<p>To that end, Cloudian is working on a new project that will incorporate analytics capabilities. It has a working name of the Hyperstore Analytics Platform, the project would incorporate frameworks like Spark or TensorFlow to assist with the intelligent streaming and processing of data. A beta was expected by the end of the year (at least that was the timeline that Ogasawara shared in early March before the COVID-19 lockdown.)</p>



<h3 class="wp-block-heading">Object’s Evolution</h3>



<p>Cloudian is not the only object storage vendor looking at how to evolve its product to adapt to emerging data challenges. In fact, its not just object storage vendors who are trying to tackle the probolem.</p>



<p>The folks at Confluent have adapted their Kafka-based stream processing technologies (which excel at processing event data) to work more like a database, which is good at managing stateful data. MinIO has SQL extensions that allow its object store to function like a database. NewSQL database vendor MemSQL has long had hooks for Kafka that allow it to process large amounts of real-time data. The in-memory data grid (IMDG) vendors are doing similar things for processing new event data within the context of historic, stateful data. And let’s not even get into how the event meshes are solving this problem.</p>



<p>According to Ogasawara, adapting Cloudian’s HyperStore offering is a logical way to tackle today’s emerging data challenges. “You’ve done very well at building this storage infrastructure,” he says. “Now, how do you make the data usable and consumable? It’s really about providing better access APIs to get to that data, and almost making the object storage more intelligent.”</p>



<p>Object stores are moving beyond their initial use case, which was reading, writing, and deleting data at massive scale. Now customers are pushing object storage vendors to support more advanced workflows, including complex machine learning workflows. That will most likely require an extension to the S3 API (something that Cloudian has brought up with AWS, but without much success).</p>



<p>“How do you look into those objects? Those types of APIs need more and more [capabilities],” Ogasawara says. “And even letting AI or machine learning-type workflows, doing things like a sequence of operations — those types of language constructs, everyone is starting to look at and trying to figure out how do we make it easier for users and customers to make that data analysis possible.”</p>
<p>The post <a href="https://www.aiuniverse.xyz/object-stores-starting-to-look-like-databases/">Object Stores Starting to Look Like Databases</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/object-stores-starting-to-look-like-databases/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>5 core components of microservices architecture</title>
		<link>https://www.aiuniverse.xyz/5-core-components-of-microservices-architecture/</link>
					<comments>https://www.aiuniverse.xyz/5-core-components-of-microservices-architecture/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Sat, 11 Apr 2020 10:49:37 +0000</pubDate>
				<category><![CDATA[Microservices]]></category>
		<category><![CDATA[applications]]></category>
		<category><![CDATA[Architecture]]></category>
		<category><![CDATA[Databases]]></category>
		<category><![CDATA[Developers]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=8122</guid>

					<description><![CDATA[<p>Source: A microservices architecture &#8212; as the name implies &#8212; is a complex coalition of code, databases, application functions and programming logic spread across servers and platforms. <a class="read-more-link" href="https://www.aiuniverse.xyz/5-core-components-of-microservices-architecture/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/5-core-components-of-microservices-architecture/">5 core components of microservices architecture</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: </p>



<p> A microservices architecture &#8212; as the name implies &#8212; is a complex coalition of code, databases, application functions and programming logic spread across servers and platforms. Certain fundamental components of a microservices architecture bring all these entities together cohesively across a distributed system. </p>



<p>In this article, we review five key components of microservices architecture that developers and application architects need to understand if they plan to take the distributed service route. Start with microservices themselves, then learn about service mesh as an additional layer, app management via service discovery, container-based deployment and API gateways.</p>



<h3 class="wp-block-heading">1. Microservices</h3>



<p>Microservices make up the foundation of a microservices architecture. The term illustrates the method of breaking down an application into generally small, self-contained services, written in any language, that communicate over lightweight protocols. With independent microservices, software teams can implement iterative development processes, as well as create and upgrade features flexibly.</p>



<p>Teams need to decide the proper size for microservices, keeping in mind that an overly granular collection of too-segmented services creates high overhead and management needs. Developers should thoroughly decouple services in order to minimize dependencies between them and promote service autonomy. And use lightweight communication mechanisms like REST and HTTP.</p>



<h3 class="wp-block-heading">2. Containers</h3>



<p>Containers are units of software that package services and their dependencies, maintaining a consistent unit through development, test and production. Containers are not necessary for microservices deployment, nor are microservices needed to use containers. However, containers can potentially improve deployment time and app efficiency in a microservices architecture more so than other deployment techniques, such as VMs.</p>



<p>The major difference between containers and VMs is that containers can share an OS and middleware components, whereas each VM includes an entire OS for its use. By eliminating the need for each VM to provide an individual OS for each small service, organizations can run a larger collection of microservices on a single server.</p>



<p>The other advantage of containers is their ability to deploy on-demand without negatively impacting application performance. Developers can also replace, move and replicate them with fairly minimal effort. The independence and consistency of containers is a critical part of scaling certain pieces of a microservices architecture &#8212; according to workloads &#8212; rather than the whole application. It also supports the ability to redeploy microservices in a failure.</p>



<p>Docker, which started as an open-source platform for container management, is one of the most recognizable providers in the container space. However, Docker&#8217;s success caused a large tooling ecosystem to evolve around it, spawning popular container orchestrators like Kubernetes.</p>



<h3 class="wp-block-heading">3. Service mesh</h3>



<p>In a microservices architecture, the service mesh creates a dynamic messaging layer to facilitate communication. It abstracts the communication layer, which means developers don&#8217;t have to code in inter-process communication when they create the application.</p>



<p>Service mesh tooling typically uses a sidecar pattern, which creates a proxy container that sits beside the containers that have either a single microservice instance or a collection of services. The sidecar routes traffic to and from the container, and directs communication with other sidecar proxies to maintain service connections.</p>



<p>Two of today&#8217;s most popular service mesh options are Istio, a project that Google launched alongside IBM and Lyft, and Linkerd, a project under the Cloud Native Computing Foundation. Both Istio and Linkerd are tied to Kubernetes, though they feature notable differences in areas such as support for non-container environments and traffic control capabilities.</p>



<h3 class="wp-block-heading">4. Service discovery</h3>



<p>Whether it&#8217;s due to changing workloads, updates or failure mitigation, the number of microservice instances active in a deployment fluctuate. It can be difficult to keep track of large numbers of services that reside in distributed network locations throughout the application architecture.</p>



<p>Service discovery helps service instances adapt in a changing deployment, and distribute load between the microservices accordingly. The service discovery component is made up of three parts:</p>



<ul class="wp-block-list"><li>A&nbsp;<strong>service provider</strong>&nbsp;that originates service instances over a network;</li><li>A&nbsp;<strong>service registry</strong>, which acts as a database that stores the location of available service instances; and</li><li>A&nbsp;<strong>service consumer</strong>, which retrieves the location of a service instance from the registry, and then communicates with that instance.</li></ul>



<p>Service discovery also consists of two major discovery patterns:</p>



<ul class="wp-block-list"><li>A&nbsp;<strong>client-side discovery pattern</strong>&nbsp;searches the service registry to locate a service provider, selects an appropriate and available service instance using a load balancing algorithm, and then makes a request.</li><li>In a&nbsp;<strong>server-side discovery pattern</strong>, the router searches the service registry and, once the applicable service instance is found, forwards the request accordingly.</li></ul>



<p>Data residing in the service registry should always be current, so that related services can find their related service instances at runtime. If the service registry is down, it will hinder all the services, so enterprises typically use a distributed database, such as Apache ZooKeeper, to avoid regular failures.</p>



<h3 class="wp-block-heading">5. API gateway</h3>



<p>Another important component of a microservices architecture is an API gateway. API gateways are vital for communication in a distributed architecture, as they can create the main layer of abstraction between microservices and the outside clients. The API gateway will handle a large amount of the communication and administrative roles that typically occur within a monolithic application, allowing the microservices to remain lightweight. They can also authenticate, cache and manage requests, as well as monitor messaging and perform load balancing as necessary.</p>



<p>Additionally, an API gateway can speed up communication between microservices and clients by standardizing messaging protocols translation and freeing both the client and the service from the task of translating requests written in unfamiliar formats. Most API gateways will also provide built-in security features, which means they can manage authorization and authentication for microservices, as well as track incoming and outgoing requests to identify any possible intrusions.</p>



<p>There are a wide array of API gateway options on the market to choose from, both from proprietary cloud platform providers like Amazon and Microsoft and open source providers such as Kong and Tyk.</p>
<p>The post <a href="https://www.aiuniverse.xyz/5-core-components-of-microservices-architecture/">5 core components of microservices architecture</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/5-core-components-of-microservices-architecture/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Why Big Data Is a Big Opportunity for Employment</title>
		<link>https://www.aiuniverse.xyz/why-big-data-is-a-big-opportunity-for-employment/</link>
					<comments>https://www.aiuniverse.xyz/why-big-data-is-a-big-opportunity-for-employment/#comments</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Mon, 13 May 2019 06:03:21 +0000</pubDate>
				<category><![CDATA[Big Data]]></category>
		<category><![CDATA[algorithms]]></category>
		<category><![CDATA[Big data]]></category>
		<category><![CDATA[Business]]></category>
		<category><![CDATA[data mining]]></category>
		<category><![CDATA[data science]]></category>
		<category><![CDATA[Databases]]></category>
		<category><![CDATA[DSA]]></category>
		<category><![CDATA[Programming]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=3487</guid>

					<description><![CDATA[<p>Source:- insidebigdata.com As data has transitioned from “nice to have” to a full-blown commodity, we’ve seen demand for data science and analytics (DSA) talent soar — along with <a class="read-more-link" href="https://www.aiuniverse.xyz/why-big-data-is-a-big-opportunity-for-employment/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/why-big-data-is-a-big-opportunity-for-employment/">Why Big Data Is a Big Opportunity for Employment</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>Source:- insidebigdata.com</p>
<p>As data has transitioned from “nice to have” to a full-blown commodity, we’ve seen demand for data science and analytics (DSA) talent soar — along with opportunities for job-seekers and companies alike.</p>
<p>This is such a critical moment for data science, in fact, that industries across the board are bracing for talent shortages. Here’s why big data is such a massive opportunity — and one that needs the backing of educators at every level.</p>
<p><strong>The Data Science Skills Gap</strong></p>
<p>There are multiple data- and computer science-related skill areas which experts expect to remain in high demand throughout 2019 and beyond. These include:</p>
<ul>
<li>Programming languages</li>
<li>Data visualization</li>
<li>AI and machine learning</li>
<li>Data mining</li>
<li>Quantitative analysis</li>
<li>Algorithms</li>
<li>Databases and data structures</li>
</ul>
<p>There are more, too. But what about job numbers?</p>
<p>One 2019 report from Indeed indicated a 344% rise in the demand for data science specialists since 2013 and a 29% year-over-year increase. The U.S. Bureau of Labor Statistics expects the operations research analysis and computer research information science fields to grow by 27% and 19%, respectively, by 2026, which they deem “much faster than average” growth.</p>
<p>Other job websites corroborate these numbers and note that it isn’t just tech companies looking for data scientists — job listings span the gamut of industries. With good reason, too. Data is the new oil, as they say — and it powers every corner of commerce and human enterprise.</p>
<p><strong>Big Data Is Big Business (And Big Opportunity)</strong></p>
<p>For job-seekers everywhere, the implications are clear — if you have a head for numbers and an eye for detail, then big data and data science could be extremely lucrative and stable career tracks.</p>
<p>The fact is, big data has become a competitive advantage among companies in virtually every industry, from health care to logistics to supply chain management to merchandising analysis to marketing campaigns.</p>
<p>Data fuels companies large and small and keeps them fed with historical and real-time insights upon which to build business decisions, launch newer and more personalized products, engage in expansions and acquisitions, fine-tune outreach and marketing, and much more.</p>
<p>But the problem, according to Michael Chui of the McKinsey Institute, is that “it’s super hard to find the right talent.” Moreover, this lack of talent results in companies not fully understanding the opportunities that big data represents in the first place, such as clearer customer insights and leaner operations. In other words, it’s a catch-22 — we need more big data expertise to prove the value of big data expertise.</p>
<p>Nevertheless, the word is getting out. Data science is “sexy” now, and smaller and mid-sized companies are adding their own slate of job listings alongside big names like Google, Microsoft and Apple. And the arms race for big data talent has placed such professionals in an incredible position to market themselves and find a company and career path that truly suits them.</p>
<p>So the question becomes — What’s the best way to close this talent gap and connect employers with qualified and motivated candidates?</p>
<p><strong>How to Cultivate Big Data Talent</strong></p>
<p>In a report called “The Quant Crunch,” IBM lays out the scope of the problem and provides several ways industry and educational entities can marshal their forces to address the impending shortage of qualified DSA job applicants.</p>
<p>First up is the challenge of communicating to K-12 students, college students and those looking for new opportunities. Since data science is everywhere, these jobs are everywhere too. IBM identified the following industries as those with the highest demand for DSA talent. The percentages express the proportion of company job openings which fall into the DSA category:</p>
<ul>
<li>Finance and Insurance: 19%</li>
<li>Professional, Scientific, and Technical Services: 18%</li>
<li>Information services: 17%</li>
<li>Management of Companies and Enterprises: 13%</li>
<li>Manufacturing: 12%</li>
<li>Utilities: 10%</li>
<li>Wholesale Trade: 9%</li>
<li>Mining, Quarrying, and Oil and Gas Extraction: 9%</li>
<li>Public Administration: 7%</li>
</ul>
<p>From insurance to scientific research to retail to public service, there isn’t a single industry that isn’t being disrupted by this shift in the job market. Moreover, notes IBM, the human race generates 2.5 quintillion bytes of data daily. Netflix is ostensibly an entertainment company, but every decision they make relies on the apparent mountain of user data they sit on. Uber isn’t a transportation company — they are, in the words of their leaders, a technology company.</p>
<p>We’re looking at a whole new kind of economy here. So how can we prepare people for it? IBM has some ideas on that front:</p>
<ul>
<li><strong>Start Data Literacy Education Early</strong>: Organizations realize the benefits of adding big data talent to their value chain. But the general public might be lagging behind. IBM recommends adding “baseline data literacy” to our educational institutions as early as junior high. Doing so would prepare graduates with a firm understanding of the basics and ready them for additional, more finely-tuned training in an industry that appeals to them.</li>
<li><strong>More Focus on DSA Skills in Higher Learning</strong>: Higher learning institutions must, in turn, add new courses and degree tracks to their offerings — many already doing so. IBM observes that 42% of open data scientist positions require candidates with a graduate degree and 20% of positions require at least six years of hands-on experience. We need more colleges with classroom experiences and internship opportunities specifically tailored to big data.</li>
<li><strong>Lower Hiring Requirements</strong>: Recognizing the value of driven and self-motivated problem-solvers, some major employers like Microsoft and Hasbro, are lowering their experience and education requirements for job applicants or splitting existing roles which had such requirements into multiple entry-level positions which do not.</li>
</ul>
<p>Of course, even these recommendations aren’t enough — not when technology skills in general, and DSA skills specifically, change so regularly. eLearning has long been seen by many as one of the remedies for a rapidly changing economy and the emergence of brand-new disciplines. We already see an explosion in eLearning options for students even in K-12schools. This could help school districts more finely tune their course offerings, cast a wider net for high-quality educators and experiences, and stay more agile as the needs of the job market change further.</p>
<p>The importance of big data expertise couldn’t be more clear — for companies or for job-seekers. Thankfully, a more complete picture is emerging for how we can prepare many more people for a role in this exciting and quickly growing field.</p>
<p>The post <a href="https://www.aiuniverse.xyz/why-big-data-is-a-big-opportunity-for-employment/">Why Big Data Is a Big Opportunity for Employment</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/why-big-data-is-a-big-opportunity-for-employment/feed/</wfw:commentRss>
			<slash:comments>7</slash:comments>
		
		
			</item>
	</channel>
</rss>
