<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>data quality Archives - Artificial Intelligence</title>
	<atom:link href="https://www.aiuniverse.xyz/tag/data-quality/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.aiuniverse.xyz/tag/data-quality/</link>
	<description>Exploring the universe of Intelligence</description>
	<lastBuildDate>Wed, 19 Jun 2024 12:15:17 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>
	<item>
		<title>Overcoming the Challenges in Training Generative AI Models: A Comprehensive Guide</title>
		<link>https://www.aiuniverse.xyz/overcoming-the-challenges-in-training-generative-ai-models-a-comprehensive-guide/</link>
					<comments>https://www.aiuniverse.xyz/overcoming-the-challenges-in-training-generative-ai-models-a-comprehensive-guide/#respond</comments>
		
		<dc:creator><![CDATA[Maruti Kr.]]></dc:creator>
		<pubDate>Wed, 19 Jun 2024 12:14:59 +0000</pubDate>
				<category><![CDATA[AI]]></category>
		<category><![CDATA[AI model training]]></category>
		<category><![CDATA[AI scalability]]></category>
		<category><![CDATA[Computational resources]]></category>
		<category><![CDATA[data quality]]></category>
		<category><![CDATA[Ethical AI]]></category>
		<category><![CDATA[Generative AI]]></category>
		<category><![CDATA[Hyperparameter tuning]]></category>
		<category><![CDATA[Model complexity]]></category>
		<category><![CDATA[Model interpretability]]></category>
		<category><![CDATA[Training instability]]></category>
		<guid isPermaLink="false">https://www.aiuniverse.xyz/?p=18924</guid>

					<description><![CDATA[<p>Training generative AI models presents a variety of challenges and limitations. Key among these are: Data Quality and Quantity Computational Resources Model Complexity Training Stability and Performance <a class="read-more-link" href="https://www.aiuniverse.xyz/overcoming-the-challenges-in-training-generative-ai-models-a-comprehensive-guide/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/overcoming-the-challenges-in-training-generative-ai-models-a-comprehensive-guide/">Overcoming the Challenges in Training Generative AI Models: A Comprehensive Guide</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<figure class="wp-block-image size-large"><img fetchpriority="high" decoding="async" width="1024" height="512" src="https://www.aiuniverse.xyz/wp-content/uploads/2024/06/image-13-1024x512.png" alt="" class="wp-image-18925" srcset="https://www.aiuniverse.xyz/wp-content/uploads/2024/06/image-13-1024x512.png 1024w, https://www.aiuniverse.xyz/wp-content/uploads/2024/06/image-13-300x150.png 300w, https://www.aiuniverse.xyz/wp-content/uploads/2024/06/image-13-768x384.png 768w, https://www.aiuniverse.xyz/wp-content/uploads/2024/06/image-13.png 1100w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>Training generative AI models presents a variety of challenges and limitations. Key among these are:</p>



<h3 class="wp-block-heading">Data Quality and Quantity</h3>



<ol class="wp-block-list">
<li><strong>Data Availability</strong>: Generative models often require vast amounts of data to learn effectively. Accessing large, diverse datasets can be challenging, particularly in specialized domains.</li>



<li><strong>Data Quality</strong>: High-quality, well-labeled data is crucial. Poor-quality data can lead to biased or inaccurate models. Ensuring data cleanliness, dealing with missing values, and addressing inconsistencies are significant hurdles.</li>



<li><strong>Data Privacy and Security</strong>: Many datasets contain sensitive information. Ensuring data privacy and security while maintaining data utility for training is a complex issue, especially with regulations like GDPR.</li>
</ol>



<h3 class="wp-block-heading">Computational Resources</h3>



<ol class="wp-block-list">
<li><strong>High Computational Requirements</strong>: Training state-of-the-art generative models, such as GPT or GANs, demands substantial computational power. This includes powerful GPUs or TPUs, large memory, and extensive storage capabilities.</li>



<li><strong>Energy Consumption</strong>: The computational resources required translate into high energy consumption, raising concerns about the environmental impact and the sustainability of large-scale AI models.</li>
</ol>



<h3 class="wp-block-heading">Model Complexity</h3>



<ol class="wp-block-list">
<li><strong>Architecture Design</strong>: Choosing the right model architecture is crucial and non-trivial. It involves selecting appropriate neural network structures, layers, and parameters, which requires deep expertise and experimentation.</li>



<li><strong>Hyperparameter Tuning</strong>: Optimizing hyperparameters (learning rate, batch size, etc.) is essential for model performance but is often a time-consuming and resource-intensive process.</li>
</ol>



<h3 class="wp-block-heading">Training Stability and Performance</h3>



<ol class="wp-block-list">
<li><strong>Training Instability</strong>: Generative models, especially GANs, can suffer from instability during training. Issues such as mode collapse, vanishing gradients, and non-convergence are common.</li>



<li><strong>Scalability</strong>: As models and datasets grow, ensuring scalability of the training process becomes challenging. Efficient parallelization and distributed training are necessary but complex to implement.</li>
</ol>



<h3 class="wp-block-heading">Interpretability and Evaluation</h3>



<ol class="wp-block-list">
<li><strong>Model Interpretability</strong>: Understanding and interpreting the inner workings of generative models is difficult, making it hard to diagnose and fix issues.</li>



<li><strong>Evaluation Metrics</strong>: Evaluating generative models is less straightforward compared to discriminative models. Metrics like Inception Score (IS) and Frechet Inception Distance (FID) are used, but they have limitations and do not always correlate with human judgment.</li>
</ol>



<h3 class="wp-block-heading">Ethical and Social Implications</h3>



<ol class="wp-block-list">
<li><strong>Bias and Fairness</strong>: Generative models can inadvertently learn and propagate biases present in training data, leading to unfair or unethical outcomes.</li>



<li><strong>Misuse Potential</strong>: Generative models can be used to create misleading or harmful content (e.g., deepfakes), raising ethical concerns and necessitating robust safeguards.</li>
</ol>



<h3 class="wp-block-heading">Development and Maintenance Costs</h3>



<ol class="wp-block-list">
<li><strong>Resource Investment</strong>: Developing state-of-the-art generative models requires significant financial investment in terms of hardware, software, and human expertise.</li>



<li><strong>Continuous Updates</strong>: Maintaining and updating models to improve performance, address biases, and incorporate new data is an ongoing challenge.</li>
</ol>



<p>Addressing these challenges requires a multidisciplinary approach, combining advances in machine learning, data engineering, computational infrastructure, and ethical frameworks.</p>
<p>The post <a href="https://www.aiuniverse.xyz/overcoming-the-challenges-in-training-generative-ai-models-a-comprehensive-guide/">Overcoming the Challenges in Training Generative AI Models: A Comprehensive Guide</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/overcoming-the-challenges-in-training-generative-ai-models-a-comprehensive-guide/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Getting Serious About Data and Data Science</title>
		<link>https://www.aiuniverse.xyz/getting-serious-about-data-and-data-science/</link>
					<comments>https://www.aiuniverse.xyz/getting-serious-about-data-and-data-science/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Tue, 29 Sep 2020 07:36:12 +0000</pubDate>
				<category><![CDATA[Data Science]]></category>
		<category><![CDATA[Big data]]></category>
		<category><![CDATA[Change Management]]></category>
		<category><![CDATA[data quality]]></category>
		<category><![CDATA[data science]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=11844</guid>

					<description><![CDATA[<p>Source: sloanreview.mit.edu Data science, including analytics, big data, and artificial intelligence, is no longer a novel concept. Nor is the important foundation of high-quality data. Both have <a class="read-more-link" href="https://www.aiuniverse.xyz/getting-serious-about-data-and-data-science/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/getting-serious-about-data-and-data-science/">Getting Serious About Data and Data Science</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: sloanreview.mit.edu</p>



<p>Data science, including analytics, big data, and artificial intelligence, is no longer a novel concept. Nor is the important foundation of high-quality data. Both have contributed to impressive business successes — particularly among digital natives — yet overall progress among established companies has been painfully slow. Not only is the failure rate high, but companies have also proved unable to leverage successes in one part of the business to reap benefits in other areas. Too often, progress depends on a single leader, and it slows dramatically or reverses when that individual departs the company. In addition, companies are not seizing the strategic potential in their data. We’d estimate that less than 5% of companies use their data and data science to gain an effective competitive edge.</p>



<ol class="wp-block-list"><li>Over the years, we have worked with dozens of companies on their data journeys, advising them on the approaches, techniques, and organizational changes needed to succeed with data, including quality, data science, and AI. From our perspective, these are the two biggest mistakes organizations make:</li><li>They underinvest in the organization (people, structure, or culture), process, and the strategic transformations needed to get on offense — in other words, to take full advantage of their data and the data analytics technologies at their disposal.<br>They address data quality improperly, which leads them to waste critical resources (time and money) dealing with mundane issues. Bad data, in turn, breeds mistrust in the data, further slowing efforts to create advantage.</li></ol>



<p>Although the details at each company differ, seeing data too narrowly — as the province of IT or the data science organization, not of the entire business — is a recurring theme. This causes companies to overlook the transformative potential in data and therefore underinvest in the organizational, process, and strategic changes cited above. Similarly, they blame technology for their quality woes and failures to capitalize on data, when the real problem is poor management.</p>



<p>We’ve all observed how companies behave when they are truly serious about something — how the goal changes from incremental progress to rapid transformation; how they muster both breadth and depth of resources; how they align and train people; how they communicate new values and new ways of working; and how senior leaders drive the effort. Indeed, it almost seems as if companies go overboard when they are truly serious about something. Amazon’s Project D initiative to develop the Echo/Alexa smart speaker is a great illustration of that seriousness, with hundreds of employees, several startup acquisitions, heavy CEO involvement, and no expense spared. DBS Bank’s journey to being named World’s Best Digital Bank by Euromoney is another good example. The company’s CEO, Piyush Gupta, said the following upon receiving that award in 2018:</p>



<p>At DBS, we believe that banks tomorrow will look fundamentally different from banks today. That’s why we have spent the past three years deeply immersed in the digital agenda. This has been an all-encompassing journey, whether it is changing the culture and mindsets of our people, re-architecting our technology infrastructure, or leveraging big data, biometrics, and AI to make banking simple and seamless for customers.</p>



<p>The contrast with most companies’ data programs is stark — one can only conclude that many are not yet serious about data and data science. For those only beginning to explore data, this may be understandable. But, if you’ve been at it for three years or more, it is time to either get serious in addressing mistakes or invest your resources elsewhere — and expect to lose out to competitors.</p>



<h3 class="wp-block-heading">Stop Wasting Effort on Data Quality</h3>



<p>The obvious approach to addressing these mistakes is to identify wasted resources and reallocate them to more productive uses of data. This is no small task. While there may be budget items and people assigned to support analytics, AI, architecture, monetization, and so on, there are no budgets and people assigned to waste time and money on bad data. Rather, this is hidden away in day-in, day-out work — the salesperson who corrects errors in data received from marketing, the data scientist who spends 80% of his or her time wrangling data, the finance team that spends three-quarters of its time reconciling reports, the decision maker who doesn’t believe the numbers and instructs his or her staff to validate them, and so forth. Indeed, almost all work is plagued by bad data</p>



<p>The secret to wasting less time and money involves changing one’s approach from the current “buyer/user beware” mentality, where everyone is left on their own to deal with bad data, to creating data correctly — at the source. This works because finding and eliminating a single root cause can prevent thousands of future errors and eliminate the need to correct them downstream. This saves time and money — lots of it! The cost of poor data is on the order of 20% of revenue, and much of that expense can be eliminated permanently. That’s more than enough to fund the needed investments.</p>



<h3 class="wp-block-heading">Get On Offense</h3>



<p>Now consider the budgets for AI (as an example of “offense-minded” data efforts). It appears to us that, in many cases, the data science work to develop a new algorithm is funded well enough. Algorithm development is getting cheaper anyway, given that automated machine learning programs are doing more of the work. But useful algorithms die on the vine because the work to build processes, train people, address fear of change, and adapt the culture is substantially underfunded. Based on our experience, a good rule of thumb is that you should estimate that for every $1 you spend developing an algorithm, you must spend $100 to deploy and support it. A few of these dollars will go to building algorithms into work processes, and many more to training, building a culture that embraces data, and change management. Most companies aren’t spending this money yet, and it explains their lack of production AI deployments.</p>



<h3 class="wp-block-heading">Make Bold Moves</h3>



<p>What tangible steps should business leaders take to demonstrate that they are serious about data? First, they should more tightly couple their business and data strategies with an eye toward driving revenue growth. From the data perspective, opportunity abounds in fully exploiting proprietary data, driving analytics into every nook and cranny of the company, and augmenting virtually every decision using AI. You cannot — and should not — do them all, so you must select those most closely aligned with your business strategies. One sign that you’re on the right track is that there will be fewer data efforts. But those you do have will be far larger, more comprehensive, and more closely managed.</p>



<p>Second, get everyone fully engaged. After all, everyone is technically involved in your data efforts already. They interpret data correctly, or they do not; they create data correctly, or not; they use data to improve their work, or not; and they contribute to larger data initiatives, or not. Today, there are far too many “nots.” Similarly, managers push back against the nots, or they do not, and more senior leaders get in front of them, or not. So you must reach out to people, educate them, and enroll them in the effort, even as you grow increasingly intolerant of the inefficiencies stemming from bad data. This is going to take some time. One sign that you’re on the right track is that morale will improve. In our experience, once people get the hang of it, most of them find data work quite enjoyable. Importantly, in the data space, talent wins.</p>



<p>Third, draw a clear distinction between the management of data and the management of technology. Just as a movie is a different sort of asset than streaming technology, data and tech are different sorts of assets. Each demands its own specialized management. Yet today, too many companies subordinate data to tech. The result is that topics such as data architecture do not get the attention they deserve, leading to such absurdities as a bank having 130,000 databases, not including spreadsheets. Meanwhile, technology programs spend too much time dealing with the consequences of having systems that don’t talk to one another and spending too little time introducing new technologies to employees. One sign that you’re on the right track is that technology departments will become more effective and, in time, strategic.</p>



<p>Finally, now is a good time to start thinking about the longer-term roles data will play in your company. It is easy enough to recite the mantra “Data is the new oil.” And according to The Economist, data is now worth up to $2 trillion in the U.S. alone. But, of course, not all data is created equally. Some data — such as proprietary data, data needed to run the company, and data associated with other key assets — is so important that it should be treated as an asset in its own right. At the very least, you should make sure that end-to-end accountabilities for this data are clear.</p>



<p>We fully recognize how challenging these recommendations will prove to be. Yet they signal great opportunity, especially for the first companies in their sectors to embrace them. The needed approaches, methods, and technologies are widely available and have proved themselves over and over among digital natives and at the department level for established companies. It is clear enough that the future depends on data, so sooner or later, you have no real choice. As in all things, audentes Fortuna iuvat — fortune favors the brave.</p>
<p>The post <a href="https://www.aiuniverse.xyz/getting-serious-about-data-and-data-science/">Getting Serious About Data and Data Science</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/getting-serious-about-data-and-data-science/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Artificial Intelligence Tool Diagnoses Alzheimer’s with 95% Accuracy</title>
		<link>https://www.aiuniverse.xyz/artificial-intelligence-tool-diagnoses-alzheimers-with-95-accuracy/</link>
					<comments>https://www.aiuniverse.xyz/artificial-intelligence-tool-diagnoses-alzheimers-with-95-accuracy/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Tue, 01 Sep 2020 08:22:09 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[data quality]]></category>
		<category><![CDATA[Medical Research]]></category>
		<category><![CDATA[neural networks]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=11356</guid>

					<description><![CDATA[<p>Source: healthitanalytics.com A team from Stevens Institute of Technology has developed an artificial intelligence tool that can diagnose Alzheimer’s disease with more than 95 percent accuracy, eliminating the need <a class="read-more-link" href="https://www.aiuniverse.xyz/artificial-intelligence-tool-diagnoses-alzheimers-with-95-accuracy/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/artificial-intelligence-tool-diagnoses-alzheimers-with-95-accuracy/">Artificial Intelligence Tool Diagnoses Alzheimer’s with 95% Accuracy</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: healthitanalytics.com</p>



<p>A team from Stevens Institute of Technology has developed an artificial intelligence tool that can diagnose Alzheimer’s disease with more than 95 percent accuracy, eliminating the need for expensive scans or in-person testing.</p>



<p>In addition, the algorithm is also able to explain its conclusions, enabling human experts to check the accuracy of its diagnosis.</p>



<p>Alzheimer’s disease can impact a person’s use of language, the researchers noted. For example, people with Alzheimer’s tend to replace nouns with pronouns, and they can express themselves in a very roundabout, awkward way.</p>



<p>The team designed an explainable AI tool that uses attention mechanisms and a convolutional neural network to accurately identify well-known signs of Alzheimer’s, as well as subtle linguistic patterns that were previously overlooked.</p>



<p>Researchers trained the algorithm using texts composed by both healthy subjects and known Alzheimer’s sufferers describing a drawing of children stealing cookies from a jar. The team converted each individual sentence into a unique numerical sequence, or vector, representing a specific point in a 512-dimensional space.</p>



<p>This kind of approach allows even complex sentences to be assigned a concrete numerical value, making it easier to analyze structural and thematic relationships between sentences.</p>



<p>Using those vectors along with handcrafted features, the AI gradually learned to spot differences between sentences composed by healthy or unhealthy individuals, and was able to determine with significant accuracy how likely any given text was to have been produced by a person with Alzheimer’s.</p>



<p>“This is a real breakthrough,” said the tool’s creator, K.P. Subbalakshmi, founding director of Stevens Institute of Artificial Intelligence and&nbsp;professor of electrical and computer engineering at the Charles V. Schaefer School of Engineering &amp; Science.</p>



<p>“We’re opening an exciting new field of research, and making it far easier to explain to patients why the AI came to the conclusion that it did, while diagnosing patients. This addresses the important question of trustability of AI systems in the medical field.”&nbsp;&nbsp;</p>



<p>The AI system can also incorporate new criteria that may be identified by other research teams in the future, making the algorithm increasingly more accurate over time.</p>



<p>“We designed our system to be both modular and transparent,” Subbalakshmi explained. “If other researchers identify new markers of Alzheimer’s, we can simply plug those into our architecture to generate even better results.”</p>



<p>In the future, AI tools may be able to diagnose Alzheimer’s using any text, from emails to social media posts. However, to develop such an algorithm, researchers would need to train it on many different kinds of texts produced by known Alzheimer’s sufferers instead of just picture descriptions.</p>



<p>While this kind of data is not yet available, increasing access to this kind of information could lead to the development of accurate, comprehensive AI tools.</p>



<p>“The algorithm itself is incredibly powerful,” Subbalakshmi said. “We’re only constrained by the data available to us.”</p>



<p>The researchers’ next steps will be gathering new data that will help the algorithm diagnose patients with Alzheimer’s disease based on speech in languages other than English. The team is also uncovering ways in which other neurological conditions, such as aphasia, stroke, traumatic brain injuries, and depression, can impact language use.</p>



<p>“This method is definitely generalizable to other diseases,” said Subbalakshmi. “As we acquire more and better data, we’ll be able to create streamlined, accurate diagnostic tools for many other illnesses too.”&nbsp;</p>



<p>Researchers expect that providers can use this AI tool to more accurately diagnose Alzheimer’s, leading to earlier treatment and reduced healthcare costs.</p>



<p>“This is absolutely state-of-the-art,” said Subbalakshmi. “Our AI software is the most accurate diagnostic tool currently available while also being explainable.”</p>
<p>The post <a href="https://www.aiuniverse.xyz/artificial-intelligence-tool-diagnoses-alzheimers-with-95-accuracy/">Artificial Intelligence Tool Diagnoses Alzheimer’s with 95% Accuracy</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/artificial-intelligence-tool-diagnoses-alzheimers-with-95-accuracy/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Six Key Steps to Ensure Data Quality for Artificial Intelligence</title>
		<link>https://www.aiuniverse.xyz/six-key-steps-to-ensure-data-quality-for-artificial-intelligence/</link>
					<comments>https://www.aiuniverse.xyz/six-key-steps-to-ensure-data-quality-for-artificial-intelligence/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Tue, 11 Feb 2020 05:47:11 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[data quality]]></category>
		<category><![CDATA[Heine Krog Iversen]]></category>
		<category><![CDATA[TimeXtender]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=6657</guid>

					<description><![CDATA[<p>Source: solutionsreview.com As a growing number of companies are looking to build out and leverage artificial intelligence solutions across their organization, they’re often delayed due to poor <a class="read-more-link" href="https://www.aiuniverse.xyz/six-key-steps-to-ensure-data-quality-for-artificial-intelligence/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/six-key-steps-to-ensure-data-quality-for-artificial-intelligence/">Six Key Steps to Ensure Data Quality for Artificial Intelligence</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: solutionsreview.com</p>



<p>As a growing number of companies are looking to build out and leverage artificial intelligence solutions across their organization, they’re often delayed due to poor data quality that exist across their business operations. This quality deficiency prevents them from proceeding with their intended AI rollout. Once AI is fully implemented, it can improve data quality throughout a company.</p>



<p>Being faced with data quality issues forces a company to shift priorities and resources from implementing AI to fixing these quality shortcomings before they can proceed. This means extensive time delays, allocation of resources, and a slow draining of the AI budget.</p>



<p>The magnitude of this problem is multiplied by the number of data sources a company possesses, and even more so when you consider the ever-growing volume of new data pouring in. By some measures, the amount of new data available to a company doubles every two years. Having an abundance of data is a strategic advantage and should be looked upon in this manner, and a data infrastructure should be in place to support all of it. One sure way to take advantage of all this data is with the application of AI.</p>



<p>To help address this challenge of having data quality in place that is suitable for AI, organizations should look to appropriately design or redesign its underlying data architecture. Following are a few key steps that can be taken.</p>



<ol class="wp-block-list"><li>Consolidate your data<br>
Consolidating and integrating all corporate data into a centralized data hub provides a single platform for all data. This helps ensure one version of the truth and a consistent data home for all users throughout an organization regardless of department.</li><li>Connect your data<br>
Having connectivity and a data exchange enables the retrieval of data in its raw form before cleansing. This allows for single connectivity to all data sources now and into the future.</li><li>Use a modern data warehouse<br>
By using a modern data warehouse (MDW), users can modify and enrich their data so that data issues can be resolved once. Data that resides in various systems can be rationalized and after rationalization, golden records can be created. And with an MDW, historical data can be preserved.</li><li>Consider a semantic layer<br>
Within a data mart or what some call a semantic layer, data can be governed and prepared for any visualization tool. In addition, with a shared data model, all users can see the same data regardless of which data visualization tool is being used.</li><li>Deploy all-in-one data management software<br>
Too many organizations have complicated systems in place which can obscure data management and can lead to data quality issues. Having numerous tools requires extensive management and coordination to ensure synchronization among all the various tools. One method recommended to fix this is to use an all-encompassing data management platform that eliminates the usage of multiple discrete tools.</li><li>Automate data management procedures<br>
With an automated data management platform, time-consuming, hand-written code is eliminated in favor of automation. This eliminates the time needed for manual coding – which improves quality. It also frees up time to work on other issues regarding data quality and ultimately the bigger AI program.</li></ol>



<p>AI, once fully in place after data quality is secured, can also improve the quality of future data. With an established AI program, the inherent intelligence in AI can be used to automate the gathering and collecting of needed data and automatically enter the data – removing the need for manual data entry. And when you remove manual tasks you generally improve data quality. AI can also identify data errors and anomalies, remove duplicate or outdated records, and identify third-party data sources that can provide value related to the data model.</p>



<p>With all this in mind, desiring to have AI within an enterprise is certainly a worthy cause, but it starts with having solid data quality across the enterprise. And with AI in place, not only can organizations enjoy all the benefits of AI such as predicting trends, identifying new opportunities, and answering tough business questions, but they can also be assured that improvement of the level of quality of their future data will transpire as well.</p>
<p>The post <a href="https://www.aiuniverse.xyz/six-key-steps-to-ensure-data-quality-for-artificial-intelligence/">Six Key Steps to Ensure Data Quality for Artificial Intelligence</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/six-key-steps-to-ensure-data-quality-for-artificial-intelligence/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Machine Learning in Drug Development Requires Data Access, Standards</title>
		<link>https://www.aiuniverse.xyz/machine-learning-in-drug-development-requires-data-access-standards/</link>
					<comments>https://www.aiuniverse.xyz/machine-learning-in-drug-development-requires-data-access-standards/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Thu, 23 Jan 2020 07:21:31 +0000</pubDate>
				<category><![CDATA[Machine Learning]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[data quality]]></category>
		<category><![CDATA[Data Standards]]></category>
		<category><![CDATA[Drug Development]]></category>
		<category><![CDATA[Drug Discovery]]></category>
		<category><![CDATA[Machine learning]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=6320</guid>

					<description><![CDATA[<p>Source: healthitanalytics.com January 22, 2020 &#8211; Machine learning algorithms have the potential to accelerate and refine the drug development process, but the industry should expand data access <a class="read-more-link" href="https://www.aiuniverse.xyz/machine-learning-in-drug-development-requires-data-access-standards/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/machine-learning-in-drug-development-requires-data-access-standards/">Machine Learning in Drug Development Requires Data Access, Standards</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: healthitanalytics.com</p>



<p>January 22, 2020 &#8211; Machine learning algorithms have the potential to accelerate and refine the drug development process, but the industry should expand data access and create consistent data standards to ensure drug companies can fully leverage these tools, according to a report from the Government Accountability Office (GAO).</p>



<p>Drug companies spend ten to 15 years bringing a drug to market, often at high costs. Only about one in every 10,000 chemical compounds initially tested for drug potential makes it through the research and development pipeline, GAO noted, and is then approved by the FDA for marketing in the US. Machine learning tools could accelerate and improve the drug development process.</p>



<p>“Machine learning can make drug development more efficient and effective, decreasing the time and cost required to bring potentially more effective drugs to market,” GAO said.</p>



<p>“Both of these improvements could save lives and reduce suffering by getting drugs to patients in need more quickly. Lower research and development costs could also allow researchers to invest more resources in disease areas that are currently not considered profitable to pursue, such as rare or orphan diseases.”</p>



<p>Although drug companies already use machine learning throughout the drug development process, there are several challenges that hinder its advancement in this area, including barriers to data access and sharing.</p>



<p>“According to one industry representative, collecting data from the early drug discovery phase can be cost prohibitive. This representative said that certain health-related data may cost tens of thousands of dollars, as compared to just cents for other consumer related data that many technology companies use,” GAO stated.</p>



<p>“Data sharing also presents unique legal issues. According to stakeholders, privacy laws such as HIPAA can make it difficult for drug companies, especially those that are not regulated by HIPAA, to share or access data.”</p>



<p>To increase data sharing and access, GAO recommended that policymakers create mechanisms or incentives to share data held by private or public sectors, while also ensuring patient information is protected.</p>



<p>“To promote greater availability of data, policymakers could consider forming or facilitating research consortia that allow for secure data sharing,” the organization wrote.</p>



<p>“Policymakers could also consider creating a data repository through encouraging an industry-driven solution, establishing a public-private partnership, or creating a repository of all data under their control.”</p>



<p>In creating new ways to share and access data, stakeholders should ensure they adhere to laws around information exchange.</p>



<p>“Improper data sharing or use could have legal consequences. Increased data sharing could therefore require a careful review of the legal ramifications, because data are often gathered through a wide variety of mechanisms and governed by multiple legal frameworks,” GAO advised.</p>



<p>In addition to data sharing and access, policymakers will need to address the lack of quality data in the drug development process.</p>



<p>“Machine learning requires a large amount of accurate and representative data. This poses a unique challenge in drug development, as much of the data were not originally collected with machine learning in mind and may not be machine-readable or model-ready,” GAO wrote.</p>



<p>“Furthermore, according to an industry representative, data collected across different organizations and environments come in different formats, and this lack of standardization in data quality is a barrier.”</p>



<p>READ MORE: Data Standards, Governance Will Address Social Determinants of Health</p>



<p>Overcoming data quality issues will require policymakers to collaborate with appropriate stakeholders to establish data standards, GAO said.</p>



<p>“For example, a standard that defines synthetic data and how they can be used can help reduce bias by allowing researchers to generate data that could be used to better represent currently underrepresented communities,” the agency stated.</p>



<p>“Similarly, a standard data format for uploading and sharing data across platforms could reduce the need for data scientists to spend time converting data sets to machine-readable formats.”</p>



<p>GAO also named drug development research gaps as an obstacle to machine learning use.</p>



<p>“Research gaps present a significant challenge to advancing the use of machine learning in drug development. These gaps fall into two broad categories: gaps in understanding of fundamental biology and chemistry, and gaps in domain-specific machine learning research,” GAO said.</p>



<p>“Experts in the field have noted that addressing these issues may be transformational for future applications of machine learning in drug development.”</p>



<p>GAO suggested that policymakers promote basic research to generate new and better data to improve understanding of machine learning in drug development.</p>



<p>“Policymakers could promote the field in multiple ways, including approaches such as support for intramural research, grants, or other subsidies. Policymakers could choose to use one of these approaches or combine them,” GAO said.</p>



<p>“Policymakers could also support collaboration across sectors. The Machine Learning for Pharmaceutical Discovery and Synthesis Consortium (MLPDS) is a collaboration between large drug companies such as Pfizer, Merck, and Novartis with the Chemical Engineering, Chemistry, and Computer Science departments at the Massachusetts Institute of Technology, and has published a variety of papers at the intersection of machine learning and drug development.”</p>



<p>With these recommendations, policymakers and other stakeholders can advance the use of machine learning in drug development, refining and speeding the process to benefit patients.</p>
<p>The post <a href="https://www.aiuniverse.xyz/machine-learning-in-drug-development-requires-data-access-standards/">Machine Learning in Drug Development Requires Data Access, Standards</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/machine-learning-in-drug-development-requires-data-access-standards/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Artificial Intelligence Success Requires Human Validation, Good Data</title>
		<link>https://www.aiuniverse.xyz/artificial-intelligence-success-requires-human-validation-good-data/</link>
					<comments>https://www.aiuniverse.xyz/artificial-intelligence-success-requires-human-validation-good-data/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Wed, 27 Nov 2019 07:22:55 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[Analytics Strategies]]></category>
		<category><![CDATA[analytics technologies]]></category>
		<category><![CDATA[data quality]]></category>
		<category><![CDATA[Data Standards]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=5423</guid>

					<description><![CDATA[<p>Source: healthitanalytics.com November 26, 2019&#160;&#8211;&#160;As the foundation of nearly every healthcare trend, process, and solution, data has a vitally important role to play in care delivery and <a class="read-more-link" href="https://www.aiuniverse.xyz/artificial-intelligence-success-requires-human-validation-good-data/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/artificial-intelligence-success-requires-human-validation-good-data/">Artificial Intelligence Success Requires Human Validation, Good Data</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: healthitanalytics.com</p>



<p>November 26, 2019&nbsp;&#8211;&nbsp;As the foundation of nearly every healthcare trend, process, and solution, data has a vitally important role to play in care delivery and success.</p>



<p>From risk stratification to chronic disease management, precision medicine and medical research, data is at the center of everyday healthcare tasks and broader industry improvements, making it an incredibly valuable resource for organizations.</p>



<p>“If you talk to any data scientist, they’ll tell you that the more quality, scientifically validated data they have, the more likely they&#8217;re going to be able to generate useful trends and insights,” said Todd Frech, CIO at Press Ganey.</p>



<p>“The core of everything we do is taking the vast amounts of data that we collect and creating value for hospitals that are trying to improve their operations.”</p>



<p> With healthcare quickly becoming a digital industry, more and more entities are gathering meaning from this big data using artificial intelligence and other advanced analytics technologies. </p>



<p>“The challenge that we&#8217;re trying to overcome is that we have more data than a human can process, and we&#8217;re trying to develop insights based on those volumes of data. This issue is a natural fit for AI, so the use of this technology is going to continue to accelerate,” Frech said.</p>



<p>“AI can augment humans’ understanding of data, not only from the perspective of generating new insights, but also in generating those insights faster than a typical human analyst processes.”</p>



<p>There are countless examples of AI outperforming humans in analyzing and extracting insights from clinical data. The technology’s potential to transform the industry has led to concerns about robots encroaching on healthcare jobs, creating an environment run entirely by machines and devoid of human interaction.</p>



<p>While AI may disrupt standard care delivery, it’s unlikely that advanced analytics tools will completely take over the role of clinicians. In a field where high-stakes situations and sensitive data are routine, technology can’t simply be left to operate on its own, Frech stressed.</p>



<p>“AI is going to play a bigger part in healthcare, and humans will also continue to play a big part,” he said.</p>



<p>“We can&#8217;t just assume that AI is making the right decisions without human validation. There&#8217;s a trend that you&#8217;re going to see more – what’s called AI augmentation, or human augmentation with AI, more than what you would call complete robotic AI, meaning that you&#8217;re letting the AI make decisions without human intervention.”</p>



<p>Recent research has demonstrated that when implementing AI tools, human intervention can lead to optimal results. A study conducted by a team at NYU School of Medicine and the NYU Center for Data Science showed that combining AI with analysis from human radiologists significantly improved breast cancer detection.</p>



<p>Using an AI augmentation approach could also help organizations analyze and measure unstructured data.</p>



<p>“We collect hundreds of thousands of survey data points in the forms of responses to questions, as well as unstructured data in the form of comments. We use AI to look at the comments that come in our surveys. Those comments are obviously in the form of unstructured text, and they convey information on perception of the providers and of the service,” explained Frech.</p>



<p>“Those aren’t yes or no questions. Those are questions that require some soft skills to interpret. We can use AI to do an initial sentiment analysis, and that provides a way for us to really measure this type of data, which is not as binary as some of the data we typically evaluate.”</p>



<p>However, data-driven technologies can’t improve care if they’re fed inaccurate or incomplete information – in fact, this could have the opposite effect.</p>



<p>“Never underestimate the importance of data quality,” said Frech. &nbsp;“No AI tool is going to work well without high-quality data. People talk about data lakes and unstructured data, and these things are great tools. But without quality data, you’re going to have more of a data swamp than a data lake.”</p>



<p>“If you&#8217;re trying to use AI to gather insights without high-quality data, obviously the results aren’t going to pan out. Or even worse, the results could potentially offer dangerous recommendations that could negatively impact people,” he added.</p>



<p>Having a solid data ecosystem ensures that any innovative tools will contribute positively to a health system’s operations, Frech said, as well as communicating openly with other organizations.</p>



<p>“When implementing artificial intelligence or any new technology, make sure that the foundation is strong. Make sure that there&#8217;s testing and validation. If that doesn&#8217;t happen, there is potential for organizations to take steps backward rather than forward,” he said.</p>



<p>“Find opportunities with your peers, find case studies, talk to people who are using the technology. The more that your organizations can collaborate and learn from each other, the more ideas and successes will increase.”</p>



<p>AI has massive potential to revolutionize the way providers deliver care and make treatment decisions. The road to industry-wide adoption won’t be without its challenges, but the technology will likely make its way into regular clinical care.</p>



<p>“There are a lot of different ways to use AI, and there has been a lot of experimentation. Over time, there will be more and more successes, and those successes will come in fits and starts, depending on how the data in the market mature. There&#8217;s too much investment in AI right now to not have some of those successes come into play,” Frech concluded.</p>
<p>The post <a href="https://www.aiuniverse.xyz/artificial-intelligence-success-requires-human-validation-good-data/">Artificial Intelligence Success Requires Human Validation, Good Data</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/artificial-intelligence-success-requires-human-validation-good-data/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>NIH Promotes Big Data to Enhance Eye Disease Research</title>
		<link>https://www.aiuniverse.xyz/nih-promotes-big-data-to-enhance-eye-disease-research/</link>
					<comments>https://www.aiuniverse.xyz/nih-promotes-big-data-to-enhance-eye-disease-research/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Thu, 01 Aug 2019 06:03:24 +0000</pubDate>
				<category><![CDATA[Big Data]]></category>
		<category><![CDATA[Big data]]></category>
		<category><![CDATA[Clinical Analytics]]></category>
		<category><![CDATA[Data Integrity]]></category>
		<category><![CDATA[data quality]]></category>
		<category><![CDATA[Genomics]]></category>
		<category><![CDATA[NIH]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=4187</guid>

					<description><![CDATA[<p>Source: healthitanalytics.com July 31, 2019 &#8211; Improving collaboration between specialists and integrating multiple datasets to leverage big data will be key for advancing research for dry age-related macular degeneration <a class="read-more-link" href="https://www.aiuniverse.xyz/nih-promotes-big-data-to-enhance-eye-disease-research/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/nih-promotes-big-data-to-enhance-eye-disease-research/">NIH Promotes Big Data to Enhance Eye Disease Research</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Source: healthitanalytics.com</p>



<p>July 31, 2019 &#8211; Improving collaboration between specialists and integrating multiple datasets to leverage big data will be key for advancing research for dry age-related macular degeneration (AMD), according to a new report from a National Institute of Health (NIH) working group.</p>



<p>Over 11 million people in the United States are diagnosed with AMD, an eye disease that ultimately results in blindness. It is the leading cause of blindness among individuals 65 years of age and older.</p>



<p>The disease can manifest in one of two forms: neovascular (wet) or non-neovascular (dry). While the neovascular form progresses more rapidly, there are several known and proven treatments for the disease. There are no preventive measures for dry AMD nor treatment options.</p>



<h4 class="wp-block-heading">Dig Deeper</h4>



<ul class="wp-block-list"><li>As Artificial Intelligence Matures, Healthcare Eyes Data Aggregation</li><li>Is Healthcare Any Closer to Achieving the Promises of Big Data Analytics?</li><li>New Project Puts an Actuarial Eye on Big Data, Healthcare Costs</li></ul>



<p>“The working group thoroughly assessed what is known about dry AMD pathobiology, and the recommendations will be informative for considering future NEI research priorities to align with promising pathways for discovering therapeutic targets,” said Director of National Eye Institute (NEI), Paul Sieving, MD, PhD, in an earlier news release.</p>



<p>The working group recommended a systems biology approach to disease treatment, an integration of genomic, preclinical, medical, pharmacological, and clinical data to inform modeling of the disease progression. Synthesizing big data from all these areas including tissue samples from clinical trials will help inform predictive modeling which can then be used to inform individual patient care.</p>



<p>A personalized approach to disease management may also be helpful, the working group recommended. Such an approach should consider the disease stage, progression, and individual risk factors to provide preventive and treatment strategies specific to the patient, the report said. Collaborating will all points of care will allow a multidisciplinary team to use a patient’s unique clinical, imaging, and genomic data to treat the disease.</p>



<p>“We propose that researchers utilize a systems biology approach, integrating the big data available from clinical registries and various fields of biology known as ‘omics’ to develop better models and ultimately treatments for patients with this blinding disease,” stated report co-author Joan W. Miller, MD.</p>



<p>Due to a lack of preventive strategies and treatment options for dry AMD, the working group noted the need for improved understanding of the disease pathology and promoted clinical trial investigations to do so. Previous research has shown a genetic link to the disease as well as several lifestyle factors including smoking, but there is no work examining the effects these factors have on dry AMD.</p>



<p>A better understanding of how these factors impact the disease will help providers be better informed to watch for risk factors and promote inventive preventive strategies. Such understanding only comes from examining data and promoting the use of big, integrated data sources to help investigators use multiple sources to answer their questions.</p>



<p>Effective disease management will need multiple targets that differ based on the disease stage progression, the report notes. A strategy overhaul needs to take place that focuses on large-scale, collaborative, systems biology in order to effectively treat the disease.</p>



<p>“This approach would integrate basic, genomic, pre-clinical, medical, pharmacological, and clinical data into mathematical models of pathological processes at different stages of dry AMD in order to ask how relevant individual components act together within the living system,” Miller said.</p>



<p>The working group was appointed by the National Advisory Eye Council, a 12-member panel that establishes guidelines for the NEI under the NIH. The group was charged with a multilayered goal: to raise public health awareness about the impact of dry AMD, review the current state of research about the disease for a better understanding of its pathology, propose future research directions, encourage scientists to focus on AMD, and promote collaboration among a network of specialized providers.</p>
<p>The post <a href="https://www.aiuniverse.xyz/nih-promotes-big-data-to-enhance-eye-disease-research/">NIH Promotes Big Data to Enhance Eye Disease Research</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/nih-promotes-big-data-to-enhance-eye-disease-research/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>How dataops improves data, analytics, and machine learning</title>
		<link>https://www.aiuniverse.xyz/how-dataops-improves-data-analytics-and-machine-learning/</link>
					<comments>https://www.aiuniverse.xyz/how-dataops-improves-data-analytics-and-machine-learning/#respond</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Fri, 21 Jun 2019 10:51:07 +0000</pubDate>
				<category><![CDATA[Data Science]]></category>
		<category><![CDATA[analyze]]></category>
		<category><![CDATA[data quality]]></category>
		<category><![CDATA[DataOps]]></category>
		<category><![CDATA[improves]]></category>
		<category><![CDATA[master]]></category>
		<category><![CDATA[Technology]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=3893</guid>

					<description><![CDATA[<p>Source:- infoworld.com A dataops team will help you get the most out of your data. Here’s how people, processes, technology, and culture bring it all together Have you noticed <a class="read-more-link" href="https://www.aiuniverse.xyz/how-dataops-improves-data-analytics-and-machine-learning/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/how-dataops-improves-data-analytics-and-machine-learning/">How dataops improves data, analytics, and machine learning</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>Source:- infoworld.com</p>
<h3>A dataops team will help you get the most out of your data. Here’s how people, processes, technology, and culture bring it all together Have you noticed that most <span class="vm-hook-outer vm-hook-default vm-hook-inview"><span class="vm-hook">organizations</span></span> are trying to do a lot more with their data?</h3>
<p>Businesses are investing heavily in data science <span class="vm-hook-outer vm-hook-default"><span class="vm-hook">programs</span></span>, self-service business intelligence tools, artificial intelligence programs, and organizational efforts to promote data-driven decision making. Some are developing customer facing applications by embedding data visualizations into web and mobile products or collecting new forms of data from sensors (Internet of Things), wearables, and third-party APIs. Still others are harnessing intelligence from unstructured data sources such as documents, images, videos, and spoken language.</p>
<div class="connatix">
<div id="cnx-adUnit-overlay">    <strong>[ The essentials from InfoWorld: What is big data analytics? Everything you need to know • What is data mining? How analytics uncovers insights. | Go deep into analytics and big data with the InfoWorld Big Data and Analytics Report newsletter. ]</strong></div>
</div>
<p>Much of the work around data and analytics is on delivering value from it. This includes dashboards, reports, and other data visualizations used in decision making; models that data scientists create to predict outcomes; or applications that incorporate data, analytics, and models.</p>
<p>What has sometimes been undervalued is all the underlying data operations <span class="vm-hook-outer vm-hook-default"><span class="vm-hook">work</span></span>, or dataops, that it takes before the data is ready for people to analyze and format into applications to present to end users.</p>
<p>Dataops includes all the work to source, process, cleanse, store, and manage data. We’ve used complicated jargon to represent different capabilities such as data integration, data wrangling, ETL (extract, transform and load), data prep, data quality, master data management, data masking, and test data management.</p>
<p>The post <a href="https://www.aiuniverse.xyz/how-dataops-improves-data-analytics-and-machine-learning/">How dataops improves data, analytics, and machine learning</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/how-dataops-improves-data-analytics-and-machine-learning/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Big data systems up ante on data quality measures for users</title>
		<link>https://www.aiuniverse.xyz/big-data-systems-up-ante-on-data-quality-measures-for-users/</link>
					<comments>https://www.aiuniverse.xyz/big-data-systems-up-ante-on-data-quality-measures-for-users/#comments</comments>
		
		<dc:creator><![CDATA[aiuniverse]]></dc:creator>
		<pubDate>Fri, 06 Oct 2017 07:35:14 +0000</pubDate>
				<category><![CDATA[Big Data]]></category>
		<category><![CDATA[Data Science]]></category>
		<category><![CDATA[Amazon Web Services]]></category>
		<category><![CDATA[Big data]]></category>
		<category><![CDATA[cloud]]></category>
		<category><![CDATA[data quality]]></category>
		<category><![CDATA[data scientists]]></category>
		<category><![CDATA[IT]]></category>
		<guid isPermaLink="false">http://www.aiuniverse.xyz/?p=1374</guid>

					<description><![CDATA[<p>Source &#8211; searchdatamanagement.techtarget.com NEW YORK &#8212; In the rush to capitalize on deployments of big data platforms, organizations shouldn&#8217;t neglect data quality measures that can ensure what&#8217;s <a class="read-more-link" href="https://www.aiuniverse.xyz/big-data-systems-up-ante-on-data-quality-measures-for-users/">Read More</a></p>
<p>The post <a href="https://www.aiuniverse.xyz/big-data-systems-up-ante-on-data-quality-measures-for-users/">Big data systems up ante on data quality measures for users</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></description>
										<content:encoded><![CDATA[<p>Source &#8211; <strong>searchdatamanagement.techtarget.com</strong></p>
<p>NEW YORK &#8212; In the rush to capitalize on deployments of big data platforms, organizations shouldn&#8217;t neglect data quality measures that can ensure what&#8217;s used in analytics applications is clean and trustworthy, experienced IT managers said at the 2017 Strata Data Conference here last week.</p>
<p>Several speakers pointed to data quality as a big challenge in their big data environments &#8212; one that required new processes and tools to help get a handle on quality issues, as both the volumes of data being fed into corporate data lakes and use of the info by data scientists and other analysts grow.</p>
<p>&#8220;The more of the data you produce is used, the more important it becomes, and the more important data quality becomes,&#8221; said Michelle Ufford, manager of core innovation for data engineering and analytics at Netflix Inc. &#8220;But it&#8217;s very, very difficult to do it well &#8212; and when you do it well, it takes a lot of time.&#8221;</p>
<p>Over the past 12 months, Ufford&#8217;s team worked to streamline the Los Gatos, Calif., company&#8217;s data quality measures as part of a broader effort to boost data engineering efficiency based on a &#8220;simplify and automate&#8221; mantra, she said during a Strata session.</p>
<p>A starting point for the data-quality-upgrade effort was &#8220;acknowledging that not all data sets are created equal,&#8221; she noted. In general, ones with high levels of usage get more data quality checks than lightly used ones do, according to Ufford, but trying to stay on top of that &#8220;puts a lot of cognitive overhead on data engineers.&#8221; In addition, it&#8217;s hard to spot problems just by looking at the metadata and data-profiling statistics that Netflix captures in an internal data catalog, she said.</p>
<section class="section main-article-chapter" data-menu-title="Calling for help on data quality">
<h3 class="section-title"><i class="icon" data-icon="1"></i>Calling for help on data quality</h3>
<p>To ease those burdens, Netflix developed a custom data quality tool, called Quinto, and a Python library, called Jumpstarter, which are used together to generate recommendations on quality coverage and to set automated rules for assessing data sets. When data engineers run Spark-based extract, transform and load (ETL) jobs to pull in data on use of the company&#8217;s streaming media service for analysis, transient object tables are created in separate partitions from the production tables, Ufford said. Calls are then made from the temporary tables to Quinto to do quality checks before the ETL process is completed.</p>
<p>In the future, Netflix plans to expand the statistics it tracks when profiling data and implement more robust anomaly detection capabilities that can better pinpoint &#8220;what is problematic or wrong&#8221; in data sets, Ufford added. The ultimate goal, she said, is making sure data engineering isn&#8217;t a bottleneck for the analytics work done by Netflix&#8217;s BI and data science teams and its business units.</p>
<section class="section main-article-chapter" data-menu-title="Calling for help on data quality">Improving data consistency was one of the goals of a cloud-based data lakedeployment at Financial Industry Regulatory Authority Inc., an organization in Washington, D.C., that creates and enforces rules for financial markets. Before the big data platform was set up, fragmented data sets in siloed systems made it hard for data scientists and analysts to do their jobs effectively, said John Hitchingham, director of performance engineering at the not-for-profit regulator, more commonly known as FINRA.</p>
<p>A homegrown data catalog, called herd, was &#8220;a real key piece for making this all work,&#8221; Hitchingham said in a presentation at the conference. FINRA collects metadata and data lineage info in the catalog; it also lists processing jobs and related data sets there, and it uses the catalog to track schemas and different versions of data in the big data architecture, which runs in the Amazon Web Services (AWS) cloud.</p>
<p>To help ensure the data is clean and consistent, Hitchingham&#8217;s team runs validation routines after it&#8217;s ingested into Amazon Simple Storage Service (S3) and registered in the catalog. The validated data is then written back to S3, completing a process that he said also reduces the amount of ETL processing required to normalize and enrich data sets before they&#8217;re made available for analysis.</p>
</section>
<section class="section main-article-chapter" data-menu-title="Data quality takes a business turn">
<h3 class="section-title"><i class="icon" data-icon="1"></i>Data quality takes a business turn</h3>
<p>The analytics team at Ivy Tech Community College in Indianapolis also does validation checks as data is ingested into its AWS-based big data system &#8212; but only to make sure the data matches what&#8217;s in the source systems from which it&#8217;s coming. The bulk of the school&#8217;s data quality measures are now carried out by individual departments in their own systems, said Brendan Aldrich, Ivy Tech&#8217;s chief data officer.</p>
<p>&#8220;Data cleansing is a never-ending process,&#8221; Aldrich said in an interview before speaking at the conference. &#8220;Our goal was, rather than getting on that treadmill, why not engage users and get them involved in cleansing the data where it should be done, in the front-end systems?&#8221;</p>
<p>That process started taking shape when Ivy Tech, which operates 45 campuses and satellite locations across Indiana, deployed the cloud platform and Hitachi Vantara&#8217;s Pentaho data integration and BI software in late 2013 to give its business users self-service analytics capabilities. And it was cemented in July 2016 when the college hired a new president who mandated that business decisions be based on data, Aldrich said.</p>
<p>The central role data plays in decision-making gives departments a big incentive to ensure information is accurate before it goes into the analytics system, he added. As a result, data quality problems are being found and fixed more quickly now, according to Aldrich. &#8220;Even if you&#8217;re cleansing data centrally, you usually don&#8217;t find [an issue] until someone notices it and points it out,&#8221; he said. &#8220;In this case, we&#8217;re cleansing it faster than we were before.&#8221;</p>
</section>
</section>
<p>The post <a href="https://www.aiuniverse.xyz/big-data-systems-up-ante-on-data-quality-measures-for-users/">Big data systems up ante on data quality measures for users</a> appeared first on <a href="https://www.aiuniverse.xyz">Artificial Intelligence</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.aiuniverse.xyz/big-data-systems-up-ante-on-data-quality-measures-for-users/feed/</wfw:commentRss>
			<slash:comments>4</slash:comments>
		
		
			</item>
	</channel>
</rss>
