<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Extraktion | Skilja</title>
	<atom:link href="https://skilja.com/de/kategorie/extraktion/feed/" rel="self" type="application/rss+xml" />
	<link>https://skilja.com/de/</link>
	<description>Document Understanding – Deep Learning</description>
	<lastBuildDate>Fri, 27 Dec 2019 14:08:21 +0000</lastBuildDate>
	<language>de</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>

<image>
	<url>https://skilja.com/wp-content/uploads/2021/06/cropped-skilja_logo_transparent_02-32x32.png</url>
	<title>Extraktion | Skilja</title>
	<link>https://skilja.com/de/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Reading Medical Reports</title>
		<link>https://skilja.com/de/reading-medical-reports/</link>
		
		<dc:creator><![CDATA[skiljaadmin]]></dc:creator>
		<pubDate>Fri, 27 Dec 2019 14:08:21 +0000</pubDate>
				<category><![CDATA[Erkennung]]></category>
		<category><![CDATA[Extraktion]]></category>
		<category><![CDATA[Technologie]]></category>
		<guid isPermaLink="false">https://skilja.com/reading-medical-reports/</guid>

					<description><![CDATA[Medical Reports are complex documents that are written by doctors who use their specific language and style to express not only facts but also hypotheses and suggestions. They are intended to be read by other doctors or experts who have a deep knowledge of the subject at hand and can make judgements based on what [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>Medical Reports are complex documents that are written by doctors who use their specific language and style to express not only facts but also hypotheses and suggestions. They are intended to be read by other doctors or experts who have a deep knowledge of the subject at hand and can make judgements based on what they learn. And in the end the information contained therein is of vital importance for many decisions that are to be taken &#8211; medication, change of life style, possible surgery but also cost of insurance. </p>



<p>So reading medical reports using Artificial Intelligence and machine learning is a challenging task. The use case described here utilizes Laera Information Extraction to assist experts at a life insurance to calculate the health risk for persons. In the end the decision needs to be taken by doctors but the system can greatly assist them to sift through the amount of text provided. Because medical reports attached to a life insurance application can easily exceed 100dreds of pages. </p>



<p>Laera Information Extraction uses advanced AI methods to assess the risks contained in these reports and points them out to the experts. In a first step the diagnoses are extracted base on the common ICD-10 code. But a simple word search is not enough because each diagnosis needs to be put into its context. Here the option of Laera to assign multiple roles to an entity becomes very useful. Once a critical diagnosis is found Laera makes an assessment based on several categories:</p>



<ul class="wp-block-list"><li>Is the polarity negative or positive? In most cases symptoms are excluded in the reports and therefore the diagnosis is negative. These are of no interest (of course the patient is happy about that). Only the positively confirmed ones are relevant for the risk assessment.</li><li>Is the diagnosis for the present or the past? Many reports contain a lot of history of what happened in the past. While the history might be interesting, the main focus is on the current situation.</li><li>Is the diagnosis for the person herself or maybe for the family? Family history (Father had heart attack) might be important but needs to be assessed differently.</li></ul>



<figure class="wp-block-image"><a href="https://skilja.com/reading-medical-reports/finding-diagnoses/"><img decoding="async" src="https://skilja.com/wp-content/uploads/2019/12/Finding-Diagnoses.png" alt="" class="wp-image-156"/></a><figcaption><em>Finding Diagnoses, polarity and roles</em></figcaption></figure>



<p>Laera intelligent extraction performs all these tasks and analyzes all pages in milliseconds using semantic and structural methods:
</p>



<ul class="wp-block-list"><li>Find all diagnose and symptoms</li><li>Find the polarity (is the diagnosis excluded or asserted</li><li>Determine the context (role e.g. self or family)</li><li>Classify the paragraph by relevance</li><li>Present and auto-summarize results in ICD10 terms</li><li>Highlight relevant areas in the document for quick visual confirmation</li></ul>



<p>Of course, just to make sure this does not get lost, the roles and assignments are not defined by rules but trained using machine learning from a few hundred labeled examples.</p>



<figure class="wp-block-image"><a href="https://skilja.com/reading-medical-reports/icd-10-summary-of-symptoms/"><img decoding="async" src="https://skilja.com/wp-content/uploads/2019/12/ICD-10-Summary-of-Symptoms.png" alt="" class="wp-image-157"/></a><figcaption><em>Summary of symptoms with polarity in ICD-10 tree</em></figcaption></figure>



<p>The customer using this system could reduce the time spent to assess an application by more than 50% as the experts get a prepared data set that allows them to quickly jump tp the relevant sections and make the decision.</p>



<p>It is also important to note that this is not a hard coded special solution, but an example of the application of the Laera Information Extraction product. Any other industry can use this approach to solve their specific requirements. Example range from contract management to court documents, but of course also standard default extraction tasks can be easily solved with Laera Information Extraction.</p>



<p>If you are interested to learn more about this use case or the application of AI extraction in your specific domain please let us know via e-Mail at info(at)skilja.com and we will be happy to provide more information.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The Magic of Online-Learning</title>
		<link>https://skilja.com/de/the-magic-of-online-learning/</link>
		
		<dc:creator><![CDATA[skiljaweb3]]></dc:creator>
		<pubDate>Sun, 03 Feb 2019 08:26:58 +0000</pubDate>
				<category><![CDATA[Erkennung]]></category>
		<category><![CDATA[Extraktion]]></category>
		<category><![CDATA[Klassifikation]]></category>
		<category><![CDATA[Technologie]]></category>
		<guid isPermaLink="false">https://skilja.com/the-magic-of-online-learning/</guid>

					<description><![CDATA[Wouldn’t it be nice if your AI enabled document processing system would continuously take the input from user interactions and use this information to improve the quality of recognition over time? And nobody would have to take care of this – even in the case of hundreds of document classes with dozens of index fields [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>Wouldn’t it be nice if your AI enabled document processing system would continuously take the input from user interactions and use this information to improve the quality of recognition over time? And nobody would have to take care of this – even in the case of hundreds of document classes with dozens of index fields each. In the best case the system would be easy to set up, run completely unattended in background and work like a charm.</p>



<p>This is what Skilja with its Laera Classification and Extraction software suites provides. We have completely implemented this new paradigm which is available either as SDKs or as integrated modules to our Vinna Document Processing Platform. But of course what looks easy for the user requires significant infrastructure and automated checks and balances to make this a reliable and stable part of your processing tasks.</p>



<p>Machine Online-Learning of document classification and recognition uses supervised and unsupervised continuous training of incoming data streams. Supervised learning will take the corrections the users have made, analyze them and apply them as new patterns as appropriate. Unsupervised learning will use the results of successful and correct classification and extraction to generate additional knowledge (expanding the space) and statistics of usage of existing knowledge. Both combined are then used to continuously improve the system. The infrastructure is set up quickly and consists of services that do the work in the background: collect statistics, collect samples, analyze the validity of the new data and publish them to the production runtime system if the AI has determined them to be valid additions.</p>



<figure class="wp-block-image"><a href="https://www.skilja.com/wp-content/uploads/2019/02/Skilja-Classification-3.0-Online-Learning.png"><img decoding="async" src="https://www.skilja.com/wp-content/uploads/2019/02/Skilja-Classification-3.0-Online-Learning-1024x640.png" alt="" class="wp-image-1339"/></a></figure>



<p>As we know that system administrators might be vary of having their setup changed automatically (at least until they have seen it really works) there is several intermediate levels of AI automation that they can chose. The most important are:</p>



<ul class="wp-block-list"><li>Have all changes and each new document manually reviewed, benchmarked and checked before explicitly publishing it. This is the box on the left</li><li>Have automatically created improvements be reviewed and explicitly published</li><li>View any conflict and resolve them manually (or at least check them)</li><li>Restrict the users that can contribute to the training to a certain group. Only corrections from this group will be taken into account while the input from less experienced users will be discarded.</li></ul>



<p>But in the end learning can run completely unattended. As in school (think exams) we need to check the validity of the new knowledge before we apply it. Therefore Laera algorithms will always analyze for conflicts that are created and try to resolve them. Im addition each new revision of the training pattern will fully automatically be quality checked in background and only be accepted if the recognition results of the new model exceed the existing one. This is an assurance for the production system: Changes in quality will always only go into one direction – better!</p>



<p>Again, this is not a black box but Laera provides precise insight of what is happening and lets you influence or even revert the suggested improvements at any stage. Laera Monitor is the tool for this, a web application that shows the continuously measured quality numbers of your system.</p>



<figure class="wp-block-image"><a href="https://www.skilja.com/wp-content/uploads/2019/02/Laera-Monitor.png"><img decoding="async" src="https://www.skilja.com/wp-content/uploads/2019/02/Laera-Monitor-1024x651.png" alt="" class="wp-image-1340"/></a></figure>



<p>The example shown here shows a typical curve for the F1 score (average quality measurement). Starting with a setup of a few hundred trained documents the quality quickly deteriorates as new and unknown samples arrive in production. Especially when the real volumes start to be processed. It is interesting to see that the precision stays high close to 95% which is very satisfying, but recall (recognition rate) goes down as the system simply does not “know” the new documents. But then online learning kicks in and uses the new samples and corrections made to quickly improve the quality to 95% after a few thousand new training documents have been processed.</p>



<p>Online Learning will make classification and extraction much easier in the future. After an initial setup AI will simply learn in background what needs to be known to arrive at he best possible automation rate within a few weeks. This makes a whole new area of processes (for example with smaller document volumes) available and will greatly improve quality for existing automation processes.</p>



<p>Please let us know if you have additional questions or need more insight or have a direct interest. Contact us under info (at) skilja.com.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Document Understanding Primer</title>
		<link>https://skilja.com/de/document-understanding-primer/</link>
		
		<dc:creator><![CDATA[Alexander]]></dc:creator>
		<pubDate>Mon, 27 Feb 2012 16:16:00 +0000</pubDate>
				<category><![CDATA[Erkennung]]></category>
		<category><![CDATA[Extraktion]]></category>
		<category><![CDATA[Grundlagen]]></category>
		<category><![CDATA[Klassifikation]]></category>
		<guid isPermaLink="false">https://skilja.com/document-understanding-primer/</guid>

					<description><![CDATA[For a long time document understanding has been a research topic in computer sciences. We have seen conferences discussing concepts and approaches to use computers and machine learning for understanding documents. Quite often this topic appears also in proceedings on text analytics or more recently document analysis. In recent times also practical applications have become [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>For a long time document understanding has been a research topic in computer sciences. We have seen conferences discussing concepts and approaches to use computers and machine learning for understanding documents. Quite often this topic appears also in proceedings on text analytics or more recently document analysis.</p>



<p>In recent times also practical applications have become available that provide basic functionality in understanding documents. Typically these applications are used by enterprises to manage large amounts of incoming documents (especially paper) and to offer some kind of automatic recognition and distribution of documents. As these early and simple solutions have proven successful we will soon see a wealth of new concepts that will allow providing much larger benefits to companies and end users making use of this. Therefore let’s take a look what this actually is.</p>



<p>The first goal of document understanding is to identify&nbsp;the function and the meaning&nbsp;of a document and its parts. Typically a document is written for a specific purpose which defines its function. An invoice is designed and created to notify a buyer on the goods bought and how much money needs to paid for them along with some other information for accounting and tax purposes. All content of the invoice follows this function. Or an application form in a bank is used to collect all information that is needed to open an account. This document is normally very structured. On the other side an e-Mail (which is also a document) conveys information, opens a discussion and calls for action in a very unstructured way.</p>



<p>So the first step in document understanding is to identify the function and separate the documents to be processed accordingly. Typically this step is called “classification”. However this is only the primary classification as more categorization according to a taxonomy can occur which do not have the purpose to define the function. It is therefore very important to distinguish these two types of classification as a lot of misunderstanding results from confusion between these. The function of a document determines the possible content and the information entities that can be found on it.</p>



<p>The second step of understanding is to identify all information entities or predicates of a document related to the function. Typically only a few entities are needed and required (e.g. the amounts on an invoice) but we would prefer to only talk about “understanding” when all entities have been identified. To identify an entity (like a tax amount) it needs to be detected and a meaning must be attached. A number without a meaning is not a predicate. Only if all entities can be labeled with a meaning the computer system really understands the meaning of the document.</p>



<p>In a third step real understanding can take place if the entities are brought into the context of the document function, the purpose of the communication and the other entities. Typically a (business) document triggers an action. Context and correlation between the discovered predicates needs to be analyzed to determine which action. An e-Mail may contain a request to send some information back (which would be an entity “request for information”). But only in context with the rest of the e-Mail, the e-Mail thread or attachments it becomes clear which actions to perform. I will discuss this in more depth in upcoming posts.</p>



<p>As you have seen it is important to know the function of a document or any text to be able to understand its content. And with the knowledge of function (=purpose) and some content we can take action and already have some kind of document understanding. This is what current solutions provide.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
