<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Essentials | Skilja</title>
	<atom:link href="https://skilja.com/category/essentials/feed/" rel="self" type="application/rss+xml" />
	<link>https://skilja.com</link>
	<description>Document Understanding – Deep Learning</description>
	<lastBuildDate>Fri, 16 Jan 2026 16:11:11 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>

<image>
	<url>https://skilja.com/wp-content/uploads/2021/06/cropped-skilja_logo_transparent_02-32x32.png</url>
	<title>Essentials | Skilja</title>
	<link>https://skilja.com</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>The End of OCR: How Word-Level Understanding Changes Everything</title>
		<link>https://skilja.com/the-end-of-ocr-how-word-level-understanding-changes-everything/</link>
		
		<dc:creator><![CDATA[Alexander]]></dc:creator>
		<pubDate>Thu, 30 Oct 2025 16:10:19 +0000</pubDate>
				<category><![CDATA[Cognition]]></category>
		<category><![CDATA[Essentials]]></category>
		<category><![CDATA[Technology]]></category>
		<guid isPermaLink="false">https://skilja.com/?p=2516</guid>

					<description><![CDATA[Modern systems don’t read characters at all—not individually. They read words, phrases, and even meaning. OCR has evolved into something fundamentally different, and this shift delivers a new level of accuracy, fluidity, and naturalness that previous generations could not reach.
This is what Skilja has done with Lesa, our deep-learning, transformer-based system designed to read like humans do: holistically, contextually, and intelligently.]]></description>
										<content:encoded><![CDATA[<p style="font-weight: 400;">For more than half a century, OCR—Optical Character Recognition—has meant one thing: machines deciphering single characters. From early template-matching in the 1960s to the statistical engines of the 2000s, OCR has always approached reading as a mechanical process. It segmented text into character-shaped fragments and tried to analyze what each one meant, letter by letter. But that era is now ending. Modern systems don’t read characters at all—not individually. They read <em>words</em>, <em>phrases</em>, and even <em>meaning</em>. OCR has evolved into something fundamentally different, and this shift delivers a new level of accuracy, fluidity, and naturalness that previous generations could not reach.</p>
<p style="font-weight: 400;">This is what Skilja has done with <strong>Lesa</strong>, our deep-learning, transformer-based system designed to read like humans do: holistically, contextually, and intelligently.</p>
<h4><strong>A Brief Look Back: From Characters to Context</strong></h4>
<p style="font-weight: 400;">Traditional OCR started as an analogous process (therefore the O = Optical) and went through several stages:</p>
<ul>
<li><strong>1950s–1980s:</strong>Rigid template matching—effective only with perfect typewritten pages.</li>
<li><strong>1990s:</strong>Feature extraction and early machine learning—better but still brittle.</li>
<li><strong>2000s–2010s:</strong>Statistical modeling and improved analysis workflows—good enough for books and printed forms and constrained hand print, but always character-bound.</li>
</ul>
<p style="font-weight: 400;">Even at its best, classical OCR remained a guessing game. It mistook <em>1</em> for <em>l</em>, turned smudges into glyphs, and struggled with anything outside its narrow expectations. Especially handwriting.</p>
<p style="font-weight: 400;">The problem wasn’t better algorithms—it was characters – but humans read words and not characters as anyone who has watched a child learn to read will confirm.</p>
<p><img fetchpriority="high" decoding="async" class="alignnone size-large wp-image-2523" src="https://skilja.com/wp-content/uploads/2026/01/Lesa-Full-Text-Recognition-1-1024x573.png" alt="" width="1024" height="573" srcset="https://skilja.com/wp-content/uploads/2026/01/Lesa-Full-Text-Recognition-1-980x548.png 980w, https://skilja.com/wp-content/uploads/2026/01/Lesa-Full-Text-Recognition-1-480x268.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) and (max-width: 980px) 980px, (min-width: 981px) 1024px, 100vw" /></p>
<h4><strong>The Deep Learning Shift: From Decoding Shapes to Understanding Language</strong></h4>
<p style="font-weight: 400;">Transformers changed everything. Instead of interpreting characters, transformer-based models interpret <strong>sequences</strong>, <strong>context</strong>, and <strong>linguistic probability</strong>. They don’t see text as isolated shapes but as parts of sentences, paragraphs, and concepts.</p>
<p style="font-weight: 400;">This allows Lesa to:</p>
<ul>
<li>recognize <strong>entire words</strong>, not just letters,</li>
<li>use surrounding text as context,</li>
<li>maintain coherence across full pages,</li>
<li>and adapt to different visual styles.</li>
</ul>
<p style="font-weight: 400;">Reading becomes a <strong>language-understanding task</strong>, not a pixel-decoding task.</p>
<h4><strong>Lesa: Built for Word-Level Intelligence</strong></h4>
<p style="font-weight: 400;">Lesa is designed from the ground up to treat a document as a linguistic landscape. Using a transformer architecture trained on diverse text and real-world images, Lesa doesn’t ask, “What is this character?” It asks, “What does this <em>word</em> say—and how does it fit into the sentence?” This matters since now we have:</p>
<ul>
<li><strong>Far fewer errors</strong>: No more character-by-character error cascades.</li>
<li><strong>Natural output</strong>: Clean spacing, correct punctuation, coherent text.</li>
<li><strong>Font and layout robustness</strong>: Works across stylized text, messy receipts, tables, multi-column pages.</li>
</ul>
<p style="font-weight: 400;">One of the most transformative outcomes of word-level understanding is that <strong>constrained handwriting</strong>—the kind found in forms, notes, medical records, delivery slips, and corporate paperwork—now works <em>much better</em>. This means that messy capital letters, box-filled handwriting, and half-printed forms suddenly become highly readable. Handwriting recognition suddenly is at our fingertips and routinely applied.</p>
<p><img decoding="async" class="alignnone size-large wp-image-2521" src="https://skilja.com/wp-content/uploads/2026/01/Lesa-Handprint-Recognition-1024x251.png" alt="" width="1024" height="251" srcset="https://skilja.com/wp-content/uploads/2026/01/Lesa-Handprint-Recognition-980x240.png 980w, https://skilja.com/wp-content/uploads/2026/01/Lesa-Handprint-Recognition-480x118.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) and (max-width: 980px) 980px, (min-width: 981px) 1024px, 100vw" /></p>
<p style="font-weight: 400;">Calling this the “end of OCR” isn’t hype. It’s a technical recognition of what has changed</p>
<p style="font-weight: 400;"><strong>Traditional OCR = character recognition</strong><br />
<strong>Modern OCR = language understanding</strong></p>
<p style="font-weight: 400;">Lesa is part of a new generation of systems that read documents the way humans do—by interpreting words, not decoding symbols.</p>
<p style="font-weight: 400;">OCR as we knew it is over. Something better has replaced it. <strong>Lesa</strong> represents this new era.</p>
<p style="font-weight: 400;">
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Auto Classification and Bias</title>
		<link>https://skilja.com/auto-classification-and-bias/</link>
		
		<dc:creator><![CDATA[skiljaadmin]]></dc:creator>
		<pubDate>Tue, 22 Jul 2025 12:01:16 +0000</pubDate>
				<category><![CDATA[Classification]]></category>
		<category><![CDATA[Cognition]]></category>
		<category><![CDATA[Essentials]]></category>
		<guid isPermaLink="false">https://skilja.com/?p=229</guid>

					<description><![CDATA[Personal bias and individual opinions are a big issue in standardized business processing if they happen to influence the outcome of a process and the decisions made. Nobody wants to be subject to random changes in the outcome of a personal request – and yet it happens. Because humans have a bias in how they see facts, based on their education, cultural background and even the mood they happen to be in at a certain time in the week. So in addition to different persons making different decisions you can even expect that the same person makes different decisions during the week. You just look differently at a task on Monday morning than on Friday evening. The reason is so-called priming which happens to all of us day-by-day through our experience, knowledge, physical condition, context and a lot of other small factors.]]></description>
										<content:encoded><![CDATA[
<p>Personal bias and individual opinions are a big issue in standardized business processing if they happen to influence the outcome of a process and the decisions made. Nobody wants to be subject to random changes in the outcome of a personal request – and yet it happens. Because humans have a bias in how they see facts, based on their education, cultural background and even the mood they happen to be in at a certain time in the week. So in addition to different persons making different decisions you can even expect that the same person makes different decisions during the week. You just look differently at a task on Monday morning than on Friday evening. The reason is so-called priming which happens to all of us day-by-day through our experience, knowledge, physical condition, context and a lot of other small factors.</p>



<p>In a recent article on the <a href="http://www.skilja.de/2015/the-meaning-of-words/">meaning of words</a> we have shown how the sound of words influences our perception. There are a lot more linguistic associations that influence the way we think and behave, which introduce bias. If for example I tell you that I am driving north across a hilly terrain, would you expect the trip to be mosty uphill or downhill? In fact most people associate movement to north with uphill and to south with downhill. An interesting study by psychologists<a href="http://papers.ssrn.com/sol3/papers.cfm?abstract_id=963159"> Leif D. Nelson from UC San Diego and Joseph Simmons from Yale</a> shows that these associations can actually be measured and produce some strange biases: People think it will take longer to travel north than south, that it will cost more to ship to a northern than to a southern location, and that a moving company will charge more for northward movement than for southward movement. A <a href="http://spp.sagepub.com/content/2/5/547">similar study </a>concluded, that people assume that property is more valuable when it sits in the northern part of town. Of course these opinions stem from the decision by the old Greeks to plot the map of the world with northern parts above the south. But also it shows us clearly how much we are biased by our language and of course north/south is only one of many linguistic associations we are exposed to.</p>



<figure class="wp-block-image"><a href="http://www.skilja.de/wp-content/uploads/2015/06/compass-152121_1280.png"><img decoding="async" class="wp-image-946" src="http://www.skilja.de/wp-content/uploads/2015/06/compass-152121_1280.png" alt="North-South Compass" /></a></figure>



<p>Ancient mapmakers introduced north and south unwittingly, but lawyers do have an intention when they describe car accidents. While the defense might call a car accident “contact”, the plaintiff might say one car “smashed” the other. Elizabeth Loftus and John Palmer showed in a classic experiment that these labels really matter. They had a group of students watch the same series of traffic accidents. Then they were asked to estimate the speed of the cars when the accident occurred. When the scene was described in a way that the cars “contacted” one another, the average speed estimation by the students was thirty-two miles an hour, whereas the estimate was forty miles an hour when they said that the cars “smashed” one another. In a another experiment, 14% of participants incorrectly remembered seeing shattered glass when told that the cars “hit” one another, whereas 32% of participants made the same error when told the cars “smashed” into one another. This shows that even a single word can change how people remember an event they witnessed only minutes earlier – making it very clear how priming can bias our decisions.</p>



<p>This brings us back to auto-classification. A classifier like the Skilja Content Classifier is trained with representative samples that are collected by a group of people. If applied to a bunch of documents it will then make the same decision over and over again. It represents – through machine learning – the average opinion about the content of a document and will repeat it without tiring. Same in Monday morning and Friday evening – with a speed of several 100.000 pages per hour. It will make errors – based on statistics – but not more than a human. And the errors are reproducible and can be corrected if necessary. If you have ever thought about compliance – this is a good example. Because compliance does not say that you cannot make errors. It says that you need a reproducible, documented procedure how you store and treat your documents. Auto-classification can help to achieve this goal. It is a great tool for boosting productivity. But it is even more helpful to avoid bias, irreproducible results and non-compliance.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>How Meanings of Words Change</title>
		<link>https://skilja.com/how-meanings-of-words-change/</link>
		
		<dc:creator><![CDATA[skiljaadmin]]></dc:creator>
		<pubDate>Fri, 02 Dec 2022 11:57:32 +0000</pubDate>
				<category><![CDATA[Cognition]]></category>
		<category><![CDATA[Essentials]]></category>
		<guid isPermaLink="false">https://skilja.com/?p=109</guid>

					<description><![CDATA[We all know that our language is fluid and words can change their meaning over time. Words get extinct and new words are created but more often existing words are adapted to new circumstances. It is interesting to see how this happens in the course of years but sometimes words change their meaning overnight. In [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>We all know that our language is fluid and words can change their meaning over time. Words get extinct and new words are created but more often existing words are adapted to new circumstances. It is interesting to see how this happens in the course of years but sometimes words change their meaning overnight.</p>



<p>In a study on “Statistically Significant Detection of Linguistic Change”, published last year (available online from&nbsp;<a href="http://arxiv.org/abs/1411.3315" target="_blank" rel="noreferrer noopener">arxiv.org</a>), the researchers Vivek Kulkarni, Rami Al-Rfou, Bryan Perozzi, Steven Skiena have used data mining to find out how the way we use words is revealing the linguistic earthquakes that constantly change our language. There findings are very interesting for anybody who works professionally with text analytics as it reveals a lot about how semantics in our language work. Kulkarni et al. have tracked these linguistic changes by mining the corpus of words stored in databases such as Google Books, movie reviews from Amazon and of course Twitter.</p>



<p>In the pre-internet times the usage and meaning of words changed relatively slowly. This is can be seen in the metamorphoses of the word “gay” from its social meaning in the fifties to the purely sexual-orientation meaning in our time. This is nicely displayed in the word cloud view below:</p>



<figure class="wp-block-image"><a href="http://www.skilja.de/wp-content/uploads/2015/12/Gay-Meaning.png"><img decoding="async" src="http://www.skilja.de/wp-content/uploads/2015/12/Gay-Meaning.png" alt="Showing the meaning of the word Gay over the century" class="wp-image-1034"/></a></figure>



<p>A faster change occurred in the 1970s to the word “mouse”, when it gained the new meaning of “computer input device” and later, the word “windows” was used internationally as the name of the Microsoft operating system within a few years.</p>



<p>Today the meaning of a word can change almost instantly. Before October 2012, the word “sandy” was an adjective meaning “covered in or consisting mostly of sand”. Then Hurricane “Sandy” approached. Almost overnight, this word gained an additional meaning as a proper noun for one of the costliest storms in US history.</p>



<p>Now this might not sound like a big deal for us – but just imagine the insurance industry and the thousands of e-Mails they suddenly receive referring to damages by Sandy! If they are using a static or rule based classification system, they might easily miss the point on these.</p>



<p>So this is a big challenge for automatic classification systems that have been trained in a machine learning algorithm on a specific set of documents containing words in a specific meaning. When these meanings suddenly change or the usage of words is widened suddenly the classifier will make wrong decisions. The only way to cope with this problem in a living and productive system is continuous learning. The system must learn from user corrections – supervised learning &#8211;&nbsp; but also from good classification that contains some new aspects and features. This is called unsupervised learning and is so important as the number of documents that can be used for training the system is much higher than the manual correction. By unsupervised learning – or better enhancement, which also includes forgetting btw – the classification system will be able to cope with changed meanings over the time. Abrupt changes as mentioned above will lead to a drop in classification rate from which the classifier will recover within a few days – like humans who will also need a short time to adapt.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>IDP: Solving bot illiteracy in the digital workforce &#8211; Part 2</title>
		<link>https://skilja.com/idp-solving-bot-illiteracy-in-the-digital-workforce-part-2/</link>
		
		<dc:creator><![CDATA[Guest]]></dc:creator>
		<pubDate>Wed, 21 Oct 2020 10:12:00 +0000</pubDate>
				<category><![CDATA[Essentials]]></category>
		<category><![CDATA[Guest Post]]></category>
		<category><![CDATA[Market]]></category>
		<guid isPermaLink="false">https://skilja.com/?p=730</guid>

					<description><![CDATA[Editor’s note: This is a guest post from Jupp Stöpetie In this post we examine the role of Intelligent Document Processing (IDP) relative to Robotic Process Automation (RPA) and how these technologies drive Digital Transformation when combined. Part one was looking at what is driving and enabling Digital Transformation. This part two is dedicated to RPA and [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p><span class="has-inline-color" style="color: #7cb955;"><strong>Editor’s note</strong>: </span>This is a guest post from Jupp Stöpetie</p>



<p><em><em>In this post we examine the role of Intelligent Document Processing (IDP) relative to Robotic Process Automation (RPA) and how these technologies drive Digital Transformation <em>when combined</em>.</em> <em>Part one was looking at what is driving and enabling Digital Transformation. This part two is dedicated to RPA and why IDP is essential.</em></em></p>



<h2 class="wp-block-heading"><strong>Robotic Process Automation </strong></h2>



<p>Since maybe 3 or 4 years RPA has become the tool of choice for automating repetitive tasks for many companies. RPA vendors claim that their systems are easy to set up and maintain without the need of coding. Vendors propagate that RPA systems can be operated by business managers with some light-weight training. Basically every manager can create bots that replace human workers. This has led to a wide spread of RPA systems in businesses. There is no need anymore for complicated, lengthy and costly automation projects managed by IT departments. RPA puts automation and the use of AI in the hands of business managers. Today the workforce in companies is increasingly a combination of humans and robots. But of course things are more complicated when you look a bit deeper than the RPA marketing collaterals. What is for example when we need to retrieve information from a document? That is no problem for human workers. But bots have no humanlike reading skills.  They only can “read” data from structured data like files and databases. That is because it is easy to instruct bots with the exact location where to find the data.</p>



<h2 class="wp-block-heading"><strong>Why bots can&#8217;t read documents</strong></h2>



<p>A document can be seen as a container for content. The content is made up of static data and explicit and hidden information describing the relationships between the data which gives meaning to the document.  A document also has meta-data being the properties of the container itself.  Content is represented in documents in a way so that humans can process the content. Note that this has an important implication. Most documents were never designed to be read by bots. When data must be extracted it doesn’t matter for human workers if a document has a fixed structure like in a form, a semi-structure like in an invoice or no structure like in a contract. With some proper instructions human workers will be able to find the data they are looking for. Depending on how much structure there is, processing time may vary significantly of course. Bots however have no cognitive reading skills. And adding OCR and data capture technology to an RPA solution is often not enough to make bots really skilled at processing documents.</p>



<figure class="wp-block-image size-large"><img decoding="async" class="wp-image-711" src="https://skilja.com/wp-content/uploads/arlington-research-kN_kViDchA0-unsplash-1-1024x683.jpg" alt="" />
<figcaption>Photo by Arlington Research on <a href="https://unsplash.com/s/photos/business?utm_source=unsplash&amp;utm_medium=referral&amp;utm_content=creditCopyText" target="_blank" rel="noreferrer noopener">Unsplash</a></figcaption>
</figure>



<h2 class="wp-block-heading"><strong>Why OCR and Data Capture often do not offer the right reading skills for bots</strong></h2>



<p>The short answer: these technologies fall short because they were not developed for RPA. OCR was not designed to understand content. OCR is a technology for converting pixels into characters. Most OCR packages can also convert document images (scans) into text files while recreating the original layout. Data Capture systems use OCR technology and also many other AI technologies. Data capture systems were designed for extracting data from large volumes of documents with the highest possible accuracy. Neither OCR nor data capture were designed for RPA users to teach their bots how to read stuff. </p>



<p>The users who usually set up data capture installations are engineers who know exactly how to use all the levers and parameters of these systems. They often will create scripts or even add self coded additions in order to achieve the highest possible accuracy. The initial set up is high but that makes a lot of sense. The ROI of data capture installations doesn’t have to be fast. Almost always these installations are set up to run for many years. </p>



<p>Batch oriented data capture systems, although very powerful when setup correctly, logically are not the first choice of RPA users when they need to add document processing capabilities to their bots. These users are looking for simple, easy, fast and flexible functionality. The volumes they need to process are small. They also have higher needs for these systems to learn while doing because not the whole spectrum of variability in the documents that need to be processed will be available at the start of the project. And often they also need a higher level of intelligence because they want to automate tasks that were formerly performed by humans. And when they design new processes RPA users want to create intelligent bots that behave just like humans. But what RPA users cannot handle and what would break the RPA paradigm is if adding reading skills to bots comes at the expense of requiring a lot of investment and special technical skills like solution design, coding, production testing etc.</p>



<p>Note that in cases where RPA systems are used for processing large volumes of documents it makes sense to use data capture systems to extract data from these documents. Because in these cases efficiency aka accuracy most likely will play a significant role. Processing large volumes of documents is data capture’s sweet spot.</p>



<p>What RPA users really need when they have to process documents is something that may look a bit like OCR and data capture <strong>but is much smarter than that because it is operating a much broader set of AI technologies</strong> and at the same time it also should be easier to use.</p>



<h2 class="wp-block-heading">IDP: a new product category for a new market</h2>



<p>The massive pervasion of RPA installations in companies that happened during the past 3 to 4 years has led to an increasingly high demand for these easy, simple, flexible but yet powerful intelligent data capture solutions. This fast growing demand has spawned a new generation of companies who has gone down a different path using different technologies than what the incumbents in the data capture market have been doing for more than 20 years. These new systems are all based on the idea that Deep Neural Networks and other forms of Machine Learning are better and much easier ways to fulfill the needs of RPA users. All you need is a lot of samples, train your neural networks and off you go. What however is unclear at this stage is if ML can actually deliver the accuracy that is needed when bots should have humanlike reading skills executing mission critical tasks. When deep neural networks make mistakes you cannot go in and correct for these mistakes. These systems are black boxes. It is noteworthy that all incumbents are updating their existing offerings with adding ML technologies. Especially with the goal to become better at processing unstructured documents. And it seems not such a ridiculous assumption that these companies who have many years of experience developing document processing systems have an advantage over the competition that is fully ML focused. It looks quite plausible that incumbents are better set up to marry old AI and new AI technologies based on their solid understanding of how to build robust document processing systems.</p>



<p>All these developments of old and new companies to develop document processing skills for RPA bots have led to a new category of products that cater to the digital transformation market and not so much to the traditional capture market. The emergence of these new products was the reason for the Everest Group, a leading management consulting and research firm to come up with a new product category: <strong>Intelligent Document Processing or IDP</strong>.</p>



<p><a href="https://www.everestgrp.com/2019-12-understanding-enterprise-grade-idp-solutions-market-insights-52033.html">Everest Group defines IDP</a> as any software product or solution that captures data from documents (e.g., email, text, pdf, and scanned documents), categorizes, and extracts relevant data for further processing using AI technologies such as computer vision, OCR, Natural Language Processing (NLP), and machine/deep learning. These solutions are typically non-invasive and can be integrated with internal applications, systems, and other automation platforms.</p>



<figure class="wp-block-image size-large"><img decoding="async" class="wp-image-721" src="https://skilja.com/wp-content/uploads/markus-spiske-3Tf1J8q9bBA-unsplash-1024x683.jpg" alt="" />
<figcaption>Photo by Markus Spiske on <a href="https://unsplash.com/s/photos/business?utm_source=unsplash&amp;utm_medium=referral&amp;utm_content=creditCopyText" target="_blank" rel="noreferrer noopener">Unsplash</a></figcaption>
</figure>



<h2 class="wp-block-heading"><strong>About OCR and Online-learning</strong></h2>



<p>In many blogs about IDP authors create a contradiction between OCR being the old fashioned unintelligent way of data extraction and modern AI based extraction methods that are state of the art and intelligent. First of all as I pointed out OCR is not data extraction. OCR is one of many AI technologies that are used in data capture systems. And there are different ways we can build intelligence into systems that will help to find and interpret data in documents. Machine learning may perform better on unstructured documents. When dealing with forms and semi-structured documents systems that use templates, classifiers and other AI technologies including machine learning will almost always outperform systems that are solely based on machine learning. Note that these comprehensive systems operating ML in a smart way will learn from automated feedback through users correcting mistakes and from the results of successful and correct classification and extraction to generate additional knowledge (expanding the space) and statistics of usage of existing knowledge. See <a href="https://skilja.com/the-magic-of-online-learning/" target="_blank" rel="noreferrer noopener">The Magic of Online-Learning</a>.</p>



<h2 class="wp-block-heading"><strong>What is the ideal IDP solution?</strong></h2>



<p>The challenge for intelligent document processing is that there seems to be no one ideal approach. Depending on the type of documents, the volumes that have to be processed and the importance of accuracy ML or more traditional AI approaches or a blend thereof will be the best choice. For example when document volumes are big accuracy is important. When a shared service centre for example is processing 50 million documents a year improving accuracy from say 95% to 96% is significant. We just reduced the number of documents that have to be corrected by 500k. Another case where accuracy is critical is in straight-through processing.</p>



<p>It seems that the best option customers have is to adopt an IDP platform that enables them to operate different solutions or combinations of such solutions while shielding users from the complexity of these solutions. Vinna is such a platform that even allows to add crowdsourcing (Human In The Loop Verification) without users ever being aware of the complexity of what goes on under the hood.</p>



<p>&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;</p>



<div class="wp-block-image">
<figure class="alignleft size-large is-resized"><img loading="lazy" decoding="async" class="wp-image-654" src="https://skilja.com/wp-content/uploads/0-3.jpeg" alt="Jupp Stöpetie" width="170" height="170" /></figure>
</div>



<p class="has-text-align-left">Jupp Stöpetie as CEO of ABBYY Europe established ABBYY&#8217;s presence in the Western European markets, growing the brand and market presence to a leadership status for +25 years. His experience includes founding and growing companies and managing all levels of business operations, sales, and marketing. Jupp left ABBYY in spring 2020 and now works as an independent consultant based in Munich, Germany.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>IDP: Solving bot illiteracy in the digital workforce &#8211; Part 1</title>
		<link>https://skilja.com/idp-solving-bot-illiteracy-in-the-digital-workforce-part-1/</link>
		
		<dc:creator><![CDATA[Guest]]></dc:creator>
		<pubDate>Wed, 15 Jul 2020 10:11:47 +0000</pubDate>
				<category><![CDATA[Essentials]]></category>
		<category><![CDATA[Guest Post]]></category>
		<category><![CDATA[Market]]></category>
		<category><![CDATA[Process]]></category>
		<guid isPermaLink="false">https://skilja.com/?p=649</guid>

					<description><![CDATA[Editor’s note:&#160;This is a guest post from Jupp Stöpetie In this post we examine the role of Intelligent Document Processing (IDP) relative to Robotic Process Automation (RPA) and how these technologies drive Digital Transformation when combined. In part one we look at what is driving and enabling&#160;Digital Transformation. Part two then is dedicated to RPA [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p><span style="color:#7cb955" class="has-inline-color"><strong>Editor’s note</strong>:&nbsp;</span>This is a guest post from Jupp Stöpetie</p>



<p><em>In this post we examine the role of Intelligent Document Processing (IDP) relative to Robotic Process Automation (RPA) and how these technologies drive Digital Transformation <em>when combined</em>. </em> <em style="font-style: italic;">In part one we look at what is driving and enabling&nbsp;Digital Transformation. Part two then is dedicated to RPA and why IDP is essential.</em><i>&nbsp;</i></p>



<h2 class="wp-block-heading">Digital Transformation</h2>



<p>Companies all over the world are redesigning and digitizing their businesses at an ever faster pace and they have many reasons for doing so:</p>



<ul class="wp-block-list"><li>They want to serve their customers faster and better. </li><li>They understand that automation and using AI technologies improves innovation, agility, scalability and cost-efficiency</li><li>They appreciate that in contrast to human labor-intensive processes digital processes<ul><li>take less time to design</li><li>need less capital investment</li><li>are faster to deploy </li><li>and (much) less costly to run. </li></ul></li></ul>



<p>The above is commonly referred to as Digital Transformation. Citing Salesforce’s definition: “Digital Transformation (DX) is the process of using digital technologies to create new — or modify existing — business processes, culture, and customer experiences to meet changing business and market requirements.“&nbsp;</p>



<p>The main drivers of Digital transformation are the ongoing globalisation and the Fourth Industrial Revolution which has led to dramatically increased levels of competition. DX greatly improves the agility of businesses so they can adapt much faster than ever before to changes in the market. Changing or even completely redesigning digital processes is a lot easier and comes at much lower cost which results in significantly increasing a business’ competitiveness. Note that even terminating digitised processes comes at a much lower cost than when a lot of capital investment and labor was involved. Basically businesses have no choice. They must become digital. And those who are slow to change find themselves in an increasingly disadvantageous position. New businesses nowadays will always start with a digital concept in mind and by doing so will avoid manual processes if that is possible and makes sense.&nbsp;</p>



<figure class="wp-block-image size-large"><img decoding="async" src="https://skilja.com/wp-content/uploads/mike-kononov-lFv0V3_2H6s-unsplash-1024x576.jpg" alt="" class="wp-image-707"/><figcaption><span class="has-inline-color has-black-color">Photo by&nbsp;Mike Kononov&nbsp;on&nbsp;</span><a rel="noreferrer noopener" href="https://unsplash.com/s/photos/business?utm_source=unsplash&amp;utm_medium=referral&amp;utm_content=creditCopyText" target="_blank"><span class="has-inline-color has-black-color">Unsplash</span></a></figcaption></figure>



<h2 class="wp-block-heading"><strong>Market Data on Digital Transformation</strong></h2>



<p>What actually makes things very different from say ten years ago is the enormous progress. Worldwide spending on technologies and services that enable digital transformation (DX) by business practices, products, and organizations is forecast to reach $2.3 trillion in 2023, according to a new update to the International Data Corporation (<a href="http://www.idc.com/">IDC</a>) <a href="http://www.idc.com/getdoc.jsp?containerId=IDC_P32575" target="_blank" rel="noreferrer noopener">Worldwide Semiannual Digital Transformation Spending Guide</a>. DX spending is expected to steadily expand throughout the 2019-2023 forecast period, achieving a five-year compound annual growth rate of 17.1%. &#8220;We are approaching an important milestone in DX investment with our forecast showing the DX share of total worldwide technology investment hitting 53% in 2023,&#8221; said <a href="https://www.idc.com/getdoc.jsp?containerId=PRF005039" target="_blank" rel="noreferrer noopener">Craig Simpson</a>, research manager with IDC&#8217;s <a href="https://www.idc.com/promo/customerinsights" target="_blank" rel="noreferrer noopener">Customer Insights and Analysis Group</a>.</p>



<p>In their July report Grand View Research reported that the global Robotic Process Automation market size which is a segment of the overall DX market was valued at USD 1.40 billion in 2019 and is projected to exhibit a compound annual growth rate (CAGR) of 40.6% from 2020 to 2027.</p>



<h2 class="wp-block-heading"><strong>Why is Digital Transformation taking place now</strong></h2>



<p>In the last decade we have seen a tsunami of digital transformation projects driven by an accelerating desire of companies to increase competitiveness and innovation. But of course one could argue that businesses always had that desire. The question one could ask is: Why is all that happening now? Haven’t companies not been automating for decades already? Yes, but not at the current pace.&nbsp;</p>



<p><em>Note: at this stage it is unclear how the Covid-19 pandemic will influence market dynamics. It seems however unlikely that the need for companies to digitize their businesses will slow down. On the contrary, it is much more likely that the opposite will happen.</em></p>



<p>What has made things totally different from say ten years ago and has really enabled Digital Transformation to take place is the enormous progress in both computer science and in the computer industry. Compared with say 10 years ago we see:</p>



<ul class="wp-block-list"><li>an enormous increase in computational power</li><li>the rise of super powerful algorithms, algorithms that need incredible amounts of computational power which is now available</li><li>vastly improved connectivity both in speed and access points and it is improving at an accelerating pace (5G)</li><li>smartphones: in 2009 170 million smartphones were sold. In 2020 1.5 billion units are estimated to be sold. That makes for an estimated 3.5 billion people having a smartphone in 2020</li><li>huge and disproportionate &#8211; compared to the rest of the economy &#8211; amounts of money, that have been invested in tech companies. Successful tech companies have rewarded their investors&nbsp; with multiples that dwarfs any other industry.</li><li>and last but not least an insistently growing appetite of consumers and businesses for more, better and faster service anywhere at any time.</li></ul>



<h2 class="wp-block-heading"><strong>Some examples of how robotic process automation and document processing drive Digital Transformation in companies</strong></h2>



<p>A large international pharmaceutical company wanted to capture all details from their purchase orders and perform lookups and validation using their ERP. Tasks were automated using a RPA system. An intelligent document processing system was needed to read the data from the POs.</p>



<p>A large financial services company wanted to use RPA to automate their KYC process which involves capturing, verifying driver licences, passports etc. Early in the process they found that they also needed an intelligent document processing system.</p>



<p>A global logistics company wanted to automate their invoice processing (millions of documents). Their RPA system was found to be up to the challenge. But the company initially backed off because of the complexity of extracting data from millions of invoices (semi-structured) with a wide range of varying lay-outs. Only after a proof of concept clarified that there was an intelligent document processing system on the market that was up to the challenge the company proceeded with the project.</p>



<p>In part 2 of this post we will discuss how the combination of RPA with IDP works and what are the challenges that need to be considered.</p>



<p>&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;&#8212;</p>



<div class="wp-block-image"><figure class="alignleft size-large is-resized"><img loading="lazy" decoding="async" src="https://skilja.com/wp-content/uploads/0-3.jpeg" alt="Jupp Stöpetie" class="wp-image-654" width="170" height="170"/></figure></div>



<p class="has-text-align-left">Jupp Stöpetie as CEO of ABBYY Europe established ABBYY&#8217;s presence in the Western European markets, growing the brand and market presence to a leadership status for +25 years. His experience includes founding and growing companies and managing all levels of business operations, sales, and marketing. Jupp left ABBYY in spring 2020 and now works as an independent consultant based in Munich, Germany.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The Meaning of Words</title>
		<link>https://skilja.com/the-meaning-of-words/</link>
		
		<dc:creator><![CDATA[skiljaadmin]]></dc:creator>
		<pubDate>Wed, 18 Mar 2015 13:07:58 +0000</pubDate>
				<category><![CDATA[Essentials]]></category>
		<guid isPermaLink="false">https://skilja.com/?p=233</guid>

					<description><![CDATA[A famous poem by German poet Christian Morgenstern starts with the line that probably everybody has heard once:&#160;“Die Möwen sehen alle aus, als ob sie Emma hießen”&#160;which is in the translation by Karl F. Ross:&#160;“The seagulls by their looks suggest that Emma is their name”. And in fact, if you think about it, isn’t the [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>A famous poem by German poet Christian Morgenstern starts with the line that probably everybody has heard once:<em>&nbsp;“Die Möwen sehen alle aus, als ob sie Emma hießen”</em>&nbsp;which is in the translation by Karl F. Ross:<em>&nbsp;“The seagulls by their looks suggest that Emma is their name”</em>.</p>



<figure class="wp-block-image"><a href="http://www.skilja.de/wp-content/uploads/2015/03/seagull-638200_1280.jpg"><img decoding="async" src="http://www.skilja.de/wp-content/uploads/2015/03/seagull-638200_1280.jpg" alt="" class="wp-image-874"/></a></figure>



<p>And in fact, if you think about it, isn’t the name Emma associated with a certain impression of a person, of a situation of a sentiment? The same is true for many other names and expressions. Because our mind is implicitly associating a linguistic meaning to words based on the sound denoting a concept that is expressed by the word going beyond the pure description.</p>



<p>Another good example is given in the&nbsp;<a href="http://newyorker.com/tech/elements/the-power-of-names" target="_blank" rel="noreferrer noopener">New Yorker</a>, that inspired this article, which quotes in the issue from May 2013:&nbsp;<em>“Dawdle” and “meander” sound as unhurried as the walking speeds they describe, and “awkward” and “gawky” sound as ungainly as the bodies they represent</em>.</p>



<p>This observation led the German Gestalt-psychologist Wolfgang Köhler in 1929 to perform a psychological experiment to measure the non-arbitrary mapping between speech sounds and the visual shape of objects. In a famous experiment, which is now known as the Bouba/Kiki effect (see Wikipedia&nbsp;<a href="https://en.wikipedia.org/wiki/Bouba/kiki_effect" target="_blank" rel="noreferrer noopener">here</a>) Köhler showed forms similar to those&nbsp; below and asked participants which shape was called “takete” and which was called “maluma”.</p>



<figure class="wp-block-image"><a href="http://www.skilja.de/wp-content/uploads/2015/03/Maluma-Takete.png"><img decoding="async" src="http://www.skilja.de/wp-content/uploads/2015/03/Maluma-Takete.png" alt="" class="wp-image-875"/></a></figure>



<p>I am sure if you think about this for a moment you go with the vast majority of the respondents who associated the round and soft shape on the left with the word “malumba” while “takete” apparently seems to represent something jagged and sharp. In the following experiments using the words “kiki” and “bouba” 95% of participants selected the curvy shape as “bouba” and the jagged one as “kiki” in groups as diverse as American college undergraduates and Tamil speakers in India, implying that there seem to be deeper concepts to words we use. And if you are reading in a foreign language (like I do), didn’t it happen to you also that you didn’t know a word but somehow understood what it had to mean based on the sound?</p>



<p>You can see this effect also very clearly in the naming of products. Consider Clorox (producer of household bleach) and Chanel (high-end perfume). Switch the name and product and you get the idea. The words bear a subconscious meaning related to the sound. Sound elements in speech are called phonemes – which are related to morphemes in Grammar.</p>



<p>What has this to do with classification and document understanding?</p>



<p>Well, a good classifier will make use of these ideas to better train concepts and understand language. Meaning that it will not use a bag-of-words approach only, but break language down into smaller parts that allow to associate meanings with the phonemes. The Skilja Classifier follows this approach in extracting features from language.</p>



<p>Especially for short texts that are not very explicit and for sentiment analysis this method is far superior to using words. Take for example social media where words very often gradually change their meaning over time or new words are invented. But even if invented the new words follow the same sound pattern as existing words. We will use this approach for sentiment analysis based on pronunciation of quasi-words in Tweets. It will be interesting to see if the observed link between a meaning of a word and its sound can be statistically used to better understand sentiment.</p>



<p>Already today we see a significant difference in the quality of document classification if we compare pure word based classification with the richer statistics gained from morphemes. This is different from language to language with English being very sensitive while for example Mandarin does not show the effect at all as the symbols used for writing do not correspond to actual sounds.</p>



<p>We will follow up this interesting topic in number of future posts sharing ideas and results. For now let me just give you the final four lines of above poem by Morgenstern:</p>



<blockquote class="wp-block-quote has-text-align-center is-layout-flow wp-block-quote-is-layout-flow"><p>O human, you will never fly<br>the way the seagulls do;<br>but if your name is Emma, why,<br>be glad they look like you.</p></blockquote>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Visual Classifiers From Random Images</title>
		<link>https://skilja.com/visual-classifiers-from-random-images/</link>
		
		<dc:creator><![CDATA[Alexander]]></dc:creator>
		<pubDate>Mon, 03 Nov 2014 13:21:47 +0000</pubDate>
				<category><![CDATA[Cognition]]></category>
		<category><![CDATA[Essentials]]></category>
		<guid isPermaLink="false">https://skilja.com/?p=247</guid>

					<description><![CDATA[Now this is an interesting experiment that leads us very close to the touch point between machine classification and human imaginations. As described in previous posts, auto-classification algorithms are using features that are extracted from the objects to be classified (images or text) and are represented in a feature space. Classification can be described as [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>Now this is an interesting experiment that leads us very close to the touch point between machine classification and human imaginations.</p>



<p>As described in previous posts, auto-classification algorithms are using features that are extracted from the objects to be classified (images or text) and are represented in a feature space. Classification can be described as finding the correct separation plane (in many dimensions) between the features for different objects. A group of researchers from MIT (<a href="http://arxiv.org/find/cs/1/au:+Vondrick_C/0/1/0/all/0/1">Carl Vondrick</a>,&nbsp;<a href="http://arxiv.org/find/cs/1/au:+Pirsiavash_H/0/1/0/all/0/1">Hamed Pirsiavash</a>,&nbsp;<a href="http://arxiv.org/find/cs/1/au:+Oliva_A/0/1/0/all/0/1">Aude Oliva</a>,&nbsp;<a href="http://arxiv.org/find/cs/1/au:+Torralba_A/0/1/0/all/0/1">Antonio Torralba)&nbsp;</a>has now used an interesting approach to get a glimpse of how this feature space might actually exist and look like in our minds. They have generated random white noise in the feature space and inverted this noise to actual images. These images then have been presented to humans and they have been asked if they resembled certain well known objects. The results are quite fascinating:</p>



<div class="wp-block-image"><figure class="aligncenter"><a href="http://www.skilja.de/wp-content/uploads/2014/11/Acquiring-Visual-Classifiers-from-Human-Imagination-1.png"><img decoding="async" src="http://www.skilja.de/wp-content/uploads/2014/11/Acquiring-Visual-Classifiers-from-Human-Imagination-1.png" alt="" class="wp-image-700"/></a></figure></div>



<p>All image patches on the left are just noise. Many thousands of them were shown to online workers and they were asked them to find ones that look like cars. See full scientific paper&nbsp;<a href="http://arxiv.org/pdf/1410.4627v1.pdf" target="_blank" rel="noreferrer noopener">here</a>.</p>



<p>Most of the time these random images will appear to people as random. &nbsp;But every now and then somebody will feel that an image does remind them of a car. So we set this image aside. And repeat. After assessing 100,000 images in this way, we end up with a set of essentially random pictures that remind people of cars. We then take the average of these and find something interesting. The resulting image does indeed look like a blurry car, not a specific kind of car but a very&nbsp;general template&nbsp;of one.</p>



<p>Mathematically speaking this noise-driven method estimates the decision boundary that the human visual system uses for recognition.</p>



<div class="wp-block-image"><figure class="alignleft"><a href="http://www.skilja.de/wp-content/uploads/2014/11/Acquiring-Visual-Classifiers-from-Human-Imagination-2.png"><img decoding="async" src="http://www.skilja.de/wp-content/uploads/2014/11/Acquiring-Visual-Classifiers-from-Human-Imagination-2.png" alt="" class="wp-image-703"/></a></figure></div>



<p>My favorite example of the ones that have been tested is the fire hydrant that emerges from random white noise images.</p>



<p>Now as the random noise was actually generated in the feature space and not in the images themselves the researchers can deduct which features actually lead to the recognition of objects. And hence get an understanding on how human recognition of objects acually works. Because humans have some remarkable capabilities to recognize objects that they have never seen, touched or smelled before. This understanding of the actual feature selection in human minds will help us in the future to derive new classfiers that more closely resemble the way we all do object recognition and includes the human bias. Because this is one of the most interesting results of the study: The object that emerges depends on the cultural background of the persons selecting the random images. For example when online workers from India were asked to find a sports ball then a red circular object emerges. Because the most popular sport in India is cricket which is played with a red ball. Ask the same question to US workers then a orange ball appears – think of football or basketball.</p>



<p>The same question of human bias we find in our daily work in classification of documents in big enterprises. Every person will classify a document set a little different from their co-workers – thus leading to a lot of inconsistencies. Which can be overcome with auto-classification, which will always make predictable and consistent decisions.</p>



<p>The research described here provides a very interesting insight into the nature of human mind. It will also allow us to work on refined methods of classification that more closely resemble the way humans make decision.</p>



<p></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Practical Semantics</title>
		<link>https://skilja.com/practical-semantics/</link>
		
		<dc:creator><![CDATA[Alexander]]></dc:creator>
		<pubDate>Fri, 17 Jan 2014 13:30:00 +0000</pubDate>
				<category><![CDATA[Cognition]]></category>
		<category><![CDATA[Essentials]]></category>
		<guid isPermaLink="false">https://skilja.com/?p=258</guid>

					<description><![CDATA[As you have seen in previous posts I was pointing out the difficulty for software to really understand English language mainly lies in the ambiguity of words and their related meanings. The technology to resolve this difficulty is called semantic analysis. I recently found an interesting and funny text summarizing these diffculties. If you have [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>As you have seen in previous posts I was pointing out the difficulty for software to really understand English language mainly lies in the ambiguity of words and their related meanings. The technology to resolve this difficulty is called semantic analysis. I recently found an interesting and funny text summarizing these diffculties. If you have read this blog up to now you will easily be able to recognize, which of the words below are homonyms and which are polysemic, which can be resolved by syntax (e.g. same word as a noun and as a verb) and which need semantics. I was not able to find the original source, but many thanks to those who compiled it. Enjoy reading (I have highlighted my favorite in blue further down):</p>



<p><em>1) The bandage was&nbsp;<strong>wound</strong>&nbsp;around the&nbsp;<strong>wound</strong>.</em><br><em>2) The farm was used to&nbsp;<strong>produce produce</strong>.</em><br><em>3) The dump was so full that it had to&nbsp;<strong>refuse</strong>&nbsp;more&nbsp;<strong>refuse</strong>.</em><br><em>4) We must&nbsp;<strong>polish</strong>&nbsp;the&nbsp;<strong>Polish</strong>&nbsp;furniture..</em><br><em>5) He could&nbsp;<strong>lead</strong>&nbsp;if he would get the&nbsp;<strong>lead</strong>&nbsp;out.</em><br><em>6) The soldier decided to&nbsp;<strong>desert</strong>&nbsp;his&nbsp;<strong>dessert</strong>&nbsp;in the&nbsp;<strong>desert</strong>.</em><br><em>7) Since there is no time like the&nbsp;<strong>present</strong>, he thought it was time to&nbsp;<strong>present</strong>&nbsp;the&nbsp;<strong>present</strong>.</em><br><em>8 ) A&nbsp;<strong>bass</strong>&nbsp;was painted on the head of the&nbsp;<strong>bass</strong>&nbsp;drum.</em><br><em>9) When shot at, the&nbsp;<strong>dove</strong>&nbsp;<strong>dove</strong>&nbsp;into the bushes.</em><br><em>10) I did not&nbsp;<strong>object</strong>&nbsp;to the&nbsp;<strong>object</strong>.</em><br><em>11) The insurance was&nbsp;<strong>invalid</strong>&nbsp;for the&nbsp;<strong>invalid</strong>.</em><br><em>12) There was a&nbsp;<strong>row</strong>&nbsp;among the oarsmen about how to&nbsp;<strong>row</strong>.</em><br><em>13) They were too&nbsp;<strong>close</strong>&nbsp;to the door to&nbsp;<strong>close</strong>&nbsp;it.</em><br><em>14) The buck&nbsp;<strong>does</strong>&nbsp;funny things when the&nbsp;<strong>does</strong>&nbsp;are present.</em><br><em>15) A seamstress and a&nbsp;<strong>sewer</strong>&nbsp;fell down into a&nbsp;<strong>sewer</strong>&nbsp;line.</em><br><em>16) To help with planting, the farmer taught his&nbsp;<strong>sow</strong>&nbsp;to&nbsp;<strong>sow</strong>.</em><br><em>17) The&nbsp;<strong>wind</strong>&nbsp;was too strong to&nbsp;<strong>wind</strong>&nbsp;the sail.</em><br><em>18) Upon seeing the&nbsp;<strong>tear</strong>&nbsp;in the painting I shed a&nbsp;<strong>tear</strong>.</em><br><em>19) I had to&nbsp;<strong>subject</strong>&nbsp;the&nbsp;<strong>subject</strong>&nbsp;to a series of tests.</em><br><em>20) How can I&nbsp;<strong>intimate</strong>&nbsp;this to my most&nbsp;<strong>intimate</strong>&nbsp;friend?</em></p>



<p><em>Let’s face it – English is a crazy language. There is no egg in eggplant, nor ham in hamburger; neither apple nor pine in pineapple. English muffins weren’t invented in England or French fries in France . Sweetmeats are candies while sweetbreads, which aren’t sweet, are meat. We take English for granted. But if we explore its paradoxes, we find that quicksand can work slowly, boxing rings are square and a guinea pig is neither from Guinea nor is it a pig.</em></p>



<p><em>And why is it that&nbsp;<font color="blue"> writers write but fingers don’t fing</font>, grocers don’t groce and hammers don’t ham? If the plural of tooth is teeth, why isn’t the plural of booth, beeth? One goose, 2 geese. So one moose, 2 meese? One index, 2 indices? Doesn’t it seem crazy that you can make amends but not one amend? If you have a bunch of odds and ends and get rid of all but one of them, what do you call it?</em></p>



<p><em>If teachers taught, why didn’t preachers praught? If a vegetarian eats vegetables, what does a humanitarian eat? Sometimes I think all the English speakers should be committed to an asylum for the verbally insane. In what language do people recite at a play and play at a recital? Ship by truck and send cargo by ship? Have noses that run and feet that smell?</em><br><em>How can a slim chance and a fat chance be the same, while a wise man and a wise guy are opposites? You have to marvel at the unique lunacy of a language in which your house can burn up as it burns down, in which you fill in a form by filling it out and in which, an alarm goes off by going on.</em></p>



<p><em>English was invented by people, not computers, and it reflects the creativity of the human race, which, of course, is not a race at all. That is why, when the stars are out, they are visible, but when the lights are out, they are invisible.</em></p>



<p><em>PS. – Why doesn’t ‘Buick’ rhyme with ‘quick’?</em></p>



<p>Taken from:&nbsp;<a href="https://plus.google.com/+DavidDStanton/posts/dFxLRjeXmvb">https://plus.google.com/+DavidDStanton/posts/dFxLRjeXmvb</a></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Understanding the Weather</title>
		<link>https://skilja.com/understanding-the-weather/</link>
		
		<dc:creator><![CDATA[Alexander]]></dc:creator>
		<pubDate>Thu, 11 Apr 2013 14:00:47 +0000</pubDate>
				<category><![CDATA[Essentials]]></category>
		<guid isPermaLink="false">https://skilja.com/?p=272</guid>

					<description><![CDATA[If you live in Europe you will agree with me that the weather this spring has been the worst since a long time. In fact it was freezing and snowing all through March right into April. And it has not been much better on the East coast in the U.S. We experienced some really strange [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>If you live in Europe you will agree with me that the weather this spring has been the worst since a long time. In fact it was freezing and snowing all through March right into April. And it has not been much better on the East coast in the U.S. We experienced some really strange weather here in Germany, with cold air hovering over the country creating a permanent hazy mist in the air. And a lot of snow everywhere, even at the Easter holiday which normally is green. I have never experienced such conditions before.</p>



<p>But not only was the weather bad. Also the weather forecasts were very unreliable and often downright wrong. For a long time they predicted an imminent arrival of spring – and then it snowed the other day, then they predicted fog and snow when we actually had some sunny, although cold days. Why did this happen?</p>



<p>The answer is – because pattern matching failed. As many do not know, weather forecasts rely heavily on historical data. While meteorlogists will claim that they run complicated simulations on huge supercomputers to predict the weather, in fact most of the predictions are based on experience. Based on a huge collection of data similar meteorological situations of the pasts are searched and matched. Then the historically observed development is extrapolated to the future. More or less in the same way our grandmothers or every farmer is doing it. Don’t get me wrong, this approach still requires enormous computing power and elaborate algorithms, but – it fails if the situation is very unusual. This is what happened this year. Pattern matching failed, because there was no existing pattern to match with. What remains was the pure “model”. But weather is a highly non-linear phenomenon which cannot be predicted by a model, however complex it is. This has been brilliantly shown by nobel prize winner&nbsp;<a href="http://en.wikipedia.org/wiki/Ilya_Prigogine" target="_blank" rel="noreferrer noopener">Ilya Prigogine</a>&nbsp;in his book “The end of Certainty”.</p>



<p>We have a similar situation in document understanding. As we know, statistical systems that have been trained by a user (in so-called supervised learning), work well as long as the documents to process somehow match to what has been trained. But beware if there is an unexpected or unknown input. The statistical classifier will in the best case raise a flag, in the worst case produce a false positive result. Because pattern matching needs patterns to match with.</p>



<p>Fortunately human language is not as complex as the weather as it is fairly linear. Therefore it is possible to describe language with a model that can actually describe and understand the text. This is the&nbsp;<strong>Semantic Model</strong>. With syntactic and semantic analysis of text, software can go beyond pure matching and really understand text as a human does it. Therefore also unknown input can be reliably analyzed and correctly processed. Semantic analysis is the key to understanding text in the future.</p>



<figure class="wp-block-image"><a href="http://www.skilja.de/wp-content/uploads/2013/04/1-DSC08124.jpg"><img decoding="async" src="http://www.skilja.de/wp-content/uploads/2013/04/1-DSC08124-1024x768.jpg" alt="" class="wp-image-564"/></a><figcaption>Springtime in Freiburg – in the upper Rhine valley</figcaption></figure>



<p>P.S. Finally – as of today – we are back to normal with stable Atlantic west-wind conditions. Forecasts are very precise again and spring has arrived at last.&nbsp;&nbsp;Springtime in Freiburg – in the upper Rhine valley</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Why Semantics is Important for Classification</title>
		<link>https://skilja.com/why-semantics-is-important-for-classification/</link>
		
		<dc:creator><![CDATA[Alexander]]></dc:creator>
		<pubDate>Wed, 13 Mar 2013 14:04:00 +0000</pubDate>
				<category><![CDATA[Classification]]></category>
		<category><![CDATA[Cognition]]></category>
		<category><![CDATA[Essentials]]></category>
		<guid isPermaLink="false">https://skilja.com/?p=277</guid>

					<description><![CDATA[Many automatic classification systems out there today use a pure bag of words approach for finding relevant features that determine the meaning of a document. Few are using correlation and collocation – to account for the fact that words have a different meaning based on their context. None of them is using full semantic analysis [&#8230;]]]></description>
										<content:encoded><![CDATA[
<p>Many automatic classification systems out there today use a pure bag of words approach for finding relevant features that determine the meaning of a document. Few are using correlation and collocation – to account for the fact that words have a different meaning based on their context. None of them is using full semantic analysis of the meaning of words. But this is very much needed to be able to accurately classify a document.</p>



<div class="wp-block-group"><div class="wp-block-group__inner-container is-layout-flow wp-block-group-is-layout-flow">
<div class="wp-block-image"><figure class="alignright is-resized"><a href="http://www.skilja.de/wp-content/uploads/2013/03/1-DSC00461.jpg"><img loading="lazy" decoding="async" src="http://www.skilja.de/wp-content/uploads/2013/03/1-DSC00461-187x300.jpg" alt="" class="wp-image-523" width="209" height="328"/></a><figcaption>Ambiguous Text on Viking Ship</figcaption></figure></div>
</div></div>



<div class="wp-block-group"><div class="wp-block-group__inner-container is-layout-flow wp-block-group-is-layout-flow">
<p>The main reason is that (especially English) language is so ambiguous. English nouns have on average 5-8 close synonyms. There are words – example “strike” – that have more than 30 common meanings (strike a baseball, strike price buying stock, going on strike as an employee etc.). Now if you use a simple bag of words as features the software will never be able to make a clear distinction between an important fact (strike = work stoppage) and irrelevant information (baseball). Hence the classification result is also ambiguous and not very precise.</p>
</div></div>



<p>This can be solved by a full semantic analysis. A good definition of semantic is given by Wikipedia: “Linguistic semantics is the study of meaning that is used for understanding human expression through language”. Sounds exactly like what we want to achieve in document classification and understanding. Linguistic semantics is actually able to resolve the ambiguity of expressions and assign a unique meaning to the words. This is achieved by analyzing the relations the word has within a text.</p>



<p>The word “plant” is a good example. If used in the meaning of a factory then this is what you can do: You can enter the plant, you can build a plant or close it. But you cannot pick it or eat it. This is the other plant – the “living organism lacking the power of locomotion”. Thus by analyzing these relations a semantic analyzer can exactly determine the meaning of words – in the same way a human brain is doing it – rule based. Therefore we can make sure that we can distinguish the meaning of plants. Not to speak about Apple. – Just looking for Apple in a text without semantics you will find also all the rotten apples, apple pies and of course Big Apple in your results. Only semantics can identify Apple as a company through its relations in a text.</p>



<p>Using semantic analysis as a feature generator for classification greatly improves the precision of classification algorithm and at the same time allows distinguishing between important and irrelevant features in a text. In the same way you are doing through reading. It is obvious that future intelligent algorithms will need to use this technology.</p>



<p>In an upcoming post we will explain how semantic analysis can be used to generalize concepts to find topics which are another important aspect for good classification.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
