Jun

Abundance of Information Often is a Liability

A massive change has occurred in the world during the last ten to twenty years. Until recently and throughout the history of mankind information was hard to access. Obtaining and sharing information was either a laborious process or impossible, and the underlying assumption was that information can never be enough.

Today, of course, we have the opposite picture. Not only information is easily available, it keeps pouring in from a growing number of sources, and we continuously find ourselves in situations when there is more information than we want or able to process.

A major task we, as species, are facing is therefore how to reduce or filter out relevant information. It is, to repeat, in direct opposition to the task we’ve been accustomed to during all previous centuries, which was how to obtain information.

Since this change took place only recently, within a lifetime of one generation, we didn’t have time to develop efficient set of procedures to address the new problem. But the work has started and will only accelerate with time.

Interested in reading more? Check out our other blogs:

AI unmasked: Have chatbots failed?

It is becoming increasingly popular to say that chatbots have failed and are overhyped.

While it is true that in many cases expectations from chatbots significantly exceed the results on the ground, the anticipation of chatbots’ demise are somewhat premature. 

One of the main problems for chatbots is that the market is inundated with low quality solution providers who deliver low quality results. This happened because conversational AI seems to have low entry barriers. Unlike other recent technological darlings such as space technology or renewable energy, conversational AI is purely software and therefore does not require vast sums of initial investment. 

What this approach is missing however,  is that conversational AI, in addition to being a software, also requires an accurate understanding of how language works. And there is a limited number of people in the world that do have such understanding.

When conversational AI is delivered by AI experts who understand the way human language works, the results are good and convincing, just as how you would expect them to be.

Suffering from unsatisfactory product quality is a common problem for many new and emerging industries.  The rules of the market dictate that most of the low quality players will eventually disappear. Poorly created chatbots will therefore not be around for too long.

READ MORE

Beware the lure of crowdsourced data

Crowdsourced data can often be inconsistent, messy or downright wrong 

We all like something for nothing, that’s why open source software is so popular. (It’s also why the Pirate  Bay exists). But sometimes things that seem too good to be true are just that. 

Repustate is in the text analytics game which means we needs lots and lots of data to model certain  characteristics of written text. We need common words, grammar constructs, human-annotated corpora  of text etc. to make our various language models work as quickly and as well as they do. 

We recently embarked on the next phase of our text analytics adventure: semantic analysis. Semantic  analysis the process of taking arbitrary text and assigning meaning to the individual, relevant components.  For example, being able to identify “apple” as a fruit in the sentence “I went apple picking yesterday” but to  identify “Apple’ the company when saying “I can’t wait for the new Apple product announcement” (note:  even though I used title case for the latter example, casing should not matter)

READ MORE