Jul

nmodes Technology - Overview

                                                       

nmodes ability to accurately deliver relevant messages and conversations to businesses is based on its ability to understand these messages and conversations. Once a system understands a sentence or text, it can easily perform a necessary action, i.e. bring a sentence about buying a car to the car dealership, or a complaint about purchased furniture to the customer service department of the furniture company.

Understanding sentences is called semantics. nmodes has developed a strong semantic technology that stand out in a number of ways.

Here is how nmodes technology is different:

1. Low computational power. We don’t use methods and algorithms deployed by almost everyone else in this space. The algorithms we are using allow us to achieve high level of accuracy while significantly reducing the computational power. Most accurate semantic systems, e.g. Google’s, or IBM’s, rely on supercomputers. By comparison our computational requirements are modest to the extreme, yet we successfully compete with these powerhouses in terms accuracy and quality of results.

2. Private data sources. We work extensively with Twitter and other social networks, yet at the same time we process enterprise data.  Working with private data sources means system should know details specific only to this particular data source. For example, when if a system handles web self-service solution for online electronics store it learns the names, prices, and other details of all products available at this store.  

3. User driven solution. Our system learns from user’s input. Which makes it extremely flexible and as granular as needed. It supports both generic topics, for example car purchasing, and conversations concentrating on specific type of car, or a model.

Interested in reading more? Check out our other blogs:

The Automation Is Coming

                                                         

A close look at the history of humanity leaves us with no choice but to admit that the majority of jobs as we know them now will be transferred to automated systems. This is part of the technological and scientific progress our civilization is undertaking and it is irreversible.

Artificial intelligence became mainstream in 2016. For the first time artificial intelligence is not only available to big companies like Google, Amazon or Apple, but to the majority of businesses worldwide.  Startups have started building products and services using artificial intelligence en masse.

The essence of artificial intelligence is massive, intuitive computing power: machines so smart that they can learn and become even smarter.  The machines are becoming quicker and more nimble. They cover wider range of conversation topics. They now connect to robotic systems and online interactive systems. There is literally very little they cannot, or will not be able to, do as applied to industrial workforce.

With all the good that’s going to come with automation, we are suddenly faced with a new problem: the elimination of many low and middle class jobs. Many jobs that have already been severely impacted by computers (manufacturing, administrative support, retail, and transportation) will continue to diminish. In the nearest future routine-based jobs (telemarketing, sewing) and work that can be solved by smart algorithms (tax preparation, data entry keyers and insurance underwriters) are most likely to be eliminated.

What to do? It is fruitless to fight automation, we need to find ways to work with automation rather than against it.

The solution is to become more creative as species. Creativity is the natural advantage of humans over machines. Automation is about to change the course of the world, it’s going to be a great disruptor and impact the workforce like nothing we’ve seen before. We can sit around and gradually become obsolete, or accept the challenge and use the tool of creativity, which we are in unique possession of, to maintain our superiority.

 

READ MORE

Towards smarter data - accuracy and precision

                                                   

There is a huge amount of information out there. And it is growing. To make it efficient and increase our competitive advantage we need to evolve and start using information in a smart way, by concentrating on data that drives business value because it is accurate, actionable, and agile. Accuracy is an important measure that determines the quality of data processing solutions.

How accuracy is calculated?

It is easy to do with structured data, because the requirements are formalizable. It is less obvious with unstructured data, e.g. a stream of social feeds, or any data set that involves natural language. Indeed, the sentences of natural language are subject to multiple interpretations, and therefore allow a degree of subjectivity. For example, should a sentence ‘I haven’t been on a sea cruise for a long time’ be qualified for a data set of people interested in going on a cruise? Both answers, yes and no, seem valid.

In these cases an argument was put forward endorsing a consensus approach which polls data providers is the best way to judge data accuracy. This approach essentially claims that attributes with the highest consensus across data providers is the most accurate.

At nmodes we deal with unstructured data all the time because we process natural language messages, primarily from social networks. We do not favor this simplistic approach, as it is considered biased, inviting people to make assumptions based on what they already believe to be true, and making no distinction between precision and accuracy. Obviously the difference is that precision measures what you got right, and accuracy measures both what you got right and what you got wrong. Accuracy is a more inclusive and therefore more valuable characteristic.

Our approach is

a) to validate data against third party independent sources (typically of academic origin) that contain trusted sets and reliable demography. Validating nmodes data against third party sources allows us to verify that our data achieves the greatest possible balance of scale and accuracy.

b) to enrich upon the existing test sets by purposefully including examples ambiguous in meaning and intent, and providing additional levels of categorization to cover these examples.

Accuracy is becoming important when businesses move from rudimentary data use, typical of the first Big Data years, to a more measured and careful approach of today. Understanding how it is calculated and the value it brings helps in achieving long-term sustainability and success.

 

READ MORE