Last mile for web search
The last few decades have been dominated by a certain search paradigm - the user enters keywords into a search box and gets back a number of links as the result. The search engine did all the hard work to sift through the enormous pool of the information that is on the web, but left the last mile to the user - the job of filtering the links and the ads to a few relevant ones, opening them and then taking the time to consume the content and deduct the actual information needed.
While the current search paradigm is still appropriate for some tasks, there are growing numbers of situations in our increasingly busy and changing lives that require a new paradigm - returning a short, to-the-point answer to our questions. It is not that hard to imagine that in the future, the AIs will take the job of answering even the hardest questions we may ask them by digesting pages of documents, even entire books, only to 'compress' the answer into a given budget of words or sentences.
We are witnessing this technology in action today in the form of AI assistants launched by all major tech companies. With Kagi.ai, we are excited to join the race.
"Sorry, I didn't quite get that" is the primary source of frustration with users when using AI assistants today. When the users do not get at least an attempt at an answer, they are left without a path to proceed and start to gradually lose interest in the platform. This leads to the effect where the engagement with AI assistants converges to weather, time and music - definitely not what the creators of these platforms envisioned! This in turn causes the AI assistant adoption to suffer.
We believe that the ability to answer any kind of general questions, including open-ended, is table stakes for any AI assistant or search engine in the future. This is the crucial ingredient to keeping the engagement levels up.
In order to better asses the current situation we present our research in tracking the answering accuracy of all the state-of-the-art AI assistants on the market today, including Google, Siri, Alexa, Cortana, Bixby and SoundHound. The test consists of randomly sampled 100 natural language factual questions sourced from relevant public research data sets. To score correctly, the AI must provide a direct, non-ambiguous answer to the test question.
We present the results below.
Current AI factual question-answering accuracy
(higher is better; last updated July 2019)
If we may, we'd like to provide short commentary on these results.
Apple is doing some wonderful things with understanding and communicating emotions and is probably leading the pack with social-context-aware communication. But this focus, while important, caused Siri to trail the pack by a large margin with its general question answering abilities.
Alexa pioneered the voice-first paradigm and is detailed when the answer is in her knowledge base, but that does not happen as often as it should for a platform as wide-spread. The users are also starting to get lost in the skills jungle and the engagement seems to be converging towards weather, music and time the quickest.
Not surprisingly, Google does best among peers in our test. The progress has been steady over the last 6 months since we have been tracking the performance. But there is a fundamental business model question on the horizon - if Google was able to create a perfect AI that can answer any question with 100% accuracy, who would be left to click on ads?
We have witnessed Microsoft and Samsung make big leaps forward in the quality of answers and range of questions they can address, and steadily gaining ground on competition. But when you are thinking about AI assistants, are you thinking about Cortana and Bixby?
And wouldn't it be nice if my AI was be able to answer a simple question like "How to approach a girl I like?"?
Do you have wise friends/colleagues that can clearly convey something in a few sentences? We all do. Being that wise friend that you can always rely on for clear and concise answer is our goal with Kagi.
You can see some of this in action if you tried our technology demo. Now, imagine for a moment that you could break down each of those answers and sentences and then explore them individually, unrolling new ideas and concepts? It is not just about how we receive the information but also how we use it to broaden our horizons.
We want to establish the connection between the human and a machine at the magic and emotive level where the HCI interaction happens. We are thinking deeply what the essence of human and machine dialogue should be.
Kagi and the new paradigm of AI
At Kagi, we believe that combination of the following two factors will effectively remove the barrier for AI assistant adoption and enable a new search paradigm to emerge:
Let's deconstruct this.
With Kagi we've basically eliminated the 'Sorry, I didn't quite get that' responses. Our approach allows us to take advantage of the entirety of available human knowledge. This is in contrast to using methods like scripted responses or knowledge bases, which by design have a horizon of nature and types of questions they are able to answer. It also makes them hard to scale and when they do, they threaten to collapse under the weight of their own complexity.
There are vast amounts of information currently available on the web and Kagi is attempting to make sense of all that. It acts as gateway to the giant repository of human knowledge, always happily waiting to serve our requests. We have set the goal of 80% answering accuracy as we believe this is the tipping point where engagement is unleashed. We are not quite there yet (our current accuracy will be revealed in a minute) but there is a visible road ahead.
You may ask why not shoot for 100%? It turns out there are certain limitations in the way we as humans interact and process information, that makes it impossible to always get the answer that we expect. Let's give an example that we encountered in the early days of working on Kagi.ai.
We were training Kagi to answer yes/no questions and one that we asked was "Do cats have wings?". To our surprise it responded "Yes", and when we debugged why, it turns out that under certain circumstances cats can grow something resembling wings (it is documented online) and people were calling it "wings". So we rephrased the question to better match our original intention and asked "Can cats fly?" and it again responded "Yes"! This time we were sure it was a mistake, but when we debugged it, it turns out that cats can 'fly' - if you buy them a plane ticket... This would make a good joke in human company, let alone in AI's. We decided to capture the moment with this cartoon.
Also sometimes meaning is not simple to derive from just the original question itself. Take the following example - when were the Twin towers built?
These make for interesting challenges and it is going to boil down to the human computer interaction design as we discussed before.
To summarize, our approach represents a paradigm where AI's scope of the possible inputs is kept as open and as general as possible, while working on improving the quality and accuracy of outputs. We could say that it is a top-down approach. This is in contrast to a bottoms-up approach largely in use today, where the user is forced to memorize commands and types of interactions and questions with AI that are allowed.
With that, we are excited to put our current results into perspective.
Current AI factual question-answering accuracy
(higher is better; last updated July 2019)
Your Future AI
Lets take a moment and envision a future where AI assistants might be graded by their ability, and as a result potentially come at different price points. You can imagine being able to purchase a beginner, advanced and an expert level AI. They will come with character traits, different tact, charm and wit, certain 'pedigree' and 'interests' and have controllable bias - like for example you will be able to choose between an AI with a conservative or liberal (or agnostic!) view of the world.
In this future, instead of everyone sharing the same Siri, we will own our truly own Mike or Julia, or maybe Donald - the AI. And when you ask your own AI a question like "does God exist?" it will answer it relying on biases you preconfigured. When you ask it to recommend a good restaurant nearby, it will do so knowing what kind of food you like to eat. The same will happen when you ask it to recommend a good coffee maker - it will know the brands you like, your likely budget and the kind of coffee you usually drink. All this information will be volunteered to the AI by you - similar to how you would volunteer your information to a human assistant - but this time to a much larger extent. And you will also do it without fear as the business models will change away from ad-driven ones. This will make AI assistants indispensable in our future even busier lives.
We invite you to play with Kagi and have some fun asking it weird questions you would never ask Siri or Alexa. Also if you like what we are building, please share it with your friends. Thank you.