但為Alexa的學識進行眾包是個好主意嗎？從種族主義網絡釣魚者“策反”的微軟聊天機器人，到類似Alexa Answers但因充斥不良信息而惡名在外的Yahoo Answers，用戶提供數據的系統走入歧途的案例在過去幾年隨處可見。因此不難想象出最糟糕的情景：在Alexa操控下，智能音箱“愉快地”播報著假新聞、危險的陰謀論或者白人至上主義者的觀點。
亞馬遜負責Alexa Information的副總裁比爾·巴頓向商業雜志《Fast Company》描述Alexa Answers時態度樂觀。他說：“我們注入的是貢獻者的正能量和善意，我們用機器學習和算法來排除少數聒噪者和壞家伙。”
實際上，從愛因斯坦的襪子到芥末的顏色，Google Assistant目前可以回答幾乎所有上述問題，只是它們都直接取自Amazon Answers網站。谷歌的答案來自谷歌搜索引擎找到的結果、谷歌featured snippet以及知識圖譜。亞馬遜想利用公眾提供的答案在這個領域追上谷歌。
除了為每個問題找到答案，來自Alexa Answers的數據還將用于訓練亞馬遜語音助手背后的人工智能系統。該公司發言人告訴《財富》雜志：“Alexa Answers不僅是擴展Alexa知識面的另一條途徑，還可以讓她更好地幫助其他用戶并為他們提供更多信息。”當初發布Alexa Answers時，亞馬遜將其稱為“變得更聰明的”Alexa。
但自動化系統可以輕松識別并阻攔個人發表的冒犯性語句，盡管在這方面也有負面風險。在一次測試中，筆者在Alexa Answer上回答問題時想提到‘20世紀90年代的搖滾樂隊Porno for Pyros’，結果答案被拒，原因不是不準確，而是因為包含了‘porno’這個詞。系統提示稱：“Alexa不會用這個詞”。
網絡釣魚并非Alexa在精神衛生方面面臨的唯一風險。如果Alexa沒有正確解讀提問者的話語，就連初衷良好的問題也可能變得荒謬。比如，上周五上午在Alexa Answers出現了一個問題，內容是“What is a piglet titus?”。看來用戶實際上問的是“What is Epiglottitis?”（答案是：急性會厭炎，一種罕見咽喉疾病）。如果有足夠多的用戶嘗試回答這個毫無意義的問題，比如小熊維尼的粉絲或者急于獲得分數的用戶，他們就可能讓數據池變得混亂，而不是得到改善。
Did Albert Einstein wear socks? How do you prevent tears when cutting an onion? Did Burt Reynolds marry Sally Field? What makes wasabi green? The average person might not know the answer to these questions, but Amazon Alexa, through the new Alexa Answers portal that was announced Thursday, might. Well, more accurately, an Alexa user could.
An online community where anyone who logs in can suggest answers to user-supplied questions posed to the voice-activated Alexa A.I. assistant, Alexa Answers is designed to answer the tough questions that can’t already be answered by the voice-enabled assistant. Once the answers are submitted, they are vetted for accuracy, scored, and if they are good enough, make their way back to Alexa users.
But is crowdsourcing Alexa's smarts a good idea? From a Microsoft chatbot subverted by racist trolls to Yahoo Answers, a similar service to Alexa Answers that has become notoriously rife with bad information, the past few years have been littered with cases of user-generated data systems gone bad. So it's not hard to imagine the worst-case scenario: an Alexa-backed smart speaker blithely spouting fake news, dangerous conspiracy theories, or white supremacist talking points.
Describing Alexa Answers to Fast Company, Bill Barton, Amazon’s Vice President of Alexa Information, struck an optimistic tone. “We’re leaning into the positive energy and good faith of the contributors," he said. "And we use machine learning and algorithms to weed out the noisy few, the bad few.”
Experts on data use and its impacts are markedly less cheery.
“We have plenty of examples of why this is not going to play out well,” says Dr. Chris Gillard, who studies the data policies of Amazon and other tech companies at Macomb Community College near Detroit. Crowdsourcing data, and then using that data in training the Alexa algorithm, he says, presents “pitfalls that Amazon seem intent on stepping right into.”
The race to beat Google
While better assistants and smart speakers drive sales of accessories like voice-activated lights, Google’s decades in the search business seem to have given it an advantage over Amazon when it comes to understanding queries and returning data. Google's smart speaker has steadily gained market share against the Echo, and Google Assistant has almost uniformly outperformed Alexa in comparison tests.
In fact, almost all of the questions above, from Einstein's socks to wasabi's color, are are currently answered with Google Assistant, though they were taken directly Amazon Answers' website. Google's answers come from its search engine's results, featured snippets, and knowledge graph. Amazon is trying to use crowd-supplied answers to catch up in this space.
“Amazon’s not Google,” says Dr. Nicholas Agar, a technology ethicist at Victoria University of Wellington, New Zealand. “They don’t have Google’s [data] power, so they need us.”
Beyond just providing missing answers to individual questions, data from Alexa Answers will be used to further train the artificial intelligence systems behind the voice assistant. “Alexa Answers is not only another way to expand Alexa's knowledge,” an Amazon spokesperson tells Fortune, “but also... makes her more helpful and informative for other customers.” In its initial announcement of Alexa Answers, Amazon referred to this as Alexa “getting smarter.”
Money for nothing, facts for free
As important as Alexa Answers might be for Amazon, contributors won’t get any financial compensation for helping out. The system will have human editors who are presumably paid for their work, but contributed answers will be rewarded only through a system of points and ranks, a practice known in industry parlance as ‘gamification.’
Agar believes this will be effective, because Amazon is leveraging people’s natural helpfulness. But he also thinks a corporation leveraging those instincts should give us pause. “There’s a difference between the casual inquiry of a human being, and Amazon relying on those answers," he says. "I think it’s an ethical red flag.”
Gillard also thinks Amazon should pay people to provide answers, whether its one of its own workers or partner with an established fact-checking group.
Amazon certainly has the infrastructure to do it. The ecommerce giant already runs Mechanical Turk, a ‘gig’ platform that pays “Turkers” for performing small, repetitive tasks, and would seem well-suited to supplementing Alexa’s training.
But Gillard believes that relying on a ‘community’ model insulates Amazon if Alexa starts spouting bad or offensive answers, based on crowd input. “I think not paying people lets you say, well, it was sort of the wisdom of the crowd,” he says. “If you pay people, you’re going to be accused of bias.”
A gamified incentive system, though, is not without its own risk. In 2013, Yahoo Answers disabled part of its user voting system. That's allegedly because some participants created fake accounts to upvote their own (not necessarily accurate) answers. (Source: Quora. Also, this is a good example of how crowd-sourcing information impacts reliability.)
The biggest question facing Alexa Answers is whether Amazon can effectively prevent abuse its new platform. Amazon declined to answer questions from Fortune about the precise role of human editors in the system. But their presence alone represent an acceptance that automated systems in their current state can't reliably detect offensive content, or evaluate the accuracy of facts.
Amazon has never grappled with these challenges as directly as companies like Facebook and Twitter, though according to some critics, it has failed even to consistently detect fake reviews in its own store. Barton told Fast Company that Amazon will try to keep political questions out of the system, a subtle task Gillard says will likely fall to humans. “A.I. can’t do those things," he says, "It can’t do context.”
Yet automated systems can easily detect and block individual offensive terms, though even that has its downsides. In a test, this reporter attempted to reference the ‘90s rock band Porno for Pyros when suggesting an Alexa Answer. The answer was rejected, not because of inaccuracy, but because of the word ‘porno.’ According to a notification, “Alexa wouldn’t say that.”
Not everything has an answer
Barton told Fast Company that “we’d love it if Alexa can answer any question people ask her,” but that’s clearly impossible. Alexa cannot be expected to know, for instance, what the meaning of life is, and crowdsourcing answers to questions that are enigmas could make the entire system more fragile. In a 2018 study, researchers found that search queries with limited relevant data, which they called “data voids,” were easier for malicious actors to spoof with fake or misleading results.
And trolls aren’t the only risk to Alexa’s mental hygiene. Even well-intentioned questions can wind up nonsensical, if Alexa doesn’t properly interpret the questioner’s speech. For example, the question “What is a piglet titus?” appeared on Alexa Answers Friday morning. It seems likely the user actually asked “What is Epiglottitis?” (Answer: a rare throat condition). If enough users tried to answer the nonsense question—perhaps Winnie the Pooh fans, or users hungry for points—it could muddy the data pool, instead of improving it.
It’s unclear how Alexa's overall performance might be impacted by messy or malicious data—those answers are a ways away yet. Bit it's a wonder if, after all the stumbles of similar systems, Amazon is taking the risks of crowdsourced answers seriously.