员工呼入并提供行程信息以保存在数据库中

时间:2022-03-09 16:55:29

I would like to code something up where my employees can call in and Watson will ask them the important questions, and they can just tell Watson the information and Watson then output that information into a CSV, XLS or etc. format possibly even a database.

我想在我的员工可以打电话的地方编写代码,Watson会问他们重要的问题,他们可以告诉Watson信息,然后Watson将这些信息输出为CSV,XLS等格式,甚至可能是数据库。

It seems that I should be able to do this because of the way it can converse with people through messenger etc.

似乎我应该能够做到这一点,因为它可以通过信使等方式与人交谈。

I know it is probably a 3 pronged approach.

我知道这可能是一个三管齐下的方法。

Ideas?

2 个解决方案

#1


1  

@Florentino DeLaguna, in this case, you can use Conversation Service and Text to Speech and Speech to text API's from IBM Watson. See options you can use for that:

@Florentino DeLaguna,在这种情况下,您可以使用Conversation Service和Text to Speech and Speech来从IBM Watson发送文本API。查看可用于此的选项:

  1. In theory, you would have to built an application that integrates with one URA (using Asterisk for example), convert the Speech to Text, send that text for Conversation Service, and the response of the Conversation you would have to transform into voice and send to the URA . In practice, there are some conversational problems, especially from Speech to Text. But the return voice you can use some effects using the IBM Watson Text to Speech (faster and slower voices, control of pauses, put emotions ...).
  2. 从理论上讲,您必须构建一个与一个URA集成的应用程序(例如使用Asterisk),将Speech转换为Text,将该文本发送给Conversation Service,以及您必须转换为语音和发送的Conversation的响应到市建局。在实践中,存在一些会话问题,尤其是从语音到文本。但是返回声音你可以使用IBM Watson Text to Speech(更快更慢的声音,控制暂停,放置情感......)来使用一些效果。

Obs.: The URA audios are in narrowband, 8khz, and most Speech to Text services only accept broadband, 16khz.

Obs。:URA音频是窄带,8khz,大多数语音到文本服务只接受宽带,16khz。

Obs II.: You app (like Asterisk) need to be able to consume a REST API and / or make use of Web Sockets then it will be able to invoke the Watson Speech to Text service.

Obs II。:您的应用程序(如Asterisk)需要能够使用REST API和/或使用Web套接字,然后它才能调用Watson Speech to Text服务。

  1. Another option is to route a call out of Asterisk to the new IBM Voice Gateway which is a SIP endpoint that fronts a Watson self-service agent by orchestrating Speech To Text, Text To Speech and the Watson Conversation service. You can think of IBM Voice Gateway as a stand alone, Cognitive IVR system. Go here for more details.

    另一个选择是将呼叫从Asterisk路由到新的IBM语音网关,这是一个SIP端点,通过编排语音到文本,文本到语音和沃森对话服务来面向Watson自助服务代理。您可以将IBM Voice Gateway视为独立的Cognitive IVR系统。到这里了解更多详情。

  2. Another potential option is to use MRCP. IBM has a services solution that will allow you to reach the Watson STT and TTS engines using MRCP. Not sure if Asterisk supports MRCP but that is typically how traditional IVRs integrate with ASRs.

    另一个可能的选择是使用MRCP。 IBM有一个服务解决方案,允许您使用MRCP访问Watson STT和TTS引擎。不确定Asterisk是否支持MRCP,但这通常是传统IVR与ASR集成的方式。

Important: The options 2 and 3 are answered for another person, see the official answer.

重要提示:选项2和3可供其他人使用,请参阅官方回答。

See more about these API's:

了解有关这些API的更多信息:

#2


1  

Have a look to the Voximal solution, it integrates all the SpeechToText Cloud API (and TextToSpeech) as an Asterisk application throw a VoiceXML standard browser.

看看Voximal解决方案,它集成了所有SpeechToText Cloud API(和TextToSpeech)作为Asterisk应用程序抛出一个VoiceXML标准浏览器。

All is integrated in the VoiceXML interpreter, you got the full text result of the transcription, and you can push it to a chatbot to detect the intent of the users and pick dynamic parameters like date, number, city, and more... for example by using api.ai.

所有都集成在VoiceXML解释器中,您获得了转录的全文结果,您可以将其推送到聊天机器人以检测用户的意图并选择动态参数,如日期,数字,城市等等...例如,使用api.ai.

Voximal supports the STT from Google, Microsoft, IBM/watson (and soon Amazon).

Voximal支持谷歌,微软,IBM /沃森(以及很快亚马逊)的STT。

The 3 API listed by Sayuri are embedded in the solution.

Sayuri列出的3 API嵌入在解决方案中。

#1


1  

@Florentino DeLaguna, in this case, you can use Conversation Service and Text to Speech and Speech to text API's from IBM Watson. See options you can use for that:

@Florentino DeLaguna,在这种情况下,您可以使用Conversation Service和Text to Speech and Speech来从IBM Watson发送文本API。查看可用于此的选项:

  1. In theory, you would have to built an application that integrates with one URA (using Asterisk for example), convert the Speech to Text, send that text for Conversation Service, and the response of the Conversation you would have to transform into voice and send to the URA . In practice, there are some conversational problems, especially from Speech to Text. But the return voice you can use some effects using the IBM Watson Text to Speech (faster and slower voices, control of pauses, put emotions ...).
  2. 从理论上讲,您必须构建一个与一个URA集成的应用程序(例如使用Asterisk),将Speech转换为Text,将该文本发送给Conversation Service,以及您必须转换为语音和发送的Conversation的响应到市建局。在实践中,存在一些会话问题,尤其是从语音到文本。但是返回声音你可以使用IBM Watson Text to Speech(更快更慢的声音,控制暂停,放置情感......)来使用一些效果。

Obs.: The URA audios are in narrowband, 8khz, and most Speech to Text services only accept broadband, 16khz.

Obs。:URA音频是窄带,8khz,大多数语音到文本服务只接受宽带,16khz。

Obs II.: You app (like Asterisk) need to be able to consume a REST API and / or make use of Web Sockets then it will be able to invoke the Watson Speech to Text service.

Obs II。:您的应用程序(如Asterisk)需要能够使用REST API和/或使用Web套接字,然后它才能调用Watson Speech to Text服务。

  1. Another option is to route a call out of Asterisk to the new IBM Voice Gateway which is a SIP endpoint that fronts a Watson self-service agent by orchestrating Speech To Text, Text To Speech and the Watson Conversation service. You can think of IBM Voice Gateway as a stand alone, Cognitive IVR system. Go here for more details.

    另一个选择是将呼叫从Asterisk路由到新的IBM语音网关,这是一个SIP端点,通过编排语音到文本,文本到语音和沃森对话服务来面向Watson自助服务代理。您可以将IBM Voice Gateway视为独立的Cognitive IVR系统。到这里了解更多详情。

  2. Another potential option is to use MRCP. IBM has a services solution that will allow you to reach the Watson STT and TTS engines using MRCP. Not sure if Asterisk supports MRCP but that is typically how traditional IVRs integrate with ASRs.

    另一个可能的选择是使用MRCP。 IBM有一个服务解决方案,允许您使用MRCP访问Watson STT和TTS引擎。不确定Asterisk是否支持MRCP,但这通常是传统IVR与ASR集成的方式。

Important: The options 2 and 3 are answered for another person, see the official answer.

重要提示:选项2和3可供其他人使用,请参阅官方回答。

See more about these API's:

了解有关这些API的更多信息:

#2


1  

Have a look to the Voximal solution, it integrates all the SpeechToText Cloud API (and TextToSpeech) as an Asterisk application throw a VoiceXML standard browser.

看看Voximal解决方案,它集成了所有SpeechToText Cloud API(和TextToSpeech)作为Asterisk应用程序抛出一个VoiceXML标准浏览器。

All is integrated in the VoiceXML interpreter, you got the full text result of the transcription, and you can push it to a chatbot to detect the intent of the users and pick dynamic parameters like date, number, city, and more... for example by using api.ai.

所有都集成在VoiceXML解释器中,您获得了转录的全文结果,您可以将其推送到聊天机器人以检测用户的意图并选择动态参数,如日期,数字,城市等等...例如,使用api.ai.

Voximal supports the STT from Google, Microsoft, IBM/watson (and soon Amazon).

Voximal支持谷歌,微软,IBM /沃森(以及很快亚马逊)的STT。

The 3 API listed by Sayuri are embedded in the solution.

Sayuri列出的3 API嵌入在解决方案中。