Re: Asterisk and UniMRCP Licensing

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]


On Sep 5, 2014, at 10:33 AM, Matthew Jordan <mjordan@xxxxxxxxxx> wrote:

On Fri, Aug 29, 2014 at 2:52 AM, Ben Merrills <b.merrills@xxxxxxxxxxxxxxxx> wrote:
Great news!

Will this go any way towards helping get some general support for UniMRCP via ARI? I know when I first brought this up a few months ago, my main motivation was to try and see if there was a path for getting speech functionality exposed into ARI.

So, in my spare time, I have been playing around with this.

Putting an ARI wrapper around the res_speech API is pretty easy. I haven't really decided on what the resources would look like, but right now my working theory would be that it would look similar to the playbacks resource - that is, you can start speech detection via some operation on the channels resource, which hands you back a handle to a speech detection control resource. You can then use a speech detection resource to manipulate what is happening.

That's the theory anyway.

I've taken a glance at the Asterisk modules that were written by various folks that the UniMRCP project distributes. Unfortunately, it doesn't look like they made use of the res_speech API. As such, they can't just be used directly behind ARI - some substantial rework would have to be done.

The first step would still be to get ARI able to wrap res_speech. Once that's in place, if Arsen and others are interested, we can discuss how to tweak the UniMRCP Asterisk modules such that they fit into the overall speech architecture in Asterisk.

Is it really required to use res_speech? If so, can we change the interfaces that ARI presents?

Over the last few years we’ve evaluated res_speech vs. the various UniMRCP applications (SynthAndRecog primarily). We’ve always come to the conclusion that the res_speech API either couldn’t give us what we needed, or was not as performant.  SynthAndRecog isn’t perfect, but it does a couple of crucial things, perhaps most importantly is the combined lifecycle of TTS + ASR so that you can “barge” into a TTS playback before it is finished.

Would now be the appropriate time to talk about the ideal application interface for speech in Asterisk?  Is it safe to assume that ARI’s speech APIs can break with the past (both res_speech and unimrcp-asterisk)?



Matthew Jordan
Digium, Inc. | Engineering Manager
445 Jan Davis Drive NW - Huntsville, AL 35806 - USA
asterisk-app-dev mailing list

Attachment: signature.asc
Description: Message signed with OpenPGP using GPGMail

asterisk-app-dev mailing list

[Index of Archives]     [Asterisk SS7]     [Asterisk Announcements]     [Asterisk Users]     [PJ SIP]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [ECOS]     [Linux API]

  Powered by Linux