[asterisk-app-dev] Asterisk and UniMRCP Licensing

Ben Klang bklang at mojolingo.com
Fri Sep 5 09:47:35 CDT 2014


On Sep 5, 2014, at 10:33 AM, Matthew Jordan <mjordan at digium.com> wrote:

> 
> 
> 
> On Fri, Aug 29, 2014 at 2:52 AM, Ben Merrills <b.merrills at mersontech.co.uk> wrote:
> Great news!
> 
> Will this go any way towards helping get some general support for UniMRCP via ARI? I know when I first brought this up a few months ago, my main motivation was to try and see if there was a path for getting speech functionality exposed into ARI.
> 
> 
> So, in my spare time, I have been playing around with this.
> 
> Putting an ARI wrapper around the res_speech API is pretty easy. I haven't really decided on what the resources would look like, but right now my working theory would be that it would look similar to the playbacks resource - that is, you can start speech detection via some operation on the channels resource, which hands you back a handle to a speech detection control resource. You can then use a speech detection resource to manipulate what is happening.
> 
> That's the theory anyway.
> 
> I've taken a glance at the Asterisk modules that were written by various folks that the UniMRCP project distributes. Unfortunately, it doesn't look like they made use of the res_speech API. As such, they can't just be used directly behind ARI - some substantial rework would have to be done.
> 
> The first step would still be to get ARI able to wrap res_speech. Once that's in place, if Arsen and others are interested, we can discuss how to tweak the UniMRCP Asterisk modules such that they fit into the overall speech architecture in Asterisk.
> 

Is it really required to use res_speech? If so, can we change the interfaces that ARI presents?

Over the last few years we’ve evaluated res_speech vs. the various UniMRCP applications (SynthAndRecog primarily). We’ve always come to the conclusion that the res_speech API either couldn’t give us what we needed, or was not as performant.  SynthAndRecog isn’t perfect, but it does a couple of crucial things, perhaps most importantly is the combined lifecycle of TTS + ASR so that you can “barge” into a TTS playback before it is finished.

Would now be the appropriate time to talk about the ideal application interface for speech in Asterisk?  Is it safe to assume that ARI’s speech APIs can break with the past (both res_speech and unimrcp-asterisk)?

/BAK/

> Matt
> 
> -- 
> Matthew Jordan
> Digium, Inc. | Engineering Manager
> 445 Jan Davis Drive NW - Huntsville, AL 35806 - USA
> Check us out at: http://digium.com & http://asterisk.org
> _______________________________________________
> asterisk-app-dev mailing list
> asterisk-app-dev at lists.digium.com
> http://lists.digium.com/cgi-bin/mailman/listinfo/asterisk-app-dev

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.digium.com/pipermail/asterisk-app-dev/attachments/20140905/5fe18b81/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 841 bytes
Desc: Message signed with OpenPGP using GPGMail
URL: <http://lists.digium.com/pipermail/asterisk-app-dev/attachments/20140905/5fe18b81/attachment-0001.pgp>


More information about the asterisk-app-dev mailing list