I’ve been reading up on interactive voice response systems (IVRs), since they seem very similar in design to how Jibo’s SDK approaches it, and there’s been one item that always stands out: the tuning of the skill after it was launched. Most places ( e.g. https://msdn.microsoft.com/en-us/library/ee800143.aspx#BKMK_postlaunchtuning ) view it as a key aspect of the process.
Now, for example the Microsoft page assumes that the skill developer has access to the speech recordings and the logs, but am I incorrect in assuming that that is out of the question for Jibo? I can imagine a total privacy disaster if Jibo shared people’s recordings with anonymous skill developers.
So, I was wondering, how is that planned to happen? According to these sites it’s hard to guess how users will use a skill and what they say to it, but without seeing that data I can’t imagine how one would improve their own skill.
In the same token, I have no idea how Echo or Google Home go about it. They are obviously faced with the same conundrum.