so I’ve been looking around and with the Mk II coming up I have a few questions about the current workings of and near future plans for Mycroft. Since I found no better place to ask I’ll just post them here.
Is the device capable of multi-language interaction and being addressed via multiple languages without re-configuration?
E.g. German and English?
How smart is the current STT logic?
Does it try to “understand” what I’m saying or does it require me to remember and say specific triggers for skills?
1). Can it tell the difference between stopping the playback of music and stopping to run skills when I say “stop”?
2). Can I query tasks like “listen to radio station example” or does it depend on a specific skill for each station?
Is Mycroft multi-user capable?
As in is it able to differentiate between users and act accordingly within Skills? For instance can it tell me apart from friends or family?
Since Mycroft devices are registered to an online-account and given names - can they be addressed? Examples:
1). Can I transmit a message from the bath to the kitchen? (Hey Mycroft, let me talk to the kitchen device)
2). Can I set an alarm on my bedroom device from the living room one? (Hey Mycroft, wake me up 7am on my bedroom device)
4). Can I set a timer in the kitchen so that it rings on all devices in the house once complete? (Hey Mycroft, set a timer for 60 minutes on all my devices)
Also related: how would you go about setting up a multi-room environment?
Are there ecological plans to interface multiple microphone arrays / speakers to one unit (what ivee tried to do) or do you have to buy as many devices as coverage is needed and keep them all sitting there using electricity?
Can the Mycroft unit control regular PCs or are there plans to do it?
E.g. can it be interfaced as a HID device forwarding keypresses (hotkeys)?
Have there been talks to have Mycroft be included to the lineup of hardware vendors for power sockets etc?
For example idevices etc.?
The last question is a bit “personal” I guess but I will ask it since the core issue is still there.
Have there been talks to the hello Ivee / Ivee voice team to gather lessons learned about what caused them to go belly up?
They gave in under the pressure from Amazon, Google and Apple as to how well they were able to integrate their devices with their target platforms. And it’s certainly a valid risk - in fact I’m seeing the same issue in my contact circles: people usually buy the AI that suits their target platform (Apple, Google or “just” heavy Amazon users (mostly PC guys). It will be no surprise to see this trend continue as Amazon aquired the Cryengine (calling it Lumberyard) and thus we will probably see the Echo being improved in the future to even act as an input device for PC gaming.
So are there plans to avoid the same errors that lead to the cancellation of Ivee voice?
Thanks in advance and best regards - really looking forward to the Mk. II Kickstarter