The Talkbox layer allows the state-of-the-art neural networks (NLP, Vision, AGI) to run locally on your device (smartphone, tablet, watch). This means ultra-low latency and 100% uptime.
GPU-based Cloud can become very expensive very quickly. Since all of our AI runs locally on the device, there are negligible on-going cloud costs. Feedback and related data could optionally be aggregated to push a refined model downstream every update cycle (such as a few weeks/months).
On-device means your data always stays on the device. This offers the maximum possibly privacy for the users. Also from a cybersecurity perspective, this offers the lowest attack-surface and hence more secure than when the neural networks run on the cloud. Optionally, some aggregate data could be transferred to the cloud to fine-tune the neural netowrks or for analytics. This is fully configurable and optional.