Within the realm of digital communications, digital voice assistants are gaining momentum, as they permit natural hands-free interactions, allowing us to communicate easily with voice-enabled devices. The beneficial strategy in a customer-centric functional scenario would be to deploy voice assistants and voice bots (computer programs that handle voice conversations and are driven by natural language processing and artificial intelligence. Examples of well-known voice assistants are Amazon’s Alexa, Google Assistant, Microsoft’s Cortana, Samsung’s Bixby, and Apple’s Siri.
For the above scenarios, from a developer’s perspective, certified development kits serve a very useful purpose. They can expeditiously and cost effectively set up solutions/applications and prototypes for smart gateways, smart lighting/plugs, IoT sound sensors, smart speakers, smart thermostats, and other devices that support voice services.
The ensuing discussion will specifically focus on the voice-enabled front-end audio system kit for Alexa-enabled products offered by Microchip Technology Inc. (through its Microsemi subsidiary), the Microchip AcuEdge ZLK38AVS2 Development Kit for Amazon AVS.
The development kit consists of:
The kit is equipped with a Timberwolf ZL38063 audio processor, powered by AcuEdge technology and Sensory’s TrulyHandsFree “Alexa” wake-word engine, for embedded and cloud based automatic speech recognition (ASR), and it supports audio enhancement features and functions such as:
Let us, now conduct the functional testing of the Microchip AcuEdge ZLK38AVS2 Development Kit for Amazon AVS. During the testing we will use the wake word, “Alexa,” as this kit is engineered for evaluation of voice-enabled front-end audio systems for Alexa-enabled products.
The test setup consisted of the front-end audio development kit, Microchip AcuEdge ZLK38AVS2 Development Kit for Amazon AVS and the following systems (please refer to Fig. B.):
We first assembled the kit with the ZLE38000-004 evaluation board Rev 401 (interfaces with the Raspberry Pi 3), the Raspberry Pi 3 and JBL Clip 2 portable speaker, using the provided installation accessories. We connected the speaker cord to the 3.5mm jack on the ZLE38000-004 board. We then deployed a pre-built Raspberry Pi image (the image has the Amazon Alexa sample application pre-configured to work with the Microchip ZLK38AVS2 kit) based on Raspbian Linux distribution, onto a 32GB SD card for our use with the kit. The image has both VNC and SSH enabled. We inserted the SD card and powered up the Raspberry Pi. In our setup, the Microchip PDS-408G/ PDS-208G digital ceiling PoE switches provided PoE, and network connectivity and Wi-Fi connectivity were provided by the Grandstream networks Wi-Fi access points GWN 7630/GWN7610/GWN7600.
We opened a browser, created and logged into our Amazon developer account. Within the Alexa Voice Service developer console, we clicked on products and created and registered a product by filling in the relevant fields such as: product name, product ID, product type (device with Alexa built-in), end user interaction (Hands-free to allow users to interact with Alexa by using voice at a close distance and Far-field toallow users to interact with Alexa by using their voice from a longer distance), and we created a security profile. Once the product creation and registration were complete, we took note of the product ID and client ID, as we would have to input those values and associate the Alexa sample application running on the kit with our Amazon developer account.
We used VNC to access the Raspberry Pi desktop using the required credentials. We then opened a terminal session on the Raspberry Pi, entered the commands “cd ~/ZLK38AVS” and “make avs_config” and accepted the ensuing license agreements. We were prompted for the product ID and client ID (it is located under “Other devices and platforms” for the product that we registered earlier on the Amazon developer site), and the Alexa sample app installation was completed.
We started the Alexa sample application by using the command “make start_alexa.” The first time the sample application started, it prompted us for a web-based authorization, providing a link and the associated code that needed to be entered to complete the authorization. The authorization and registration were completed, and the kit responded with Connecting …Authorized…Alexa is currently idle on the terminal session (refer to Fig. C.).
We were up and running in a speedy manner, the development kit was now listening (but we could mute it). It starts recording when the wake word “Alexa” is heard – no audio is stored or sent to the cloud unless the device detects the wake word. We had to simply say the wake word and begin our query. We said, “Alexa, what is the weather in Mississauga now?” The 12 RGB LED ring on the unit indicated the recording status. The recording was sent to the cloud for processing and storage, and the response was sent back to the kit using an SSL encrypted traffic stream (refer to Fig. D.).
It is also noteworthy that it offered us options such as Tap to talk and Hold to talk. On the terminal, for Tap to talk, we had to press ‘t' and the ENTER key followed by our query, “What is the time in Mississauga now?” (refer to Fig. E.). Next, for Hold to talk, we had to press ‘h’ followed by the ENTER key (simulates holding a button), followed by our query, “How far is Mississauga from New York?” (refer to Fig. F.). We then pressed ‘h’ followed by the ENTER key (simulates releasing a button) all without the wake word.
Functions such as the voice communication, audio barge-in during device playback, far-field 360-degree audio pick-up in the presence of interfering noise sources (including external HVAC noise), device playback, pre-loaded “Alexa” wake-word detection all worked satisfactorily. For us in our test setup, the kit represented an excellent smart speaker application.
The front-end audio development kit from Microchip – the Microchip AcuEdge ZLK38AVS2 Development Kit for Amazon AVS – was very easy to set up and test. It supports cloud-based and embedded ASR solutions. This development kit for AVS includes all the building blocks (software, voice processing technologies, chipsets, etc.) that leverage AVS APIs to help developers easily, quickly and cost effectively build and deploy smart speaker prototypes and applications and solutions that leverage Amazon AVS voice support.
For more information, education, and networking opportunities around how businesses are leveraging voice recognition, AI and automation to improve existing business models and develop new ones, don’t miss the Future of Work Expo in Fort Lauderdale, Florida February 12-14, 2020. The event, part of the TechSuperShow, will explore how AI, machine learning, and automation are being used throughout a host of vertical markets and are rapidly shaping the future of business.
Future of Work Contributor
COVID restrictions were originally meant to be temporary measures, so using video conferencing tools like Zoom or Teams to keep things moving for a fe…
While artificial intelligence (AI) is helping many businesses get significant value from their data and enabling better fraud prevention and improved …
NICE has announced five new technology partners for its Evidencentral Marketplace. The open evidence management ecosystem is designed to accelerate th…
Although COVID-19 has transformed the workforce virtually overnight, there is no clear consensus on what the workplace will look like moving forward. …
The BYOD trend is steadily evolving to a bring your own environment (BYOE) work culture as a result of the COVID-19 pandemic. Businesses are rethinkin…