How does UNIHIKER use Azure speech recognition and speech synthesis

0 10572 Medium

UNIHIKER is a single-board computer(SBC) that features a 2.8-inch touchscreen, Wi-Fi and Bluetooth. It is equipped with light sensor, accelerometer, gyroscope and microphone. With built-in co-processor, it is able to communicate with various analog/digital/I2C/UART/SPI sensors and actuators.


Through official information, we learned that UNIHIKER has speech recognition, speech synthesis, wifi and other functions. Therefore, it can be used for speech recognition and speech synthesis projects.


I have shared the Azure method I often use to make UNIHIKER an intelligent conversation assistant. 


Official website link: https://azure.microsoft.com/en-us

Step 1: Account Login

Login with GitHub account

Login with email address

Login with phone or Skype

Click Next to complete the login

Step 2 Create resources

Step 3 Set parameters

Basics create

Subscription-> Default selection

Resource group-> Create a new resource group

Regin-> Select according to your region

Create a new project name

Pricing tier-> Depending on your choice of pricing

Then select Next.

Network selection

Then select Next.

Identity selection

Open or close according to your needs

Then select Next.

Tags choice

Default empty

Then select Next

Review + create

Review + Create

Then select Create.

Creation complete

Click on GO to resource

Get the key and Location/Region

Click the arrow to get the key and Location/Region

The Unihiker uses Azure services



Before installation, ensure that UNIHIKER is connected to wifi. For details, see the official documentation of UNIHIKER

pip install azure.cognitiveservices.speech
import time
import os
from azure.cognitiveservices.speech import  SpeechConfig
AZURE_SPEECH_KEY = "" # Fill key 
AZURE_SPEECH_REGION = "" # Enter Location/Region 
    import azure.cognitiveservices.speech as speechsdk 
except ImportError:
    Importing the Speech SDK for Python failed.
    Refer to
    https://docs.microsoft.com/azure/cognitive-services/speech-service/quickstart-python for
    installation instructions.
    import sys
# Set up the subscription info for the Speech Service:
# Replace with your own subscription key and service region (e.g., "japaneast").
speech_key, service_region = AZURE_SPEECH_KEY, AZURE_SPEECH_REGION
def tts(text):
    speech_config = SpeechConfig(subscription=speech_key, region=service_region)
    speech_config.speech_synthesis_language = "en-US" 
    speech_config.speech_synthesis_voice_name ="en-US-JennyNeural"
    # Creates a speech synthesizer for the specified language,
    # using the default speaker as audio output.
    speech_synthesizer = speechsdk.SpeechSynthesizer(speech_config=speech_config)
    # Receives a text from console input and synthesizes it to speaker.
    result = speech_synthesizer.speak_text_async(text).get()
def speech_recognize_once_from_mic():
    speech_config = speechsdk.SpeechConfig(subscription=speech_key, region=service_region, speech_recognition_language="en-US")
    audio_config = speechsdk.audio.AudioConfig(use_default_microphone=True)
    speech_recognizer = speechsdk.SpeechRecognizer(speech_config=speech_config, audio_config=audio_config)
    print("Speak into your microphone.")
    speech_recognition_result = speech_recognizer.recognize_once_async().get()
    if speech_recognition_result.reason == speechsdk.ResultReason.RecognizedSpeech:
        print("Recognized: {}".format(speech_recognition_result.text))
    elif speech_recognition_result.reason == speechsdk.ResultReason.NoMatch:
        print("No speech could be recognized: {}".format(speech_recognition_result.no_match_details))
    elif speech_recognition_result.reason == speechsdk.ResultReason.Canceled:
        cancellation_details = speech_recognition_result.cancellation_details
        print("Speech Recognition canceled: {}".format(cancellation_details.reason))
        if cancellation_details.reason == speechsdk.CancellationReason.Error:
            print("Error details: {}".format(cancellation_details.error_details))
            print("Did you set the speech resource key and region values?")
while True:
All Rights