Professional Documents
Culture Documents
Programming Voice Controlled Iot Applications With Alexa and Raspberry Pi John Allwork All Chapter
Programming Voice Controlled Iot Applications With Alexa and Raspberry Pi John Allwork All Chapter
Programming Voice Controlled Iot Applications With Alexa and Raspberry Pi John Allwork All Chapter
Programming
# constants
and gained his MSc at UMIST.
”
T%H:%M:%S.00Z
skill and terminology such as utterances, intents, slots, and conversations After two years working for ICL as
m - % d
are explained. Debugging your code, saving user data between sessions, a design engineer, he returned to
UTC_FORMAT = “ % Y - %
a z o n . c o m / a u t h/o2/token”
m
https://api.a
S3 data storage and Dynamo DB database are discussed. UMIST where he graduated with a
PhD in ‘Design and Development of
Microprocessor Systems’.
TOKEN_URI = “
In-skill purchasing, enabling users to buy items for your skill as well as
s constants
certification and publication is outlined. Creating skills using AWS Lambda
# Token acces
He worked for several years in
and ASK CLI is covered, along with the Visual Studio code editor and technical support and as manager
n t i a l s . k e y [ ‘ CLIENT_ID’]
rede T’]
CLIENT_ID = c ‘CLIENT_SECRE
local debugging. Also covered is the process of designing skills for visual in electronics distribution, working
closely with Intel engineers and
e n t i a l s . k e y [
= cred
displays and interactive touch designs using Alexa Presentation Language.
CLIENT_SECRET
later designing Inmos Transputer
systems.
The second half of the book starts by creating a Raspberry Pi IoT “thing”
to control a robot from your Alexa device. This covers security issues and Having taught electronics at
s_to ken():
methods of sending and receiving MQTT messages between an Alexa Manchester Metropolitan University,
he retired in 2011 but retained
def get_acces
device and the Raspberry Pi.
his interest in electronics and
= {
token_params t_credentials
”,
programming. His other occupations
Creating a smart home device is described including forming a security consist of traveling, walking,
p e ” : “ c l i e n
profile, linking with Amazon, and writing a Lambda function that gets geocaching and spending time on “grant_ty a c t i ve_events”,
e x a : : p r o
“scope”: “al
triggered by an Alexa skill. Device discovery and on/off control is his allotment.
LIENT_ID,
demonstrated.
l i e n t _ i d ” : C
“c
Next, readers discover how to control a smart home Raspberry Pi display
n t _ s e c r e t ” : CLIENT_SECRET
from an Alexa skill using Simple Queue Service (SQS) messaging to “clie
switch the display on and off or change the color.
A node-RED design is discussed from the basic user interface right up to
}
configuring MQTT nodes. MQTT messages sent from a user are displayed Elektor International Media
on a Raspberry Pi. www.elektor.com
token_headers
= {
o n / j s o n ; c h a r set=UTF-8”
”: “applicati
A chapter discusses sending a proactive notification such as a weather “Content-Type
alert from a Raspberry Pi to an Alexa device. The book concludes by
}
explaining how to create Raspberry Pi as a stand-alone Alexa device.
I , h e a d e r s = t oken_headers
s
Dr John Allwork
ts.post(TOKEN
_UR
n s e = r e q u e
respo
● All rights reserved. No part of this book may be reproduced in any material form, including photocopying, or
storing in any medium by electronic means and whether or not transiently or incidentally to some other use of this
publication, without the written permission of the copyright holder except in accordance with the provisions of the
Copyright Designs and Patents Act 1988 or under the terms of a licence issued by the Copyright Licencing Agency
Ltd., 90 Tottenham Court Road, London, England W1P 9HE. Applications for the copyright holder's permission to
reproduce any part of the publication should be addressed to the publishers.
● Declaration
The Author and Publisher have used their best efforts in ensuring the correctness of the information contained in
this book. They do not assume, and hereby disclaim, any liability to any party for any loss or damage caused by
errors or omissions in this book, whether such errors or omissions result from negligence, accident, or any other
cause.
All the programs given in the book are Copyright of the Author and Elektor International Media. These programs
may only be used for educational purposes. Written permission from the Author or Elektor must be obtained before
any of these programs can be used for commercial purposes.
● F irst edition
© Copyright 2023: Elektor International Media B.V.
Editor: Alina Neacsu
Prepress Production: Jack Jamar | Graphic Design, Maastricht
Elektor is part of EIM, the world's leading source of essential technical information and electronics products for pro
engineers, electronics designers, and the companies seeking to engage them. Each day, our international team develops
and delivers high-quality content - via a variety of media channels (including magazines, video, digital media, and social
media) in several languages - relating to electronics design and DIY electronics. www.elektormagazine.com
●4
Contents
Contents
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
●5
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
●6
Contents
4.14 AMAZON.SearchQuery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
4.15 ASK SDK Utilities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
4.16 Intent error logging . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82
4.17 Language understanding NLU and Automatic speech recognition ASR . . . . . . . . 82
4.18 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82
●7
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
●8
Contents
●9
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
● 10
Contents
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 273
● 11
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
Dr. John Allwork was born in 1950 in Kent, England and became interested in electronics
and engineering at school. He went to Sheffield University on their BEng Electrical and
Electronic Engineering course. There he developed an interest in computers and continu-
ed his education on an MSc course in Digital Electronics and Communication at UMIST.
After two years working for ICL as a design, commissioning and test Engineer he returned
to UMIST where he graduated with a Ph.D. in ‘Design and Development of Microprocessor
Systems’.
He worked for several years in technical support and as a manager in electronics distribu-
tion, working closely with Intel Application Engineers and followed this with design work
using the Inmos Transputer systems.
Having taught at Manchester Metropolitan University he retired in 2011 but has kept up
his interest in electronics and programming as well as his other occupation of travelling,
walking, geocaching and spending time on his allotment.
● 12
Introduction
This book is aimed at anyone who wants to learn about programming for Alexa devices
and extending that to Smart Home devices and controlling hardware, in particular the
Raspberry Pi.
It covers Alexa programming concepts from the basic concepts of Alexa Voice service, the
interaction model and the skill code which runs on AWS (Amazon Web Services) Lambda.
It takes the reader through all stages of creating skills to certification and publishing,
including writing skills that involve in-skill purchasing. It discusses different ways of crea-
ting skills, then moves on to creating visual skills using APL (Alexa Presentation Langua-
ge) for screen-based Alexa devices.
The book then moves on to cover different ways of controlling hardware including the
Internet of Things and Smart Home devices. There are interfaces with the Raspberry Pi
using MQTT and SQS communication, displaying on the Pi using Node-RED and Python
code.
Although mostly based on Python, Node.js examples or links are also provided. The full
code is provided in a separate document.
Please note that Alexa skill development, the developer console and APL versions have
changed since writing this book, so please bear with the author if there are slight diffe-
rences.
I do not pretend to know all there is about Alexa and Raspberry Pi programming – they
seem to advance faster than I can follow! I have a background in hardware and software
design. I am sure that there are areas where some programmers may be offended by my
code and that there may be better ways to write it, but I have written and tried all the
examples and know they work. I hope the examples will spur you on to find solutions to
your own problems. Should you need more information then please try the online help
and the Raspberry Pi or Alexa forums: alexa.design/slack is particularly good. There are
plenty of programmers out there willing to help solve your problems, often extremely
quickly; certainly faster than I would get back to you!
I wish to thank my friends for encouraging me, especially Dr. Hugh Frost, Andy Marsh
and Dr. John Nichols; the Alexa staff: in particular Jeff Nunn, Jeff Blankenburg and Ryan J
Lowe; helpers on the alexa.design/slack group, including Andy Whitworth; subscribers of
my YouTube and GitHub channels who have made encouraging comments; and the many
anonymous people on the internet, forums, blogs and websites who have answered many
questions, not just my own – keep up the good work. Not least of all I would like to thank
my wife Penny, for supporting me throughout.
● 13
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
In 2017, Amazon released a combination of the Dot and Show, called the Echo Spot. In
the same year, the Echo Show was released and featured a slanted, 7-inch touchscreen,
camera and speaker. This later changed to a 10-inch screen (Echo Show 10), and more
recently, added a 360-rotating display.
The Echo Show 5 came in 2019, (2nd gen in 2021), as well as Echo Show 8 and an Echo
Show 15 in 2021 designed for wall mounting.
There are other devices too, including the Button, Flex, Input, Look and recently the Astro
robot.
Here are some of my devices (not including smart devices). From the top: Echo Show 8,
Fire TV stick, Echo Auto, my original Echo dot, and the Echo Spot.
● 14
Chapter 1 ● Alexa History and Devices
Even though many devices have a screen, you should always design for ‘voice first’.
The interaction model is what your users say and how they communicate with your skill.
AWS Lambda is a serverless, event-driven computing service that lets you run your code.
Lambda can be triggered by many AWS services and you only pay for what you use.
When a user interacts with an Echo device, AVS sends a request to the skill which is run-
ning on AWS Lambda. The skill replies with a response that is turned into a speech and/
or visual response back to the user.
1.2 Pricing
Although there are charges for AWS Lambda, the AWS Lambda free tier includes one
million free requests per month and 400,000 GB-seconds of compute time per month, as
well as 500 Mb storage. As you can see, this is more than enough for a beginner. For more
information, see https://aws.amazon.com/lambda/pricing/
For developers whose skills use more than this, Amazon provides Promotional Credits
which reward those who build cloud-hosted applications, software, or tools for sustaina-
bility-related work.
For FAQ see reference 1.
● 15
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
We’ll be concentrating on Custom skills. Blueprint pre-built skills are also available and
easy to develop but have reduced options for user experience. We’ll also study smart
home skills, of course.
1.5.2 Invocation
The ‘invocation’ is the phrase used to trigger your skill, e.g.: ‘Alexa, open Johns’ weather
skill’ or ‘Alexa, launch my cooking skill’.
1.5.3 Utterances
Utterances are the phrases that your user says to make a request. There can be many
ways to achieve the same result, e.g.: What’s the time? What’s the time now? What time
is it? – you will have to think of as many possible ways that your user can interact with
your skill. Nevertheless, Alexa will build your model and try to find similar utterances.
All the possible ways to do this can be difficult to describe (considering, for instance, how
many different ways and types of pizza someone might order), so Amazon has recently
announced Alexa Conversations to help with this. For more information, see reference 3.
● 16
Chapter 1 ● Alexa History and Devices
• Built-in Intents
- Standard built-in intents: These are provided by default by Amazon that every skill
must have, e.g.: AMAZON.StopIntent, AMAZON.CancelIntent, AMAZON.Fallback-
Intent, etc. and include: AMAZON.YesIntent, and AMAZON.NoIntent, intents for
screen control (e.g., scroll up/ down/ left / right) and media intents (pause, repeat,
resume), and also an AMAZON.SendToPhoneIntent. You can see these when you
add an intent, and select “Use an existing intent from Alexa’s built-in library”
- The Alexa Skills Kit also provides a library of specific built-in intents and includes
intents such as Actor intents, Books, Calendar, LocalBusiness, Music, TV, Series,
WeatherForecast, etc.
These intend to add functionality to your skill without you having to provide any sample
utterances. For example, the WeatherForecast includes a search action (What is), an ob-
ject (WeatherForecast), location (London) and date (tomorrow) .
We won’t cover them in this book, see:
https://developer.amazon.com/en-US/docs/alexa/custom-skills/built-in-intent-library.html
• Custom Intents
• A Launch request that runs when our skill is invoked (as a result of the user saying,
‘Alexa open …’ or ‘Alexa, launch ...’).
• An Intent request which contains the intent name and variables passed as slot values.
• A SessionEnded request, which occurs when the user exits the skill, or there is an
unmatched user’s response (although you may be able to trap this out with AMAZON.
FallbackIntent).
This information is all packaged and sent as a request (and returned as a response) as a
JSON file. We’ll look at the JSON code later.
1.5.5 Slots
A slot is a variable that contains information that is passed to an intent. The user might
say ‘What’s the time in London’. Here ‘London’ (or Paris or Rome) is passed as a slot
variable to the intent code.
● 17
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
Amazon provides built-in slot types, such as numbers, dates and times, as well as built-
in list types such as actors, colors, first names, etc. In the previous example, we could use
AMAZON.GB_CITY which provides recognition of over 15,000 UK and world-wide cities
used by UK speakers.
However, some of these slots are being deprecated (including AMAZON.GB_CITY in favour
of AMAZON.CITY), so check. The full list is covered at ‘List Slot Types:
https://developer.amazon.com/en-US/docs/alexa/custom-skills/slot-type-reference.html#list-slot-types
Developers can create custom slots for variables that are specific to their skill.
When we define our utterances, slots are shown in curly braces: {city}, e.g.:
Example:
Intent GetTimeIntent
Utterance What is the time in
Slot {city}
The GetTimeIntent will trigger a function in your skill (which you might sensibly call Get-
TimeIntentFunction).
Slots are used to pass data from your VUI (voice user interface) to your program. As an
example, we might have an Alexa skill that asks for your name.
● 18
Chapter 1 ● Alexa History and Devices
At this point, the slot data is lost unless you save it. You can save data in a temporary
folder but more often data is stored in session attributes, you will find out later.
1.5.7 Endpoints
The endpoint is where your code is hosted. You can choose an Amazon-hosted, AWS
Lambda ARN (Amazon Resource Name) site or host it yourself on an HTTPS site that you
manage.
If you choose an AWS site, it will give you an ID beginning arn:aws:lambda and look
like: arn:aws:lambda:<region>:function:<functionID>. You skill also has an ID looking
something like this: amzn1.ask.skill.a0093469-4a50-4428-82e6-abcde990fgh3.
1.5.8 Regions
If using an AWS-hosted site, you should host your code in a region near to your user.
We’ll see that there are lots of regions, but for some skills currently only North Virginia
is available.
If the shouldEndSession parameter is ‘true’ the skill terminates, otherwise, the session
remains open and expects the user to respond. If no user input occurs, a reprompt is
sent if included in the code. If the user still doesn’t respond (after about 8 seconds), the
session ends 4.
Skill connections and progressive responses may override these rules. For example, if
a skill has to get further information from another source, e.g., when your taxi will be
available, or pizza delivered.
● 19
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
The basic JSON request contains information on the active session, the context, the sys-
tem information on the application (ID), user, deviceID, and the request itself:
{
“version”: “1.0”,
“session”: {
( ..session parameters)
}
“context”: {
( ..information on the Alexa device)
}
“System”: {
…
“request”: {
“type”: “IntentRequest”,
“requestId”: “amzn1.echo-api.request.745d…9a”,
“locale”: “en-GB”,
“timestamp”: “2022-04-14T09:27:01Z”,
“intent”: {
“name”: “HelloWorldIntent”,
“confirmationStatus”: “NONE”
}
}
The reply JSON contains is the response speech and reprompt, as well as the state
of the EndSession and session attributes.
{
“body”: {
“version”: “1.0”,
“response”: {
“outputSpeech”: {
“type”: “SSML”,
“ssml”: “<speak>Welcome, you can say Hello or Help.</speak>”
},
“reprompt”: {
“outputSpeech”: {
“type”: “SSML”,
“ssml”: “<speak>Welcome, you can say Hello or Help.</speak>”
}
},
● 20
Chapter 1 ● Alexa History and Devices
“shouldEndSession”: false,
“type”: “_DEFAULT_RESPONSE”
},
“sessionAttributes”: {},
“userAgent”: “ask-python/1.11.0 Python/3.7.12”
}
}
The response can get more complicated if there is a dialog session occurring (i.e., if the
program hasn’t enough information to complete the intent request and has to ask for
more).
We’ll look at the information passed in the JSON request and response and how to extract
it in a later chapter.
The current categories are: At home, Kids recommended, Learning and knowledge, Fun
and Games, Storyteller, Greetings and Occasions, Communities and Organizations, and
Business 6. We won’t cover them here.
● 21
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
1.10 Summary
We’ve seen how the Alexa devices have developed from the original voice-only device to
screen-based and robot devices, how the Alexa Voice service works and looked at termin-
ology – Invocation, Utterances, Intents and Slots. Finally, we looked at a skill session and
how data is passed and saved during a session and between sessions.
In the next chapter, we’ll see how to set up an Alexa account before moving on to our
first Alexa skill.
1.11 References:
1. https://aws.amazon.com/lambda/faqs/
2. https://developer.amazon.com/en-US/docs/alexa/ask-overviews/list-of-skills.html
3. https://www.youtube.com/watch?v=1nYfRvg976E
4. https://developer.amazon.com/en-US/docs/alexa/custom-skills/
manage-skill-session-and-session-attributes.html
5. https://www.json.org/json-en.html
6. https://blueprints.amazon.com
● 22
Chapter 2 ● Creating your Amazon Account
2.1 Introduction
There are three ways to create an Alexa Skill.
1. Using Alexa-Hosted Skill – this is most probably the easiest way. You don’t need
an AWS account – you can just get started.
2. Using AWS Lambda and the Developer Console.
3. Using an IDE (such as Visual Studio code) and ASK CLI (Alexa Skills Kit and Com-
mand Line Interpreter).
We’ll use an Alexa-Hosted skill where we can. It provides most of the requirements we
need including extra storage (S3) and a database (DynamoDB). We’ll study these in a
later chapter.
● 23
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
Complete the form and create your account. You may have to authenticate it with an OTP.
You can now provide payment information, user roles, etc., or leave until later. If you
want to earn money from In Skill Purchases, you will have to complete this.
● 24
Chapter 2 ● Creating your Amazon Account
We will be using the Alexa Skills Kit but note the other items here.
• Alexa Voice Service (AVS) is used to integrate Alexa features and functions into a
product (see chapter 17).
• Amazon Appstore is an app store for Android and provides apps for Fire Tablets.
• Amazon Dash console enables you to manage Amazon dash buttons.
• Amazon GameOn is for mobile gamers to share their experiences (high scores, tips,
tricks, etc.) and to participate in challenges.
• Login with Amazon (LWA) is used to authenticate users using their Amazon accounts.
You need this if you create any smart home devices.
● 25
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
2.4 Hosting
Click on the Hosting tab to see what you’re allowed to access.
You’ll find the storage and data transfer more than enough for now.
● 26
Chapter 2 ● Creating your Amazon Account
2.5 Summary
In this chapter, we saw the different ways of creating and Alexa skill, created our develo-
per account, and saw how to start an Alexa-Hosted skill.
● 27
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
3.1 Introduction
In the last chapter, we saw how to start an Alexa-Hosted Skill. This provides access to
three AWS Lambda endpoints, an S3 bucket for storage and access to a DynamoDB data-
base. We’ll discuss these in later chapters.
In the next screen, you’re asked for a skill name, the locale (language and location) for
your skill, and if you want more than one locale, to sync them. This will link all the utte-
rances. The user interface has changed slightly since originally writing this chapter, but
the process is the same.
● 28
Chapter 3 ● Creating an Alexa Skill
There are rules as to what names you can give your skill. It must be at least two charac-
ters and less than fifty, and brand names are only allowed if you show proof of rights.
When you publish your skill, these will be checked.
It’s sensible to give your skill a two- or three-word name. Try ‘<your name> first skill’.
I’ve used John’s first demo skill.
● 29
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
In the top right-hand corner, you can select the hosting region. It’s sensible to select one
nearest to you.
Click ‘Create skill’. The next screen provides some templates. There are more Node.js
templates than Python. Some provide examples of intents, data persistence and perso-
nalization. Choose ‘Start from scratch’ and ‘Continue with Template’.
● 30
Chapter 3 ● Creating an Alexa Skill
Your Alexa skill will be created. It takes about a minute and then you’re taken to the de-
veloper console build tab.
The first thing you notice is the video showing you how to build an Alexa Skill. You can
view this now if you want to.
At the top left is your skill name. To the left of this, you can return to your skill list. To the
right, you can move between the build tab (currently displayed), the Code and Test tabs
which we’ll look at soon and the Distribution, Certification and Analytics tabs which are
used when publishing the skill.
● 31
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
This template skill can now be tested and executed – you don’t have to do anything. But
before we run our skill, let’s take a look at the invocation and intents. Remember, the
invocation is what your user says to invoke (start) your skill, for instance: “Alexa, open
John’s first demo skill”; the intents are what your users say when the skill is running du-
ring a conversation e.g., “My favorite color is red”.
On the right-hand side, there is a skill builder checklist that shows that four stages have
been completed.
We won’t monetize our skill just yet – we will cover that in a later chapter. These actions
are also available in the left-hand column, which also shows the currently selected lan-
guage. The items labelled ‘Custom’ can be expanded for more detail.
We’ll look at the Invocation Name and Build Model in this chapter, and the Intents and
Endpoint in the next.
● 32
Chapter 3 ● Creating an Alexa Skill
Sometimes (as here) the Invocation Name says, “change me”. If necessary, change this
to your skill name (it can be anything, but it seems sensible). I’ve used john’s first demo
skill. The user will say “Alexa, open john’s first demo skill” to invoke the skill. Note that
the Invocation Name can only be lowercase.
As you’ve changed the invocation, you will have to rebuild your model. Save the model
and rebuild it. Use the option we saw above (3. Build Model), or these buttons.
Whilst it’s doing that (It won’t take long), take a look at the Skill Launch Phrases and
Intent Launch Phrases.
Skill and Intent launch phrases enable your users to interact with your skill without sta-
ting the launch name. Neither is available until the skill is published.
● 33
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
If you click on an intent, you get a list of the utterances that will trigger the intent.
The Python code example here begins by importing the libraries we need. For example,
import logging imports the libraries that enable us to output debugging information to
Cloudwatch logs, see section 3.8. To do this we use the code
logger.info(info_to_display)
● 34
Chapter 3 ● Creating an Alexa Skill
The handlers follow. These deal with the launch request and intents such as Hello or Help.
class LaunchRequestHandler(AbstractRequestHandler):
class HelpIntentHandler(AbstractRequestHandler):
The rest of the intents follow. These include the Stop or Cancel intent (what to do when
the user says “stop” or “cancel”); the Fallback intent (what to do if the user’s utterance
doesn’t match any voice model intent); the SessionEnded intent (what to do when the
session ends), on so on.
The intent handlers are registered using the sb.add_request_handler code, where sb is
the SkillBuilder() code. This is imported by the command:
● 35
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
This example uses Classes, you can also use Decorators. For a more detailed description
of both, and the code, see:
https://developer.amazon.com/en-US/docs/alexa/alexa-skills-kit-sdk-for-python/develop-your-first-skill
html#using-handler-classes
It’s worth noting the link mentioned in this code to more skill examples at:
https://alexa.design/cookbook
For the Node.js version, refer to section 3.9 at the end of this chapter.
Before looking at our code in more detail, let’s make the welcome message more perso-
nal, then test our skill.
● 36
Chapter 3 ● Creating an Alexa Skill
Python:
• If you are unfamiliar with Python, you MUST ensure your code is tabbed correctly.
Note that this is in the LaunchRequestHandler code, which runs when the user starts your
skill.
• Save the code. Click the deploy button in the top right-hand corner.
Node.js:
• Make sure that the quotes match and the line ends with a semi-colon.
Remember that, after making any changes to your code and before we can test it, we
must save and deploy the code.
● 37
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
You’ll get a message when it’s been deployed, e.g., Last Deployed: Apr 13, 2022, 8:26 PM.
You can now send any command that you might usually use with your Alexa device (try:
Alexa, what’s the weather).
Type (or click the microphone button and say) “Open <launch name>” e.g., “Open johns
first demo skill”. You don’t have to use the wake word (i.e., No need to say or type Alexa,
open johns first demo skill).
● 38
Chapter 3 ● Creating an Alexa Skill
Troubleshooting
If it doesn’t work, check that you have the correct invocation name, that you saved and
built the model and that you saved and deployed your code. Also, if the box under Alexa
Simulator e.g., English (GB) is greyed out, close the developer console and log in again.
In the simulator, you can see we see the invocation: “open johns first demo skill”
The response: “Welcome, you can say …” (Or whatever you changed it to)
and after saying ‘hello’, our HelloWorld intent code response: “Hello”.
The skill terminates after this conversation – we’ll see why later. If you type in Hello
again, your program doesn’t respond - Alexa responds instead.
You can save up to 30 utterances for later use by clicking the plus sign:
Note that, depending on how the user invokes your skill, it is possible that the launch
request isn’t executed.
The Device Display is used to show what the device will display if it has a screen. If you
scroll down, you’ll see that it currently displays the skill name and how to exit. We’ll see
this in action when we discuss Alexa Presentation Language in a later chapter.
● 39
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
The device log provides details of the actions of the skill, including events and directives:
Personalization. Alexa devices are capable of identifying a recognized skill user if it is set
up and agreed by the user. You can use this in your code and test using personalization.
For more details, see:
https://developer.amazon.com/en-US/docs/alexa/custom-skills/add-personalization-to-your-skill.html
Remember, communication uses JSON code requests and responses. This shows the JSON
input (request –sent to your code) and JSON output (response – received by your code).
JSON is easy to understand. If you look at the JSON output, you can see where the
<speak>Welcome, you can say … occurs.
We won’t list all the JSON input here – you can do that yourself, but I will look at a few
relevant items, and we’ll see more later on.
● 40
Chapter 3 ● Creating an Alexa Skill
It starts with the version, then the session items that identify the sessionId, the applica-
tion and the user’s userId.
{
“version”: “1.0”,
“session”: {
“new”: true,
“sessionId”: “amzn1.echo-api.session.27d.. e7cb”,
“application”: {
“applicationId”: “amzn1.ask.skill.4649.. 981”
},
“attributes”: {},
“user”: {
“userId”: “amzn1.ask.account.AFCVJZSYW…VY2BH6HI”
}
It follows with some information on the device under “context” > “Viewports”. In this
case, it refers to the Alexa simulator we’re using and tells us we have a rectangular screen
capable of showing APL, it can’t rotate and has a pixel height and width of 1280 × 800.
See if you can identify these in your display
“context”: {
“Viewports”: [
{
“type”: “APL”,
“id”: “main”,
“shape”: “RECTANGLE”,
“System”: {
“application”: {
“applicationId”: “amzn1.ask.skill.4649….981”
},
“user”: {
“userId”: “amzn1.ask.account.AFCVJZSY…H6HI”
},
“device”: {
“deviceId”: “amzn1.ask.device.AGSER…DU”,
“supportedInterfaces”: {}
● 41
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
Finally, information about the request itself is displayed. In this case, we see that it was
a launch request, and the locale was en-GB (I’m in the UK)
“request”: {
“type”: “LaunchRequest”,
“requestId”: “amzn1.echo-api.request.a7ed3cc2-4bba-9e82-3c79e5e4eb59”,
“locale”: “en-GB”,
“timestamp”: “2022-04-13T19:57:52Z”,
“shouldLinkResultBeReturned”: false
}
}
We’ll see how to extract some of this data in our code later.
If we compare the request with that after entering ‘hello’, we get this:
“request”: {
“type”: “IntentRequest”,
“requestId”: “amzn1.echo-api.request.745de8e2-7… 6fc9a”,
“locale”: “en-GB”,
“timestamp”: “2022-04-14T09:27:01Z”,
“intent”: {
“name”: “HelloWorldIntent”,
“confirmationStatus”: “NONE”
}
We can see that we have an IntentRequest and its name is “HelloWorldIntent”. We pick
this up in our code to reply.
This is the response to the open skill launch request. It’s short, so we’ll look at the whole
listing:
{
“body”: {
“version”: “1.0”,
“response”: {
“outputSpeech”: {
“type”: “SSML”,
“ssml”: “<speak>Welcome, you can say Hello or Help.</speak>”
},
“reprompt”: {
“outputSpeech”: {
● 42
Chapter 3 ● Creating an Alexa Skill
“type”: “SSML”,
“ssml”: “<speak>Welcome, you can say Hello or Help.</speak>”
}
},
“shouldEndSession”: false,
“type”: “_DEFAULT_RESPONSE”
},
“sessionAttributes”: {},
“userAgent”: “ask-python/1.11.0 Python/3.7.12”
}
}
The main parts here are the response > outputSpeech. This is SSML (Speech Synthesis
Markup Language) which tells the device what to say initially, and the reprompt > output-
Speech which tells the device what to say if the user doesn’t respond. In this case, because
“shouldEndSession” is false, the user then has 8 seconds to respond, before the device
gives up.
“outputSpeech”: {
“type”: “SSML”,
“ssml”: “<speak>This speech uses SSML</speak>”
}
There is a long list of SSML tags. They can also be used in APLA, see section 8.2. They
are listed at:
https://developer.amazon.com/en-US/docs/alexa/custom-skills/speech-synthesis-markup-language-
ssml-reference.html
● 43
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
<speak>Sorry, but
<voice name=”Brian”>I am a robot</voice>
</speak>
• speak Surrounds the text to be spoken with this tag.
<speak>This is what Alexa sounds like without any SSML</speak>
There are strict requirements and limitations for audio when used in SSML. The main ones
being: a valid MP3 file hosted on an https source, no longer than (a combined) total of
240 seconds, a bit rate of 48 kbps and sample rate of 22050 Hz, 24000 Hz, or 16000 Hz.
Alexa provides a library of sound effects that you can use. It has a wide selection of short
audio clips, from Air/Fire Extinguisher through Animal to Wood/Squeaks, see
https://developer.amazon.com/en-US/docs/alexa/custom-skills/ask-soundlibrary.html
Python:
Node.js:
● 44
Chapter 3 ● Creating an Alexa Skill
In Python, make sure they are tabbed correctly (at the same level as the speak_output).
The Node.js changes are in sections 3.9 and 3.10 and use console.log() command.
You can see how your code is steered to the correct procedure from the JSON message.
The launch request had a “type”: “LaunchRequest”, which is picked up by the code:
return ask_utils.is_request_type(“LaunchRequest”)(handler_input)
● 45
Programming Voice-controlled IoT Applications with Alexa and Raspberry Pi
We’ll look at Intents and handler_input in more detail in the next chapter.
Save and deploy your code.
requirements.txt
This imports the libraries we need:
boto3==1.9.216
ask-sdk-core==1.11.0
We’ll see in a later chapter that boto3 is used to create, configure, and manage AWS ser-
vices, such as the Amazon Simple Storage Service (Amazon S3). On a side note, you can
search for “Why is it called Boto3?”.
ask-sdk-core provides the ‘boiler plate’ code for developing our Alexa Skills.
We’ll add to requirements.txt in later chapters.
utils.py
The create_presigned_url code here provides temporary access to S3 storage (‘buckets’)
to users who do not have AWS credentials or permissions. We’ll use that later too.
● 46
Chapter 3 ● Creating an Alexa Skill
3.8 Debugging
When you’ve tested your code, Click the Code tab and then the CloudWatch logs tab.
You may have to use the drop-down box to select your region, depending on where you are.
You may also be prompted to complete your settings (language and region) for the Cloud-
Watch logs to work.
● 47
Another random document with
no related content on Scribd:
O bigode fanado posto ao ferro
Está alli num desterro,
E cada pello em solidão tão rara,
Que parece ermitão da tua cara;
De cabelleira tal affirmam cegos
Que a mandaste comprar no Arco dos pregos.
Do confessor jesuita,
Que ao ladrão do confessado
Não só absolve o peccado,
Mas os fructos lhe alcovita:
Do precursor da visita,
Que na vanguarda marchando,
Vai pedindo e vai tirando,
O demo ha de ser algoz:
Porém fique aqui entre nós.
O ladronaço em rigor
Não tem para que dizer
Furtos, que antes de os fazer
Já os sabe o confessor:
Cala-os, para ouvir melhor,
Pois, com officio alternado,
Confessor e confessado
Alli se barbeam sós:
Porém fique aqui entre nós.
Aqui o ladrão se consente
Sem castigo e com escusa,
Porque do mesmo se accusa
O confessor delinquente:
Ambos alternadamente,
Um a outro e outro a um,
O peccado, que é commum,
Confessa em commua voz:
Porém fique aqui entre nós.
Tractam de diminuir
O dinheiro a meu pezar,
Que para a cousa baixar
O melhor meio é subir:
Quem via tão alto ir,
Como eu vi ir a moeda,
Lhe prognosticou a queda,
Como eu lh’a prognostiquei:
Dizem que o mandou El-Rei,
Quer creais, quer não creiais,
Não vos espanteis, que inda lá vem mais.
Mandam a força do fado,
Por ser justo que o dinheiro
Baixe a seu valor primeiro
Depois de tão levantado:
O que se vir sublimado
Por ter mais quatro mangabas,
Hão de peza-lo ás oitavas,
E por leve hão de engeita-lo:
E si com todo este abalo
Por descontentes vos dais,
Não vos espanteis, que inda lá vem mais.
A bocca desempedrada
É a ponte de Coimbra,
Onde não entram nem sahem
Mais que mentiras.
Ouçam e olhem,
Venham, venham, verão
O Frizão da Bahia,
Que está retratado
Ás maravi, maravi, maravilhas.
Não é a lingua de vacca
Pelo maldizente e maldicta,
Mas pelo muito que corta
De tiririca.
Ouçam e olhem,
Venham, venham, verão
O Frizão da Bahia,
Que está retratado
Ás maravi, maravi, maravilhas.
No corpazil torreão
A natureza prevista
Formou a fresta da bocca
Para guarita.
Ouçam e olhem,
Venham, venham, verão
O Frizão da Bahia,
Que está retratado
Ás maravi, maravi, maravilhas.
É grande Conimbricense
Sem jámais pôr pé em Coimbra,
E sendo ignorante sabe
Mais que gallinha.
Ouçam e olhem,
Venham, venham, verão
O Frizão da Bahia,
Que está retratado
Ás maravi, maravi, maravilhas.
É dotado de um entendimento
Tão vivo e esperto, que fôra um Beliz,
Si lhe houvera o juizo illustrado
Um dedo de grego, outro de latim.
Entre gabos o triste idiota
Tão pago se mostra dos seus gorgotis,
Que nascendo sendeiro de gemma,
Quer á fina força metter-se a rossim.
Marinicolas é finalmente
Sugeito de prendas de tanto matiz,
Que está hoje batendo moeda,
Sendo ainda hontem um vilão ruim.