How Will My Voice & Video Files Be Used - 240411 - 142148

You might also like

Download as pdf or txt
Download as pdf or txt
You are on page 1of 2

How will my voice & video files from this project be used?

Great question! The recordings will be used to improve existing text-to-speech (TTS) models. Projects
like this allow companies to improve how well their AI products (think Alexa, Siri, or your in-car
navigation) can understand the human voice, and how conversational it can sound when responding to
you. But to clarify, your voice will not be heard coming out of a smart speaker for this project.

Question: Will my voice be used to synthesize & create derivative works by the client? Will the voice
be used to create derivative works by Third Parties? If so, will derivative works be broadcast,
commercialized, or distributed?

Answer: No. No new content will be created with your voice by Voices, or the client.

Question: Will the voice be used to create, improve, or expand existing speech services models?

Answer: The voice data will be used to improve existing TTS speech models.

Question: Will my voice be synthesized and/or combined with other voices (derived voices) or will the
voice be cloned in its current timbre?

Answer: No. The recordings will NOT be used to create a clone, or replication, of your voice or video.
Your voice and video will not be heard outside of Voices and/or any Voices’ clients. The only use for the
files is to train the TTS algorithm, with no synthesis or replication of your voice or video.

Question: Are these usage terms subject to change?

Answer: Not without further contracting and consent between Voices, our client and yourself. You have
the opportunity to decline any further projects or licensing beyond this project’s scope, should that
arise.

Question: How will my video file be used?

Answer: The videos serve the purpose of researching mouth movements during specific phrases for
future animation projects. The model will not create any version resembling you in the videos; rather,
the model aims to learn general movement and facial expressions. The computer studies the videos to
understand how facial expressions convey emotions, and it utilizes this knowledge to replicate

1
expressions on a screen/avatar. Importantly, the generated faces will not resemble the people featured
in the videos.

You might also like