Track B, targeted to voice interface experience designers, contains many customer case studies illustrating new and innovative applications using speech technologies, including speech in the home and in the car. Learn new insights for using speech technologies and how they could be used in your enterprise. Track B also contains discussions about the latest best practices and techniques for voice user experience design.
Monday, April 27: 10:30 a.m. - 11:15 a.m.
At its best, spoken languages connect two minds. Conscious and sentient humans have minds. Machines do not. In this session, Balentine describes a continuum of three conversational idioms— task-speech, thought-speech and social-speech—that together define a newly emerging “alien mind,” which is useful and productive, unlike many human minds. Intelligent conversation with an alien mind bridges the gap from today’s early (but mindless) successes to tomorrow’s ubiquitous interactions with agents, robots, kiosks, and disembodied entities.
Bruce Balentine, Chief Scientist Emeritus, Enterprise Integration Group
Monday, April 27: 11:30 a.m. - 12:15 p.m.
Defining the intents, dialogue responses, and variations can all increase the time to create a conversational assistant. Yet the content needed to power an assistant may already exists in places like public FAQ pages or knowledge management systems. This talk addresses how to leverage document understanding AI to extract valuable content and uses AI search to find and combine speech and visual interactions to quickly bootstrap an effective conversational interaction.
Anish K. Mathur, Senior Offering Manager, IBM Watson
Monday, April 27: 1:15 p.m. - 2:00 p.m.
High-quality Actions see high engagement with Google Assistant users. What makes them high quality? Drawing from real case studies, learn how to reduce development errors, enhance discovery of your Action, grow your userbase, and avoid mistakes along the way when building quality and engaging Actions.
Erica Cei, Conversation Designer, Google
Monday, April 27: 2:15 p.m. - 3:00 p.m.
Open to all full-conference attendees and Networking Pass holders. Check SpeechTEK.com for details or follow #SpeechTEK on Twitter.
Monday, April 27: 3:15 p.m. - 4:00 p.m.
This talk reviews techniques for gaining insights through text mining of recorded customer interactions, including the following: scoping and sampling to ensure that a text mining solution is quick to market, economic to scale, and conformable to other processes or speech solutions in the enterprise; champion/challenger testing of multiple text mining platforms; understanding the benefits of sentiment analysis; and using unsupervised clustering to illustrate regularities in elements such as products, features, events, and responses.
Dan Coyer, Senior Analyst, West Corporation
Monday, April 27: 4:15 p.m. - 5:00 p.m.
This talk discusses the benefits of performing a deep analysis of user turns in a conversation. As an example, we review a case where, through deep analysis of bot interactions, the U.S. Army learned that military personnel were frequently asking the bot about post-traumatic stress disorder (PTSD) symptoms and support. Consequently, the Defense Health Agency was informed that it needed to provide more anonymous self-service PTSD resources for soldiers.
Ian Beaver, Chief Scientist, Verint-Next IT
Tuesday, April 28: 10:45 a.m. - 11:30 a.m.
With all of the hype around smart speakers, it’s easy to get the impression that there’s nothing else interesting happening in speech technology today. In this presentation, you hear about some innovative applications of speech technology which don’t involve smart speakers.
Jeff Adams, CEO & Founder, Cobalt Speech & Language
Tuesday, April 28: 11:45 a.m. - 12:30 p.m.
Smart speakers have the fastest adoption of new technology in human history—even faster than the smart phone. However, most brands have had very limited success with their Alexa Skills and similar voice experiences. Why is that? Join Tobias Dengel as he explores how some of the largest brands in the world, such as HBO, Fox, Regal Cinemas and Synchrony Bank, are shifting their thinking from voice to multimodal to truly take advantage of the platform.
Tobias Dengel, CEO, WillowTree
Tuesday, April 28: 1:45 p.m. - 2:30 p.m.
We analyzed the emotional speech outputs of films produced by eight renowned directors, characterizing their work on an emotional spectrum. We then developed a rich recommendation system, which is more objective because it takes into account the script’s and director’s intentions. The emotional charge of a movie can influence our preferences as it gives us one more piece of information to decide what we want to see next.
Rana Gujral, CEO, Behavioral Signals
Tuesday, April 28: 2:45 p.m. - 3:00 p.m.
We describe the process of designing, implementing, and using an intelligent agent for collecting field data in the food supply chain. We present our key learnings about the process, methods, and acceptance of speech technology. We describe the advantages and limitations of speech interaction methods, as well as the value of a human-in-the loop at key stages. Lastly, we make some predictions about the future of using speech in commercial and industrial use cases.
John Swansey, Co-Founder & Chief Design Officer, AgVoice
Tuesday, April 28: 4:15 p.m. - 5:00 p.m.
Remote patient monitoring and screening overcome the shortage in mental health professionals and long waiting periods while significantly reducing medical costs. We use remote, objective, continuous patient monitoring and risk group screening based on an evidence-based vocal biomarker. The technology tracks changes in medical states by automatically monitoring patient speech patterns captured in everyday mobile interactions. Voice analysis extracts unique prosodic (non-content) speech features such as intonation, rhythm, pace, and emphasis, which glean universal voice parameters to address the physiological aspects of speech.
Yoav Degani, CEO, VoiceSense
Wednesday, April 29: 10:45 a.m. - 11:30 a.m.
Current AI and machine learning (ML) technologies are beginning to change the way we build and innovate. However, the power of our current ML technologies is not fixed. This session explores where ML is at the moment and where it is heading in the next 10 years. How are its current use cases different from those we may see in the future? This session explores these themes and their impact on speech technology applications.
Sam Ringer, Machine Learning Engineer, Speechmatics