2019 Schedule
Filter By Day
reset
Tracks
  • Managers
  • Implementers
  • Voice Interface Experience Designers
  • Sunday Apr 28
  • Monday Apr 29
  • Tuesday Apr 30
  • Wednesday May 1
 
SpeechTEK University - Preconference Workshops
 
Managers
Voice Interface Experience Designers
Implementers
 
Managers
Voice Interface Experience Designers
Implementers
 
Managers
Voice Interface Experience Designers
Implementers
1:30 PM
SpeechTEK University - Preconference Workshops
Length: 3 Hours
Speaker(s):
, Principal, Conversational Technologies
Description: Natural language understanding (along with speech recognition) is one of the foundational technologies underlying the Voice-First revolution. When it works well, the user experience is natural, frictionless, and efficient. When it doesn’t work well, the results can be frustrating and irritating. This session brings attendees up-to-date on current natural language understanding technology, explaining how it works and what’s going wrong when it doesn’t. We cover current technologies, including both traditional rule-based approaches, as well as machine learning technologies such as Deep Learning. We also review current proprietary natural language application tools such as the Amazon Alexa Skills Kit, Google Dialogflow, and Microsoft LUIS and discuss open source alternatives. Attendees come away from the session with an understanding of current natural language technology, its capabilities and future directions.

SpeechTEK University - Preconference Workshops
Length: 3 Hours
Speaker(s):
, Senior Scientist, Enterprise Integration Group
Description: This interactive session is suitable for anyone seeking deeper experience and understanding of conversational design, including anyone working with home automation devices (e.g., Alexa), chat-bots, or conversational IVR. The session takes an in-depth look at the principles underlying conversational design with particular emphasis on human-computer conversation. It is suitable for designers or decision makers who would appreciate a deeper understanding of the different aspects of conversational design. Topics include the following: principle of speech recognition, including semantics, slots, and parsing; human psychology, including memory and learning; dialogue acts, grounding, discourse markers, and confirmation; language continuity, including anaphora; and narrative voice, persona, and social identity.

SpeechTEK University - Preconference Workshops
Length: 3 Hours
Speaker(s):
, Solutions Architect, Amazon Alexa
Description: Alexa is Amazon’s cloud-based voice service and the brain behind tens of millions of devices, including the Echo family of devices, FireTV, Fire Tablet, and third-party devices with Alexa built-in. You can build capabilities, or skills, that make Alexa smarter and make everyday tasks faster, easier, and more delightful for customers. Tens of thousands of developers have built skills using the Alexa Skills Kit (ASK), a collection of self-service APIs, tools, documentation, and code samples. This is a hands-on workshop where attendees learn how to create voice experiences for Amazon Alexa. We discuss voice design best practices, show how to leverage cloud services and APIs, discuss the latest Alexa features, and share code samples to get your project started. The final hour of the session is set aside for open hacking time, where you can get one-on-one support from an Amazon Alexa solutions architect. Developers who wish to follow along should create accounts on the following sites before attending: aws.amazon.com and developer.amazon.com.

9:00 AM
Welcome & Opening Keynotes
Length: 1 Hour
Speaker(s):
, Speaker, trendwatcher, author
Description: This presentation discusses the most important technological development of the coming years: artificial intelligence. What can companies expect? What should we do as more and more human skills are being taken over by these kind of systems, such as looking, listening, talking, reading and reasoning? And what will the future of artificial intelligence look like? What is definitely possible, and what isn't? How does this change the relationship with your customer?

The future is also brightened by the autonomous "digital butler" who gives you answers before you knew you need answers, and helps you before you knew you needed help. But does it only have advantages? Or not? What are the unintended, unforeseen disadvantages of this technological trend? Just like we should have done with the introduction of social media and the smartphone, shouldn't we be asking ourselves that one important question: what do we want the era of artificial intelligence to look like?


10:00 AM
Welcome & Opening Keynotes
Length: 15 Minutes
Description: Check back for the latest details.

10:15 AM
Coffee Break
Length: 15 Minutes
10:30 AM
Managers
Length: 45 Minutes
Speaker(s):
, Principal, Conversational Technologies
, CEO, Nu Echo Inc.
Description: Choosing the right platform for a particular application is critical. We focus on the capabilities of current natural language understanding and dialog management technologies. We present evaluation criteria and compare and contrast a number of popular platforms. We consider technical and non-technical features such as cost, vendor commitment, support, and cloud vs. on-premise operation. We review the state of the art for these technologies and conclude with comments on advanced dialog features that are currently being researched.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, Human-Computer Technical Discipline Lead, NASA/Johnson Space Center
Description: This presentation provides the audience with an overview of the challenges for implementation of voice control in space applications that include the hardware, software, environment, and, more importantly, the astronaut. Past voice control applications in space are given. Learn how to apply key learnings from these applications to applications here on Earth.

Implementers
Length: 45 Minutes
Speaker(s):
, University of San Francisco
, CEO, VoiceVibes, Inc.
, Co-Founder & CEO, RankMiner Predictive Voice Analytics
, Senior Analyst, West Corporation
, CTO & Co-Founder, Cogito
Description: Learn how AI is used in a call center environment to train, assist, monitor, and advise human agents as they interact with customers as well as how to predict employee departures and prescribe targeted interventions. How a graphical representation of the client interaction assists the human agent is discussed, along with how a combination of words and non-verbal analysis can detect the emotional state of customers and agents and guide agents in-the-moment to adjust their behavior for improved outcomes.

11:30 AM
Managers
Length: 45 Minutes
Speaker(s):
, President, TMA Associates
Description: An automated digital assistant—whether a text chatbot or a voice-interactive part of your telephone customer service—can help customers receive quick results and help agents focus on more complex tasks. Today’s natural language technology can make the experience fast and pleasant—when done properly. Like any evolving technology, it can also be done poorly. This talk discusses how to achieve effective solutions using automated natural language interaction.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, 2-1-1 Manager, United Way
Description: Come hear how United Way implemented a visual IVR with Amazon Connect to augment its existing call center. The goal is to help those in need 24/7, so that more people quickly find assistance. The visual IVR provides smartphone callers with emergency shelter locations in an easy-to-understand visual format. The visual IVR shortens calls, improves call containment through visual self-service, decreases switching to live staff, and reduces follow-up calls.

Implementers
Length: 45 Minutes
Speaker(s):
, Business Analyst, First Associates Loan Servicing
Description: To improve agent performance via insights garnered through our speech AI platform, we enter a series of keywords and phrases into the program, which parses through our library of transcribed call recordings and returns the requested results. This data is then cataloged and used to generate actionable items, develop call center reports, and identify training opportunities. The automated grading system measures agent performance, while the speech AI enables us to perform rapid changes to campaign strategy to eliminate areas of risk and improve agent performance.

12:15 PM
Managers
Length: 1 Hour
Description: Check back for the latest details.

1:15 PM
Managers
Length: 45 Minutes
Speaker(s):
, CEO, Nu Echo Inc.
Description: Good NLU accuracy requires a sizeable training corpus made of sentences that represent expected responses from real users. How are new chatbots developed when there is little or no training data available? We present best practices to generate a NLU training corpus to easily train a fairly robust NLU system that can be used in a customer-facing chatbot, which makes it possible to quickly start collecting real sentences from real users.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, Head of Multidisiplinary studies, HIT Holon Institute of Technology. Israel
Description: See video clips of this fascinating avatar as a virtual personal assistant for pregnant women. Learn how focus groups revealed what women really want from this application and the importance of the user-avatar relationship to the success of the project. Learn the strategy behind its design, including technical and user interface considerations. Discover which features, including trust, reliability, and visuality, are most important to real users.

Implementers
Length: 45 Minutes
Speaker(s):
, Principal, IVR Design Group
, President, ID R&D
, Senior Manager, Commercial Security Strategy, Nuance Communications
, Solution Delivery Manager, LumenVox
, VP of Product Marketing, Pindrop
Description: How can voice and behavior biometrics seamlessly verify that users are who they claim to be in real time? How can fraudsters be detected based on their voice prints, behavior anomalies, reconnaissance tactics, and robotic dialing techniques? Explore use cases and realworld examples for establishing security, identity, and trust between your organization and your customers. We share best practices and bloopers to help you have a successful voice biometrics deployment.

2:15 PM
Managers
Length: 45 Minutes
Speaker(s):
, Professor, Ulster University
Description: We talk with text to our smartphones, smart speakers, and other devices in a conversational style; but how conversational are these interfaces? What happens when things go wrong—when the user or agent needs to ask for clarification, if something needs to be corrected, or something has been misunderstood? Learn whether and how these “edge cases” are being handled using available tools and what new approaches are being developed in research labs.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, CEO, MindSquare
Description: Over the last few years, brain injuries have moved to the forefront of health issues and concerns, particularly highlighted by challenges in professional sports. With greater attention to the issue, the ability to diagnose injuries is evolving. Now, there is an opportunity to leverage speech technology to more immediately identify injuries. Working with MindSquare, AppTek is assisting in the diagnosis of concussions (or mild traumatic brain injuries) using a mobile device.

Implementers
Length: 45 Minutes
Speaker(s):
, CEO, VoiceSense
, Chief Scientist of Speech Technology and AI, The MITRE Corporation
Description: Behavioral speech analytics identifies typical speech patterns – prosodic content, the non-content parts of speech – intonation, pace, emphasis – that reflect common behavioral patterns. Behavioral speech analytics can provide a fairly strong prediction of the individual’s anticipated behaviors in various life situations. Improved call classification and routing is achieved by combining speech technology with robust natural language understanding and other Artificial Intelligence techniques.

3:15 PM
Managers
Length: 45 Minutes
Speaker(s):
, President and CEO, Interactive Media North America
, Executive Director of the Speech Solutions Professional Services, TTEC
Description: Learn more about the latest technological developments in speech technology from leading-industry companies in Expert Perspectives sessions. These sessions are open to all full-conference attendees and Networking Pass holders. Check back for session details or follow #SpeechTEK on Twitter for updates.


Title: Give your chatbot the gift of voice
Time: 3:15 PM - 3:35 PM
Description: Millions of businesses implement chatbots, with mixed success. Chatting online is useful, but often a voice conversation is better. Speaking is faster than typing, safer and much more natural, plus everyone can use a phone. In this session, we discuss how to add voice and telephony to your bot.


Title: Utilizing AI in your Customer Service Channel Journey
Time: 3:40 PM - 4:00 PM
Description: As part of overall AI strategies, organizations are struggling with how best to leverage AI to provide a seamless customer experience throughout all their channels.  In this session we discuss how to securely empower your Speech enabled IVR with AI and seamlessly extend that AI experience to an enriched associate interaction when needed.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, Technical Product Director, Gannett
Description: Newsrooms have complex workflows to produce content at highest journalistic standards—Gannett does this efficiently and at massive scale across its 100-plus news properties. Newsroom workflows changed from print, to web, to mobile device, and now to social. The voice revolution calls for another metamorphosis to the newsroom’s workflow. Learn how Goff transformed the newsroom workflow for voice with lessons learned from the web, mobile, and social revolutions.

Implementers
Length: 45 Minutes
Speaker(s):
, VP Speech Technology, CaptionCall
, Research Fellow, Edinburgh Medical School, the University of Edinburgh, UK
, Reader, Usher Institute of Population Health Sciences and Informatics, Edinburgh Medical School, the University of Edinburgh, UK
, NLP Scientist, Interactions LLC
Description: Descriptions and demonstrations of two of the most innovative applications created at universities selected from presentations at scientific conferences. Presentations include:

- Visual, Laughter, Applause and Spoken Expression Features for Predicting Engagement Within TED Talks—How camera angles, audience’s laughter and applause, and the presenter’s speech expressions help in automatic detection of user engagement. 

- Cues to Deception in Text and Speech—How machines detect deceptive behavior. We describe a corpus for researching deceptive natural language, features that are useful cues to deception, and the role of individual differences in deceptive behavior.


4:15 PM
Managers
Length: 45 Minutes
Speaker(s):
, Strategic Sales Consultant, Radish Systems
Description: Phonebots are especially useful when enhanced with visual information, such as an instantly viewed, tappable menu of options instead of a long sequence of questions or a spoken list of options. Adding a visual component—maps, photos, video snippets, menus, graphics, diagrams, short documents—to an ordinary phone call clarifies the users’ requests and can encourage customers to stay on the automated bot to reach a satisfactory resolution.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, Product Manager, Google Cloud Al
Description: Bots have been around for a few years now, but most aren’t considered great user experiences. In this session, you'll learn about the right elements for building a bot that provides them using Dialogflow and Cloud Speech technologies. We will show you how to build a simple bot, as well as describe advanced techniques you can use to increase your completion rate.

Implementers
Length: 45 Minutes
Speaker(s):
, Conversation Design Lead, Allstate
, Conversation Designer, Allstate
Description: During this talk, conversation designers from Allstate share their experiences designing for a variety of interfaces with the goal of creating a unified experience for the audiences they serve. Get practical ideas for how your team can start sharing data, establishing common patterns, and iterating designs based on user research. Attendees also see a case study showing how designers working on separate voice and chat products find common ground when working on the same subject matter.

5:00 PM
Reception
Length: 2 Hours
Description: Join your peers as we celebrate the grand opening of the Customer Solutions Expo. Visit with conference sponsors, exhibitors, speakers, and other attendees while enjoying light hors d’oeuvres and drinks.

8:00 AM
Sunrise Discussions
Length: 45 Minutes
Speaker(s):
, Technologist and CEO, Spelamode Consulting
, Of Counsel, Tully Rinckey, PLLC
Description: New definitions associated with the Alice decision and the America Invents Act continue to be adjusted by the patent office and affected by court decisions. This morning session discusses new guidance about the Alice definition of Abstract Ideas and impacts of recent court decisions on the America Invents Act.

Sunrise Discussions
Length: 45 Minutes
Speaker(s):
, Principal, Conversational Technologies
, Senior Scientist, Enterprise Integration Group
, Chief Scientist, Enterprise Integration Group
, President, J. Markowitz Consultants
, Professor, Ulster University
, Head of Multidisiplinary studies, HIT Holon Institute of Technology. Israel
, President, TMA Associates
Description: Participate in the interactive Knowledge Café, where you can share your speech technology questions and challenges with colleagues and practitioners.

Topics and Consultants:

 - User Interfaces: David Attwater, Senior Scientist, Enterprise Integration Group

 - User Interfaces: Bruce Balentine, Chief Scientist, Enterprise Integration Group

 - Natural Language: Deborah Dahl, Principal, Conversational Technologies

 - Speaker Identification: Judith Martowitz, President, J. Markowitz Consultants

 - Intelligent Assistants: Michael McTear, Professor, Ulster University

 - multimotal systems: Nava A Shaked, Head of Multidisiplinary Studies, HIT Holon Institute of 

   Technology

 - Speech Technology Business Strategies: William Meisel, President, TMA Associates


Sunrise Discussions
Length: 45 Minutes
Speaker(s):
, Sr. Consultant, Human Factors Concentrix
Description: The ACIxD organization (formerly AVIxD) created a wiki to share best practices in VUI design, but the industry is fast outgrowing the wiki. VUI professionals now design all types of conversational interactions, and it’s time to update the wiki to reflect this. We invite all conversational designers to this interactive whiteboard session where we work to expand the wiki. Come prepared to participate and lend your expertise to brainstorm where to take the wiki next!

9:00 AM
Keynote
Length: 45 Minutes
Description: More information coming soon.

9:45 AM
Keynote
Length: 15 Minutes
Description: Check back for the latest details.

10:00 AM
Coffee Break in the Enterprise Solutions Showcase
Length: 45 Minutes
10:45 AM
Managers
Length: 45 Minutes
Description: Learn more about the latest technological developments in speech technology from leading-industry companies in Expert Perspectives sessions. These sessions are open to all full-conference attendees and Networking Pass holders. Check back for session details or follow #SpeechTEK on Twitter for updates.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, Chief Scientist, Enterprise Integration Group
Description: Non-trivial user interfaces—those that require multiple turns to accomplish complex tasks—benefit when user and machine adapt to each other. Champions of voice claim that speech uniquely exhibits this plasticity. But it doesn’t unless the interface is designed to be discoverable. Discoverability requires systemic characteristics including trust, user-initiated backup moves, rewards for experimentation, and internal transparency. This session discusses specific design techniques that allow and encourage user exploration with low risk and a likely early payoff.

Implementers
Length: 45 Minutes
Speaker(s):
, President, J. Markowitz Consultants
Description: Treating objects like smart speakers, robots, and smart devices as human is called anthropomorphism. Some users may forget that some of these devices are not human and expect human-like responses and advice. This can lead to unfortunate situations with potential social and legal repercussions. Anthropomorphism can also lead to isolation and loss of human association. Designers need to understand potential social and ethical issues surrounding anthropomorphism and take steps to minimize these problems.

11:45 AM
Managers
Length: 45 Minutes
Speaker(s):
, President, Disaggregate Corporation
, CTO & Technical Lead, TEN DIGIT Communications
Description: When all your technology resides in-house or with single vendor, you can find all the data you need to monitor performance, resolve errors, and make improvements. However, if you rely on microservices from multiple vendors, without careful planning, you might not even notice errors. This talk focuses on strategies and possible solutions to solving problems in a multi-vendor/microservices environment.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, Lead, UX Research and Design, Witlingo
Description: Consider the analogy of “bad first dates” and “failed engagements” to highlight with examples the common pitfalls designers must avoid to design voice experiences that people enjoy and, most importantly, come back to. We discuss why so many voice interactions don’t go beyond the “first date.” Using the intimate dating experience as a metaphor, Javalagi explores how the subtle conventions of human-human interactions help us outline some key, guiding principles for designing delightful and meaningful voice interactions.

Implementers
Length: 45 Minutes
Speaker(s):
, Senior Creative Technologist, Virgin Voyages
, CEO, GoVivace Inc.
, Of Counsel, Tully Rinckey, PLLC
, CEO and Founder, Voice Biometrics Group
Description: With the advent of machine learning and neural nets and the proper amount of data, we can guess accurately ID, gender, language, maybe age, and more. What are the ethics involved in creating a biometric-based lie detector, or possibility a sexual-preference detector? Where should—and how do—we draw the line?

12:30 PM
Managers
Length: 1 Hour 15 Minutes
Speaker(s):
, Product Manager, Google Cloud Al
Description: AI can now help improve contact centers in ways that up until just a few years ago where not possible. Google Cloud AI enables anyone to tap into AI built on Google tech that up until recently has been exclusive to Google employees. This includes our pre-trained ready-to-use models, including speech recognition that is now twice as accurate for phone calls, WaveNet-based neural network speech synthesis, conversational NLU, and conversational analytics. Together with partners, Google is now bringing this technology to contact centers via Contact Center AI solutions. Companies with contact centers of all sizes can now automate conversational experiences, and improve performance of human agents.

1:45 PM
Managers
Length: 45 Minutes
Speaker(s):
, EVP & CTO, Phone.com, Inc.
Description: The talk shows basic AI and ML architectures and discusses current AI limitations. Some environments such as voice networks require a different and unique AI approach to deliver value. We discuss the topic of ambient AI and how it differs from Siri or Alexa. Finally, we give the audience a few pointers about cloud-based tools that make AI accessible to any developer, while showing a demo. This presentation is a collaboration between Phone.com and Second Mind.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, Director, User Experience, Versay Solutions
Description: Spoken English is not the same as written English. This session reviews some of the academic research on the differences between spoken and written English and discusses how research results might be applied when writing material that is intended to be spoken aloud for a voice-enabled interface. Also, do these principles apply to casual text conversations such as chatbots? How might these principles be factored into a “listenability index”?

Implementers
Length: 45 Minutes
Speaker(s):
, CMO & CSO, Artificial Solutions
Description: If you listen to the scaremongers, the future of the human race is at the mercy of AI. Are we destined to become a sluggish race ruled by robots, or will our own emotional intelligence prevail? This presentation examines the constraints of conversational AI, looks at the differences in skill sets between man and machine, and discusses why humans will always have a job when it comes to customer engagement.

2:45 PM
Managers
Length: 45 Minutes
Speaker(s):
, Senior Creative Technologist, Virgin Voyages
Description: When disconnected or occasionally connected to the internet, delivering a conversational experience has its own unique requirements. How to deploy a large vocabulary speech recognition engine? How to update the content? In this session, we explore multiple options in disconnected and sometimes connected technologies, and demonstrate multiple capabilities from multiple vendors. Attendees come away with a better understanding of what is capable in a disconnected world and some of the architectures and technologies that can make this happen.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, VP of Business Development, Sensory, Inc.
Description: Learn how to create a cloud-free voice UI for your next project, including low-power sound detection, wake word recognition, small vocabulary speech recognition, natural language understanding, and biometric authenticators. Learn about the various building blocks that go into engineering a voice-enabled device, such as sourcing the right integrated circuit and voice input system hardware, accessing SDKs, building command sets, training voice models, and more. See a live demonstration of an on-device personal assistant that functions totally free of an internet connection.

Implementers
Length: 45 Minutes
Speaker(s):
, Lead Designer, Grand Studio
Description: Just because we can build something, doesn’t mean we should. Voice is positioned at the forefront of technology, and as VUI designers, we are confronted with ethical decisions. This talk walks you through the kinds of ethical considerations to incorporate into your voice designs and presents tips on how to judge whether a design is ethical. Learn how to have the hard conversations with your clients and companies.

3:30 PM
Coffee Break in the Enterprise Solutions Showcase
Length: 45 Minutes
4:15 PM
Managers
Length: 45 Minutes
Speaker(s):
, Vice President Product Management, Cyara
Description: Learn how to use AI to validate AI in the realm of outside-in CX testing. The strategy for synthesizing “virtual bot testers” from linguistic and machine learning algorithms is closer than you think. This session covers machine learning algorithms for scoring bot response accuracy and maintaining proper conversational context; conversational scenario generation to stretch the limit of NLU models; configuring bots to execute regression testing in an agile, iterative delivery cycle; and customer usage examples.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, CEO & Founder, Cobalt Speech & Language
Description: What do we do when users must shout above the din in a noisy factory or vehicle, when we have to deal with accents different from our own, or when we are trying to recognize casual speech between humans? In these circumstances, most speech recognizers will break down and give poor results. We discuss strategies for mitigating the problems for such challenging conditions.

Implementers
Length: 45 Minutes
Speaker(s):
, Senior Analyst, 451 Research
, CEO, Apprente, Inc.
, SVP Product & Solutions Marketing, RingCentral
, Founder & CEO, Orion Labs
, Chief Marketing Officer, Voicea
Description: Voice is rapidly emerging as the main user interface for many apps and devices. Speech recognition and natural language understanding will change how knowledge workers interact with computers and applications, opening opportunities for innovation in human-computer interaction, including intelligent assistants in the meeting room and for team collaboration. Learn how speech adds value to enterprise applications? Discover key opportunities and challenges for speech-enabled enterprise applications. What use cases are early adopters interested in?

5:00 PM
Reception
Length: 1 Hour 30 Minutes
8:00 AM
Sunrise Discussions
Length: 45 Minutes
Speaker(s):
, Lead, UX Research and Design, Witlingo
Description: Learn the key selection criteria for choosing the perfect invocation name. Choosing a name for your voice app is a strategic and creative process, and Javalagi highlights three key perspectives: brand identity, platform capability, and usability. With real-world examples and design exercises, this discussion equips you with best practices for naming a voice application that will be successful in the real world.

Sunrise Discussions
Length: 45 Minutes
Speaker(s):
, Speech Scientist, [24]7.ai
Description: Speech recognizers are usually not equipped to deal with poor audio quality. Poor audio quality, characteristics of which include packet loss, etc., degrades recognition accuracy significantly. However, one can address this problem in a way to make the user experience more human-like. We discuss the characteristics of poor audio, ways to automatically detect these factors, and, armed with this knowledge, how we can make our automated systems more intelligent, thus improving the user experience.

Sunrise Discussions
Length: 45 Minutes
Speaker(s):
, Director, User Experience, Versay Solutions
Description: The ACIxD (formerly AVIxD) interactive whiteboard session continues for a second day. We will discuss the direction of the wiki, its organization, and brainstorm new sections to be added. As a work product, we want to produce a set of next steps to continue to support and maintain a wiki that addresses the needs of conversational interaction designers.

9:00 AM
Keynote
Length: 1 Hour
Speaker(s):
, Principal, Conversational Technologies
Description: Virtual assistants have been around for nearly 10 years, since Siri was introduced in 2010. Now is a good time to look at what they can currently do and to think about what they could do for us if only they were smarter. How close are today’s virtual assistants to human conversational abilities, and how much closer can they get? Is it important for future systems to just be able to do more things, or should they also be more emotional and sympathetic? How important is it for them to be able to socialize informally with people and have wide-ranging conversations? This talk reviews the state of the art of virtual assistants, goes over 10 important new capabilities, and discusses the technical challenges involved in improving their abilities. We also look at some interesting current academic research and talk about at how it could be applied to future systems and applications. The talk concludes with some ideas about how the industry can help advance the state of the art.

10:00 AM
Coffee Break in the Enterprise Solutions Showcase
Length: 45 Minutes
10:45 AM
Managers
Length: 45 Minutes
Speaker(s):
, Principal Analyst, Cognilytica
, Analyst, Cognilytica
Description: Cognilytica recently tested voice assistants from Amazon, Google, Apple, Microsoft, and others and quickly realized just how un-intelligent these devices are. Many are not able to provide answers to very simple questions that require simple decision making or reasoning. These assistants provide inconsistent answers among platforms and can’t deal well with variable sentence structure and other issues. We identify where these voice assistants are failing, what sort of intelligence needs to be built into the devices to make them smarter and more useful, and the current pitfalls and opportunities for companies looking to build the next generation of voice assistant.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, Vice President of Voice Technology, Sirius XM
Description: The conventional push-to-talk speech experience is being completely redesigned in an effort to achieve a conversational interface. With autonomous vehicles on the horizon, intelligent assistants can become multimodal and effectively leverage video as a presentation modality. Learn about the current intelligent assistants for the car, what to expect with vehicle integration, and how things will change with autonomous vehicles. Finally, we discuss the optimum speech experience for the driver and what’s required to achieve this optimum experience.

Implementers
Length: 45 Minutes
Speaker(s):
, Principal Engineer, Intuit
Description: Not only what, but how, a virtual assistant speaks will determine its success. We need to create a believable illusion that a bot concerns itself with the user’s situation. We need to turn engineers, designers, and content writers into emotion-aware wordsmiths who deeply care about every word and every pause, what to emphasize, and how to respond empathically. This talk explores and demonstrates possibilities of a more personalized, contextual and likeable customer engagement by using affective computing technologies and emotions analytics.

11:45 AM
Managers
Length: 45 Minutes
Speaker(s):
, CEO, GoVivace Inc.
Description: During the past 10 years, deep neural networks have transformed the field of speech recognition. However, we are still discovering some peculiarities of these networks, such as how susceptible they are to attacks. By adding an extremely small but controlled noise that is imperceptible to humans, any regular speech or music sound could be modified to generate a transcript of your choice. We give some theoretical background on this vulnerability and provide real examples of modified audio.

Voice Interface Experience Designers
Length: 45 Minutes
Speaker(s):
, Director, Digital Home Product Management, Comcast
Description: The X1 voice remote has revolutionized the TV viewing experience. Leveraging AI to transcribe and understand what users are saying, Xfinity uses direct voice controls to connect users to the content in which they are most interested. The digital home voice experience will take on concierge-like capabilities, launching features like Phone Finder, Find My Tile, and act like a search engine to connect users with more information about available services.

Implementers
Length: 45 Minutes
Speaker(s):
, Vice President, Speech-Soft Soutions
Description: We will address AI integration methods and draw practical roadmaps for migrating to digital omni-channel architectures by leveraging existing investments in IVR, chatbots and backend database interactions. Specifically:
  • Incorporating Google + and Amazon Lex AI into existing IVR investments
  • Leveraging existing IVR business logic, flow and backend database interactions to create an AI-based chatbot in a fraction of normal time
  • Standardizing a digital omni-channel approach across voice, chat, SMS, mobile and Intelligent virtual assistant channels.


12:30 PM
Last chance to visit the Customer Solutions Expo
Length: 30 Minutes
1:30 PM
SpeechTEK University - Postconference Workshops
Length: 3 Hours
Speaker(s):
, Head of Multidisiplinary studies, HIT Holon Institute of Technology. Israel
Description: Testing and evaluation processes are crucial to the success of any NLP conversational system, but testing IVR and multimodal systems presents unique challenges. Focusing on multimodal applications that involve speech and other modalities, we describe the multiple layers of testing and QA: engine quality, functional application, VUI, interfaces and infrastructure, load balancing, backup, and recovery. Learn how to set testing goals, targets, and success factors; specify and measure metrics; test and measure “soft” and “immeasurable” targets; test documentation in all stages; manage a testing project; and identify who should be on the testing team.

SpeechTEK University - Postconference Workshops
Length: 3 Hours
Speaker(s):
, Professor, Ulster University
Description: This practical, hands-on workshop introduces attendees to the concepts, methods, and issues involved in the design and development of conversational chatbots using Google’s Dialogflow tool. Following a brief introduction to chatbots and conversational interfaces, the course explores relevant technologies and tools. The main part of the workshop is devoted to hands-on design and development of some sample conversational chatbots. Bring your laptops to learn how to develop conversational chatbots.

SpeechTEK University - Postconference Workshops
Length: 3 Hours
Speaker(s):
, Principal, Conversational Technologies
Description: This workshop provides an in-depth overview of the process for developing a natural language application with current tools such as the Alexa Skills Kit and Microsoft LUIS. We start with requirements and then discuss design considerations, such as when and how to use multimodality, how to decide what intents and entities to use (and what to do if you change your mind). We address using nested and composite entities and the effect of the design on the machine learning process. Some platforms have limitations on the number of entities allowed. All platforms have some limitations on their natural language understanding capabilities—we talk about work-arounds for both issues. Finally, we review important post-development considerations, including testing, versioning, and maintenance.

SpeechTEK University - Postconference Workshops
Length: 3 Hours
Speaker(s):
, Product Manager, Voice Platforms, Gannett
, Lead, UX Research and Design, Witlingo
, Voice User Interface Designer, Witlingo
Description: What use cases lend themselves to delivering a great Alexa skill/ Google Assistant action? How does one go about identifying such use cases? This workshop begins by diving deep into several Alexa skills and Google Assistant actions to identify which ones deliver value and which ones fall short of the mark. Then we walk through the basic characteristics and principles that help us methodically assess why some skills/actions are a good fit for the Voice-First, Far-Field interface and why some are not. Finally, the presenters work through several exercises with the workshop participants and apply those characteristics and principles to systematically move from a general use case and pinpoint experiences that are best delivered through the Voice-First interface.

Don't Miss These Special Events