SpeechTEK.com home
Monday Tuesday Wednesday
Sunrise Discussions Customer Case Studies SpeechTEK Labs
Keynotes Register Now SpeechTEK University
SpeechTEK 2010 - Wednesday, August 4, 2010
SUNRISE DISCUSSIONS
SD301 – Speech Technologies for the Developing World
8:00 a.m - 8:45 a.m
MODERATOR:
Max Froumentin, Program Manager - Web Foundation

In much of the developing world, access to the internet via personal computers remains quite low, while mobile phone use is widespread. Unlike developed countries, in which individuals rely on the personal computers for access to information and self-service transactions, in the developing world, mobile phones are the primary means of communication and data access. Join us for a discussion of how speech technologies on mobile phones can be used to improve the lives of millions of people in developing countries.

SD302 – Voice Texting While Driving: Is It Safe?
8:00 a.m - 8:45 a.m
MODERATOR:
Dr. Juan E. Gilbert, Andrew Banks Family Preeminence Endowed Professor & Chairman, Computer & Information Science & Engineering Department - University of Florida

Recent legislation in several states and municipalities around the country require drivers to use hands-free kits to use their mobile phones and have banned texting while driving. A newly developed application allows drivers to send and receive text messages using a speech-only interface, which makes the interaction similar to a hands-free phone call. Join us as we discuss the implications of voice texting and whether it sufficiently reduces the safety hazards associated with texting.

SD303 – Secret Agents and Spies, Oh My!
8:00 a.m - 8:45 a.m
MODERATOR:
Ellis 'Skip' Cave - Independent Consultant

Automated agents and human agents have their advantages and disadvantages in the call center. Can these two agents be merged? For example, can a human agent monitor automated agents, stepping in when the automated agent fails? Can users switch back and forth between human and automated agents? Is the data entered into automated agents secure if the automated agent is monitored by a human agent? What will the future of combined human/automated agents be?

SD304 – Migrating Voice-Only IVR Applications to Multimodal Applications on Moible Devices
8:00 a.m - 8:45 a.m
MODERATOR:
Dr. Matthew Yuschik, Executive Board Member - AVIOS

What are the benefits and disadvantages of migrating voice-only IVR applications to multimodal applications on mobile devices? How can voice-only applications take advantage of the small screen on many mobile devices? What changes to a voice-only application’s structure are necessary? Must the application be entirely redesigned and re-implemented?

TRACK A: BUSINESS STRATEGIES
KEYNOTE: Assessing the Mobile OS Landscape for Speech
9:00 a.m - 10:00 a.m
MODERATOR:
David Myron, Editorial Director, CRM magazine - Information Today, Inc.

Panelists from Google, Microsoft, and Symbian

Learn how you can leverage the popularity of smartphones in consumer and enterprise markets. What new customer interaction and employee productivity opportunities are enabled by these devices? What tools and training are available for developers who are looking to develop these new applications? What mix of speech, graphics, text and videos provide the best customer experience? Where are the guidelines for designing user interfaces for smartphones and how can you get started?

Break in the Exhibit Hall
10:00 a.m - 10:45 a.m
A301 – What’s Next in Speech Technology
10:45 a.m - 11:30 a.m
MODERATOR:
Dr. Ahmed Bouzid, Co-founder & President - The Ubiquitous Voice Society
Daniel Hong, Research Director & VP - Forrester Research
David Joiner, Chief Technology Officer - Autonomy

Speech technology is a quickly changing field. To make the best decisions for your business, you need to understand not just what’s happening today, but where speech technology will be tomorrow. Join us to get expert perspectives on what’s next for the speech technology industry and how to capitalize on growing trends.

A302 – Understanding Voice Biometrics
11:45 a.m - 12:30 p.m
MODERATOR:
Dr. Judith Markowitz, President - J. Markowitz Consultants
Mr. Almog Aley-Raz, CEO - PerSay
Jenny Burr, Sr. Manager, Analytic Consulting - Convergys

Speaker identification and verification (SIV) technologies offer the promise of simpler and more secure methods of giving customers access to transactions and information, but it can be challenging to understand. Attend this informative session to learn how to evaluate SIV solutions and determine if they will be right for your organization, and learn what deploying an SIV solution will mean to your customers.

Attendee Lunch
12:30 p.m - 1:45 p.m
A303 – Keeping Speech Projects on Track
1:45 p.m - 2:30 p.m
MODERATOR:
Amy Downs, Director, Customer Solutions - Voxeo, an Aspect Company
Steven Sands PMP, Director of Business Development - Walsh Media
Fran McTernan, Managing Principal - Avaya
A304 – Speech in the Cloud
2:45 p.m - 3:45 p.m
MODERATOR:
Keith Ward, CTO - Product Support Solutions (PSS)
Dr. Thomas Schalk, Vice President, Voice Technology - Agero
Dena Skrbina, Director, Solutions Marketing - Nuance Communications
Julia Webb, Executive Vice President, Sales & Marketing - VoiceVault

We’ve all heard of cloud computing, but how does speech in the cloud work, and what does it mean for your business? Attend this session to hear the views of these experts on the benefits of adopting cloud computing solutions with speech technology. Learn how speech in the cloud can help your organization improve the bottom line and simultaneously offer excellent customer service.

TRACK B: CUSTOMER EXPERIENCE
KEYNOTE: Assessing the Mobile OS Landscape for Speech
9:00 a.m - 10:00 a.m
MODERATOR:
David Myron, Editorial Director, CRM magazine - Information Today, Inc.

Panelists from Google, Microsoft, and Symbian

Learn how you can leverage the popularity of smartphones in consumer and enterprise markets. What new customer interaction and employee productivity opportunities are enabled by these devices? What tools and training are available for developers who are looking to develop these new applications? What mix of speech, graphics, text and videos provide the best customer experience? Where are the guidelines for designing user interfaces for smartphones and how can you get started?

Break in the Exhibit Hall
10:00 a.m - 10:45 a.m
B301 – Speech Beyond the Call Center
10:45 a.m - 11:30 a.m
MODERATOR:
Dr. Lizanne Kaiser, Sr. Principal Business Consultant - Genesys
Bryan Pellom, Vice President, Speech Development - Rosetta Stone

Speech technology has been a valuable asset in customer self-service applications for many years, but speech also plays an important role in other domains. Speech technology offers many benefits for both language learning applications and medical dictation systems, but optimizing speech for these contexts requires a different perspective. Join us to gain a broader view of the opportunities to use speech and gain a deeper understanding of the technology.

Multimodal Magical Data Capture and Representation
Dr. Qian Hu, Chief Scientist of Speech Technology and AI - The MITRE Corporation
B302 – Lessons From Deployments
11:45 a.m - 12:30 p.m
MODERATOR:
Kristie Goss-Flenord, Consultant, Human Factors - Convergys
David C Martin, Managing Principal/Portfolio Leader - Avaya
Howie Cahn, Principal Speech Engineer - Verizon Business

In many cases, the best way to learn is by example. Come to this session to hear stories of speech deployments from two different industries and discuss the lessons learned. Learn how the design of troubleshooting applications and an auto-attendant for retail can help you with the design, evaluation, and deployment of your speech application and enable you to provide your customers with a superior experience.

Attendee Lunch
12:30 p.m - 1:45 p.m
B303 – Breaking Barriers in Voice Interaction
1:45 p.m - 2:30 p.m
MODERATOR:
Jim Milroy, Human Factors Solutions Consultant - West Interactive Services
Karen Owens, Senior User Experience Designer - LogicTree Corporation
Michele A Burton, Accessibility Analyst - SSB BART Group

Successful speech interactions often come down to getting the details right. Experts in this session delve into details that will help you design great interactions. Learn strategies for dealing with the common problem of voice searches that return a large number of results, and how to best present these results to users of speech applications. Next learn how analyzing the self-service offerings of your competitors can help your organization form a cohesive and unique customer experience strategy.

B304 – Solve My VUI Problem!
2:45 p.m - 3:45 p.m
MODERATOR:
Dr Melanie Polkosky, Human Factors Psychologist/Consultant - IBM Corporation
Jenni McKienzie, Voice Interaction Designer - SpeechUsability
Mr David Attwater, Senior Scientist - Enterprise Integration Group
Eduardo Olvera, Sr. Manager & Global Emerging Technology Lead, UI Design, Professional Services - Nuance

Every speech application has one: that prompt or menu that just doesn’t work the way you thought it would. This expert panel has the answers! Bring your most perplexing VUI issues and let the VUI design experts debate the right solution for you. This lively session will give you the answers you need and provide an inside view of how industry experts solve VUI problems.

TRACK C: NEXT-GEN SPEECH TECHNOLOGIES
KEYNOTE: Assessing the Mobile OS Landscape for Speech
9:00 a.m - 10:00 a.m
MODERATOR:
David Myron, Editorial Director, CRM magazine - Information Today, Inc.

Panelists from Google, Microsoft, and Symbian

Learn how you can leverage the popularity of smartphones in consumer and enterprise markets. What new customer interaction and employee productivity opportunities are enabled by these devices? What tools and training are available for developers who are looking to develop these new applications? What mix of speech, graphics, text and videos provide the best customer experience? Where are the guidelines for designing user interfaces for smartphones and how can you get started?

Break in the Exhibit Hall
10:00 a.m - 10:45 a.m
C301 – Advanced Research in Human Machine Spoken Dialogue Interaction
10:45 a.m - 11:30 a.m
MODERATOR:
K.W.'Bill' Scholz, President - NewSpeech LLC

This session will provide an overview of the latest trends and research results from the academic and scientific community involved in speech and language research. Reports from the major international research conferences will highlight the state of the art, what is possible today, and provide useful insights and visions for the future of these technologies.

Spoken Natural Language Processing Research
Bhuvana Ramabhadran, Research Staff Member, Manager - IBM T.J. Watson Research Center
Roberto Pieraccini, Director of Engineering - Google Switzerland
C302 – Next-Generation Natural Language Processing
11:45 a.m - 12:30 p.m
MODERATOR:
Roberto Pieraccini, Director of Engineering - Google Switzerland
Application Concurrency for Smartphones
Kurt Fuqua, CEO - Cambridge Mobile

Spoken natural language understanding has assumed more and more importance in the development of spoken interaction application with machines during the past few years. This session will shed light on the recent advancements in the field.

Recent Advances in Natural Language Understanding
Srinivas Bangalore, Principal Technical Staff Member - AT&T Labs -- Research
Attendee Lunch
12:30 p.m - 1:45 p.m
C303 – Advanced Voice User Interface and Dialogue
1:45 p.m - 2:30 p.m
MODERATOR:
Dr. Thomas Schalk, Vice President, Voice Technology - Agero
Phil Shinn PhD, CTO - ImmunityAnalytics
Mahmoud Fahmy, CEO and Chairman of IST Group - IST Networks

As we move toward more sophisticated applications that are able to engage users in complex and long interactions, VUI design has evolved to provide better and better understanding of the principles behind human-machine communication. This session will provide a useful forum to understand and discuss the issues and the solutions of modern VUI.

C304 – Toward Anthropomorphic Systems
2:45 p.m - 3:45 p.m
MODERATOR:
Roberto Pieraccini, Director of Engineering - Google Switzerland
Mr. David L Thomson, VP Speech Technology - CaptionCall
Enrico Zovato, Chief TTS Researcher, Voice Technologies - Loquendo

There has always been a conundrum between the idea of creating systems that imitate human communication abilities and systems that focus on usability without having necessarily anthropomorphic characteristics. This session provides an interesting discussion about humanlike systems and the technology behind them.

TRACK D: VOICE INTERACTION DESIGN
KEYNOTE: Assessing the Mobile OS Landscape for Speech
9:00 a.m - 10:00 a.m
MODERATOR:
David Myron, Editorial Director, CRM magazine - Information Today, Inc.

Panelists from Google, Microsoft, and Symbian

Learn how you can leverage the popularity of smartphones in consumer and enterprise markets. What new customer interaction and employee productivity opportunities are enabled by these devices? What tools and training are available for developers who are looking to develop these new applications? What mix of speech, graphics, text and videos provide the best customer experience? Where are the guidelines for designing user interfaces for smartphones and how can you get started?

Break in the Exhibit Hall
10:00 a.m - 10:45 a.m
D301 – What’s Behind the Curtain? Observe a Live WOZ Test
10:45 a.m - 11:30 a.m
Jim Milroy, Human Factors Solutions Consultant - West Interactive Services

Wizard-of-Oz (WOZ) usability testing is a well-known technique for evaluating speech applications early in the project life cycle, but many people do not really understand how a WOZ test really works. Attend this training session as we pull back the curtain and give you a behind-the-scenes view of a live WOZ usability test. This live demonstration will let you experience the benefits of this testing method for yourself and learn how it can benefit your organization.

D302 – User Feedback and Speech Applications
11:45 a.m - 12:30 p.m
Peter Leppik, President and CEO - Vocal Laboratories Inc.
Emily Selene de Rotstein, Vice President, Sales and Marketing - Vocal Laboratories Inc.

Sooner or later, every speech application will generate user feedback — whether it’s from usability testing, customer surveys, or complaints from your mother. How can you use this feedback to improve? How do you deal with unexpected complaints? What steps should you take to manage the feedback process? Learn how to systematically build customer feedback into your speech project, or deal with the feedback you’re getting from the application you’ve already deployed.

Attendee Lunch
12:30 p.m - 1:45 p.m
D303 – 30 Ideas in 45 Minutes: Tips for Better Experiences
1:45 p.m - 2:30 p.m
MODERATOR:
Jenni McKienzie, Voice Interaction Designer - SpeechUsability
Jonathan Bloom, Voice User Interface Designer - Jibo, Inc.
Ms. Karen Kaushansky, User Experience Strategist - Microsoft Tellme
Julie Underdahl, Senior User Interface Designer - Genesys
Dr Melanie Polkosky, Human Factors Psychologist/Consultant - IBM Corporation

Calling all VUIDs! Bring your best tips and tricks and help us assemble the 30 best new ideas in voice interaction design. You’ll leave this rapid-fire session armed with the latest techniques, and lots to think about. Come to learn or just to listen in on the debate, but be sure to join us!

D304 – Effective User Research Techniques
2:45 p.m - 3:45 p.m
MODERATOR:
Susan L. Hura PhD, Principal - SpeechUsability
Mary Constance Parks, Principal Experience Designer, Automation and Control Solutions - Honeywell
Dr. Lizanne Kaiser, Sr. Principal Business Consultant - Genesys
Catherine Zhu, Principal Consultant - SpeechUsability

One of the most neglected phases in the design of speech and multimodal applications is early user research. Organizations often reject user research because they do not see the benefits it can bring to self-service projects and believe it is costly and time-consuming. Join this panel of experts as they dispel these misconceptions and explain quick and effective user research methods, and show how it can be used to guide later design and deployment decisions.

TRACK E: LESSONS
KEYNOTE: Assessing the Mobile OS Landscape for Speech
9:00 a.m - 10:00 a.m
MODERATOR:
David Myron, Editorial Director, CRM magazine - Information Today, Inc.

Panelists from Google, Microsoft, and Symbian

Learn how you can leverage the popularity of smartphones in consumer and enterprise markets. What new customer interaction and employee productivity opportunities are enabled by these devices? What tools and training are available for developers who are looking to develop these new applications? What mix of speech, graphics, text and videos provide the best customer experience? Where are the guidelines for designing user interfaces for smartphones and how can you get started?

Break in the Exhibit Hall
10:00 a.m - 10:45 a.m
E301 – Creating Conversational Mobile Applications for Android
10:45 a.m - 11:30 a.m
Kurt Fuqua, CEO - Cambridge Mobile

Create sophisticated speech-enabled applications for Android. This tutorial will walk you through the process with downloadable sample code for an actual application. Design the voice interface for natural conversation. Enable interaction with other applications. Support multiple languages. Learn how to get started and what to avoid. Taught by the man who gave Android its voice.

E302 – Developing and Deploying Portable and Interoperable Multimodal Applications
11:45 a.m - 12:30 p.m
Raj Tumuluri, President - Openstream Inc.

This session will provide hands-on training in the W3C Multimodal Architecture using interoperable components developed by Openstream, Loquendo, and other participants in the W3C Multimodal Interaction Working Group. The session will overview Openstream’s CueMe implementation of the W3C multimodal architecture and Loquendo’s MRCP speech server. Participants will be provided with a research version of the CueMe platform, documentation, and development tools that will be used in a hands-on multimodal development exercise. Both client-server and offline component distribution models will be discussed.

Attendee Lunch
12:30 p.m - 1:45 p.m
E303 – Multimodal Applications Using EMMA
1:45 p.m - 2:30 p.m
Dr. Michael Johnston, Lead Inventive Scientist - Interactions

Extensible MultiModal Annotation (EMMA) provides a common language for communication among system components, enabling plug-and-play of technologies from different vendors and rapid prototyping and iterative development of multimodal interfaces. In this hands-on tutorial, participants will use a cloud-based multimodal server supporting EMMA (AT&T Speech Mashup) and the Deutsche Telekom EMMA implementation to learn how to build multimodal applications with EMMA that incorporate advanced features such as multimodal dialogue confirmation and fusion of multiple input modalities.

E304 – Making VoiceXML Applications Portable
2:45 p.m - 3:45 p.m
Mike Monegan, Vice President Product Management - Cyara

Does application portablility matter, even with the same browser platform vendor? In practice, significant compatibility challenges often unfold when migrating a speech application from one browser platform or automatic speech recognition (ASR) engine to another. What are the top issues affecting application portability? Learn how to avoid migration problems when you initially write application code. Learn how to migrate an existing application to a new platform. Apply strategies for mitigating browser platform and ASR differences.




Connect with
SpeechTEK 2010

#STEK10

Gold Sponsors
Monday
Lunch Sponsor
Tuesday
Lunch Sponsor
Wednesday
Lunch Sponsor
Media Center
Sponsor
Tuesday
Break Sponsor
Media Sponsors