Muutke küpsiste eelistusi

E-raamat: Design of Multimodal Mobile Interfaces

Contributions by , Edited by , Contributions by , Edited by , Contributions by , Contributions by , Contributions by , Contributions by , Contributions by
  • Formaat: 239 pages
  • Ilmumisaeg: 25-Apr-2016
  • Kirjastus: De Gruyter
  • Keel: eng
  • ISBN-13: 9781501502750
  • Formaat - EPUB+DRM
  • Hind: 76,25 €*
  • * hind on lõplik, st. muud allahindlused enam ei rakendu
  • Lisa ostukorvi
  • Lisa soovinimekirja
  • See e-raamat on mõeldud ainult isiklikuks kasutamiseks. E-raamatuid ei saa tagastada.
  • Formaat: 239 pages
  • Ilmumisaeg: 25-Apr-2016
  • Kirjastus: De Gruyter
  • Keel: eng
  • ISBN-13: 9781501502750

DRM piirangud

  • Kopeerimine (copy/paste):

    ei ole lubatud

  • Printimine:

    ei ole lubatud

  • Kasutamine:

    Digitaalõiguste kaitse (DRM)
    Kirjastus on väljastanud selle e-raamatu krüpteeritud kujul, mis tähendab, et selle lugemiseks peate installeerima spetsiaalse tarkvara. Samuti peate looma endale  Adobe ID Rohkem infot siin. E-raamatut saab lugeda 1 kasutaja ning alla laadida kuni 6'de seadmesse (kõik autoriseeritud sama Adobe ID-ga).

    Vajalik tarkvara
    Mobiilsetes seadmetes (telefon või tahvelarvuti) lugemiseks peate installeerima selle tasuta rakenduse: PocketBook Reader (iOS / Android)

    PC või Mac seadmes lugemiseks peate installima Adobe Digital Editionsi (Seeon tasuta rakendus spetsiaalselt e-raamatute lugemiseks. Seda ei tohi segamini ajada Adober Reader'iga, mis tõenäoliselt on juba teie arvutisse installeeritud )

    Seda e-raamatut ei saa lugeda Amazon Kindle's. 

The smart mobile has become an essential and inseparable part of our lives. This powerful tool enables us to perform multi-tasks in different modalities of voice, text, gesture, etc. The user plays an important role in the mode of operation, so multimodal interaction provides the user with new complex multiple modalities of interfacing with a system, such as speech, touch, type and more.



The book will discuss the new world of mobile multimodality, focusing on innovative technologies and design which create a state-of-the-art user interface. It will examine the practical challenges entailed in meeting commercial deployment goals, and offer new approaches to the designing such interfaces.



A multimodal interface for mobile devices requires the integration of several recognition technologies together with sophisticated user interface and distinct tools for input and output of data. The book will address the challenge of designing devices in a synergetic fashion which does not burden the user or to create a technological overload.
Preface V
List of contributing authors XVII
Nava Shaked
Ute Winter
1 Introduction to the evolution of Mobile Multimodality
1(18)
1.1 User Interfaces: Does vision meet, reality?
1(1)
1.2 Discussion of terms: Mobility and User Interface
2(4)
1.2.1 Mobility
2(1)
1.2.2 User Interface
3(2)
1.2.3 User-centered design
5(1)
1.2.4 Teamwork
5(1)
1.2.5 Context
6(1)
1.3 System interaction: Moving to Multimodality
6(2)
1.3.1 User input and system output
6(1)
1.3.2 Multimodality
7(1)
1.3.3 Combining modalities
7(1)
1.4 Mobile Multimodality: The evolution
8(6)
1.4.1 Technology compliance to user needs
8(1)
1.4.2 Technology readiness and availability
9(1)
1.4.3 The readiness of multimodal technology
10(1)
1.4.4 User requirements and needs
11(1)
1.4.5 Cycle of mutual influence
12(2)
1.5 Conclusion
14(5)
Deborah A. Dahl
2 Integrating natural language resources in mobile applications
19(26)
2.1 Natural language understanding and multimodal applications
19(2)
2.1.1 How natural language improves usability in multimodal applications
19(2)
2.1.2 How multimodality improves the usability of natural language interfaces
21(1)
2.2 Why natural language isn't ubiquitous already
21(1)
2.3 An overview of technologies related to natural language understanding
22(2)
2.4 Natural language processing tasks
24(12)
2.4.1 Accessing natural language technology: Cloud or client?
28(1)
2.4.2 Existing natural language systems
29(2)
2.4.3 Natural language processing systems
31(4)
2.4.4 Selection Criteria
35(1)
2.5 Standards
36(4)
2.5.1 EMMA
36(2)
2.5.2 MMI Architecture and Interfaces
38(2)
2.6 Future directions
40(1)
2.7 Summary
41(4)
Kathy Brown
Anmol Walla
Prashant Joshi
Abir Chakraborty
Sanjay Agarwal
3 Omnichannel Natural Language
45(18)
3.1 Introduction
45(1)
3.2 Multimodal interfaces built with omnichannel Natural Language Understanding
46(1)
3.3 Customer care and natural language
47(3)
3.4 Limitations of standard NLU solutions
50(3)
3.5 Omnichannel NL architecture
53(5)
3.5.1 Omni-NLU training algorithm
55(1)
3.5.2 Statistical-language model
56(1)
3.5.3 Input transformation
56(1)
3.5.4 Predictive omnichannel classifier
57(1)
3.5.5 Score normalization
57(1)
3.5.6 Conversation manager
57(1)
3.6 Experimental results
58(3)
3.6.1 Current analysis segment
58(3)
3.7 Summary
61(2)
Michael Lawo
Robert K. Logan
Edna Pasher
4 Wearable computing
63(30)
4.1 Introduction to Wearable Ecology
63(3)
4.2 Human-computer symbiosis
66(1)
4.3 Interactional considerations behind wearable technology
67(5)
4.4 Training of end users
72(1)
4.5 Wearable technology in the medical sector
73(2)
4.6 Human-centered design approach
75(1)
4.7 Context of wearable computing applications
76(2)
4.8 State of the art in context-aware wearable computing
78(2)
4.9 Project examples
80(3)
4.10 Towards the TZI Context Framework
83(2)
4.11 Conclusion
85(3)
4.12 Discussion and considerations for future research
88(5)
Ming Sun
Alexander I. Rudnicky
5 Spoken dialog systems adaptation for domains and for users
93(32)
5.1 Introduction
93(2)
5.2 Language adaptation
95(20)
5.2.1 Lexicon adaptation
96(10)
5.2.2 Adapting cloud ASR for domain and users
106(9)
5.2.3 Summary
115(1)
5.3 Intention adaptation
115(6)
5.3.1 Motivation
115(1)
5.3.2 Data collection
116(1)
5.3.3 Observation and statistics
116(2)
5.3.4 Intention recognition
118(1)
5.3.5 Personalized interaction
119(2)
5.3.6 Summary
121(1)
5.4 Conclusion
121(4)
Nava Shaked
Detlev Artelt
6 The use of multimodality in Avatars and Virtual Agents
125(20)
6.1 What are A&VA Definition and a short historical review
125(4)
6.1.1 First Avatars Bodily interfaces and organic machines
126(1)
6.1.2 Modern use of avatars
127(1)
6.1.3 From virtual "me" to virtual "you"
128(1)
6.2 A relationship framework for Avatars and Virtual Agents
129(4)
6.2.1 Type 1 The Avatar as virtual me
129(2)
6.2.2 Type 2 The interaction with a personalized/specialized avatar
131(1)
6.2.3 Type 3 Me and a virtual agent that is random
132(1)
6.3 Multimodal features of A&VA categorizing the need, the challenge, the solutions
133(7)
6.3.1 About multimodal interaction technologies
133(1)
6.3.2 Why use multimodality with Avatars?
134(3)
6.3.3 Evaluation of the quality of Avatars and Virtual Agents
137(3)
6.4 Conclusion and future directions: The vision of A&VA multimodality in the digital era
140(5)
Brion van Over
Elizabeth Molina-Markham
Sunny Lie
Donal Carbaugh
7 Managing interaction with an in-car infotainment system
145(24)
7.1 Introduction
145(3)
7.2 Theoretical framework and related literature
148(3)
7.3 Methodology
151(1)
7.4 Prompt timing and misalignment A formula for interruptions
152(5)
7.5 Interactional adaptation
157(5)
7.6 Norms and premises
162(2)
7.7 Implications for design
164(5)
Yael Shmueli-Friedland
Ido Zelman
Asaf Degani
Ron Asherov
8 Towards objective method in display design
169(26)
8.1 Introduction
169(7)
8.2 Method
176(6)
8.2.1 Listing of informational elements
176(1)
8.2.2 Domain expert rating
177(1)
8.2.3 Measurement of integrative interrelationships
178(1)
8.2.4 Clustering algorithm
179(2)
8.2.5 Comparison of the two hierarchical structures
181(1)
8.2.6 Comparisons between the domain expert and the design expert analyses
181(1)
8.3 Analysis of an instrument display
182(6)
8.4 Conclusion
188(7)
8.4.1 Extension of the approach to sound- and haptic-interfaces
189(2)
8.4.2 Multimodal presentation
191(4)
Michael Heymann
Asaf Degani
9 Classification and organization of information
195(24)
9.1 Introduction
195(4)
9.1.1 Head up displays
197(1)
9.1.2 Objectives
198(1)
9.2 Characterization of vehicle information
199(5)
9.2.1 Activity
199(1)
9.2.2 Information Type
200(1)
9.2.3 Urgency
200(1)
9.2.4 Timeliness
200(1)
9.2.5 Duration of interaction
201(1)
9.2.6 Importance
201(1)
9.2.7 Frequency of use
201(1)
9.2.8 Type of user response required
201(1)
9.2.9 Activation mode
201(3)
9.3 Allocation of information
204(1)
9.4 Head up display (HUD) and its information organization
205(4)
9.4.1 Information completeness and conciseness
207(2)
9.5 Principles of HUD information organization
209(1)
9.6 Review of existing Head Up Displays (HUDs)
210(4)
9.6.1 "Sporty" head up display
210(1)
9.6.2 Simplistic HUD
211(1)
9.6.3 Colorful head up display
212(1)
9.6.4 Graphically-rich head up display
213(1)
9.7 Conclusion
214(5)
Index 219
Nava A Shaked, Holon Institute of Technology, Holon, Israel; Ute Winter, General Motors, Israel.