TeleMorph & TeleTuras: Bandwidth determined Mobile
Download
Report
Transcript TeleMorph & TeleTuras: Bandwidth determined Mobile
TeleMorph & TeleTuras:
Bandwidth determined Mobile MultiModal
Presentation
Student: Anthony J. Solon
Supervisors: Prof. Paul Mc Kevitt
Kevin Curran
School of Computing and Intelligent Systems,
Faculty of Informatics,
University of Ulster, Magee.
Objectives of Research
To develop a system, TeleMorph, that dynamically morphs between output
modalities depending on available network bandwidth:
Wireless system’s output presentation (unimodal/multimodal) depending on the
network bandwidth available
Implement TeleTuras, a tourist information guide for the city of Derry
Receive and interpret questions from the user
Map questions to multimodal semantic representation
Match multimodal representation to database to retrieve answer
Map answers to multimodal semantic representation
Query bandwidth status
Generate multimodal presentation based on bandwidth data
Wireless Telecommunications
Generations of Mobile networks:
1G - Analog voice service with no data services
2G - Circuit-based, digital networks, capable of data transmission speeds
averaging around 9.6K bps
2.5G (GPRS) - Technology upgrades to 2G, boosting data transmission
speeds to around 56K bps. Allows packet based “always on” connectivity
3G (UMTS) - digital multimedia, different infrastructure required, data
transmission speeds from 144K-384K-2M bps
Positioning Systems:
GPS
DGPS
GLONASS
GSM
Mobile Intelligent MultiMedia
Systems
SmartKom
Mobile, Public, Home/office
Saarbrucken, Germany
Combines speech, gesture and facial expressions on input & output
Integrated trip planning, Internet access, communication applications,
personal organising
VoiceLog
BBN technologies in Cambridge, Massachusettes
Views/diagrams of military vehicles and direct connection to support
Damage identified & ordering of parts using diagrams
MUST
MUltimodal multilingual information Services for small mobile Terminals
EURESCOM, Heidelberg, Germany
Future multimodal and multilingual services on mobile networks
Intelligent MultiMedia Presentation
Flexibly generate various presentations to meet individual requirements of:
1) users, 2) situations, 3) domains
Fine-grained coordination of communication media and modalities
Key research problems:
Semantic Representation
Fusion, integration & coordination
Synchronisation
Semantic representation - represents semantics
Frame-based representations:
- CHAMELEON
- REA
XML-based representations:
- SmartKom
- MUST
Fusion, integration & coordination of modalities
Integrating different media in a consistent and coherent manner
Multimedia coordination leads to effective integrated multiple media in output
Synchronisation of modalities
Time threshold between modalities
E.g. Input - “What building is this?”, Output - “This is the Millenium forum”
Not synchronised => side effect is contradiction
Intelligent MultiMedia Presentation
Systems
Automatically generate coordinated intelligent multimedia presentations
User-determined presentation
COMET
COordinated Multimedia Explanation Testbed
Generates instructions for maintenance and repair of military radio receivertransmitters
Coordinates text and 3D graphics of mechanical devices
WIP
Intelligent multimedia authoring system
presents instructions for assembling/using/maintaining/repairing devices
(e.g. espresso machines, lawn mowers, modems)
IMPROVISE
Graphics generation system
constructive/parameterised graphics generation approaches
Uses an extensible formalism to represent a visual lexicon for graphics
generation
Intelligent MultiMedia Interfaces
& Agents
Intelligent multimedia interfaces
Parse integrated input and generate coordinated output
CUBRICON
Calspan-UB Research center Intelligent CONversationalist
Air Force Command and Control
Generates & recognises Speech; natural language text; displays graphics;
interprets gestures made with a pointing device
Intelligent multimedia agents
Embodied Conversational Agents
Natural human communication - speech, facial expressions, hand gestures,
& body stance
COLLAGEN
COLLaborative AGENt
object-oriented Java middleware for building collaborative interface
agents
MIT Media Laboratory work on embodied conversational agents
Project Proposal
Research and implement a mobile intelligent multimedia presentation system
called TeleMorph
Dynamically generates a multimedia presentation determined by the bandwidth
available
TeleTuras tourist navigation aid providing testbed for TeleMorph incorporating:
route planning, maps, spoken presentations, graphics of points of interest
and animations
Output modalities used
Effectiveness of communication
TeleTuras examples:
“Where is the Millenium forum?”
“Take me to the GuildHall”
“What buildings are of interest in this area?”
“Is there a Chinese restaurant in this area?”
Architecture of TeleMorph
Comparison of Intelligent
MultiMedia Systems
Comparison of Mobile Intelligent
MultiMedia Systems
Prospective Tools
Development language - J2ME (Java 2 Micro Edition)
Speech input/output - Java Speech API – IBM’s implementation of JSAPI
“speech for Java”
US & UK English, French, German, Italian, Spanish, and Japanese
Java Speech API Markup Language (JSML)
Java Speech API Grammar Format (JSGF)
Positioning system - GPS (Global Positioning System) provides the accurate
location information necessary for a LBS (Location Based Service)
Graphics input/output - The User Interface (UI) defined in J2ME is logically
composed of two sets of APIs:
Low-level UI API
High-level UI API
Project Schedule
Conclusion
A Mobile Intelligent MultiModal presentation System called TeleMorph will be
developed
Dynamically morphing between output modalities depending on available
network bandwidth
Bandwidth and Device determined Mobile MultiModal presentation
TeleTuras will be used as a testbed for TeleMorph
Corpora of questions to test TeleTuras (prospective users/tourists)