US9813531B2 - System and method for screen orientation in a rich media environment - Google Patents

System and method for screen orientation in a rich media environment Download PDF

Info

Publication number
US9813531B2
US9813531B2 US12/018,119 US1811908A US9813531B2 US 9813531 B2 US9813531 B2 US 9813531B2 US 1811908 A US1811908 A US 1811908A US 9813531 B2 US9813531 B2 US 9813531B2
Authority
US
United States
Prior art keywords
screen
orientation
rich
angle
change event
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US12/018,119
Other versions
US20080209442A1 (en
Inventor
Vidya Setlur
Suresh Chitturi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sisvel International SA
Original Assignee
Sisvel International SA
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sisvel International SA filed Critical Sisvel International SA
Priority to US12/018,119 priority Critical patent/US9813531B2/en
Assigned to NOKIA CORPORATION reassignment NOKIA CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHITTURI, SURESH, SETLUR, VIDYA
Publication of US20080209442A1 publication Critical patent/US20080209442A1/en
Assigned to SISVEL INTERNATIONAL S.A. reassignment SISVEL INTERNATIONAL S.A. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: NOKIA CORPORATION
Application granted granted Critical
Publication of US9813531B2 publication Critical patent/US9813531B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/02Constructional features of telephone sets
    • H04M1/0202Portable telephone sets, e.g. cordless phones, mobile phones or bar type handsets
    • H04M1/0206Portable telephones comprising a plurality of mechanically joined movable body parts, e.g. hinged housings
    • H04M1/0208Portable telephones comprising a plurality of mechanically joined movable body parts, e.g. hinged housings characterized by the relative motions of the body parts
    • H04M1/0214Foldable telephones, i.e. with body parts pivoting to an open position around an axis parallel to the plane they define in closed position
    • H04M1/0222Foldable in two directions, i.e. using a two degree of freedom hinge
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/54Interprogram communication
    • G06F9/542Event management; Broadcasting; Multicasting; Notifications
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/04Indexing scheme for image data processing or generation, in general involving 3D image data

Definitions

  • the present invention relates generally to the field of wireless devices and, more particularly, presentation and display of content on such devices.
  • Rich media content generally includes content that is graphically rich and contains compound (or multiple) media including graphics, text, video and audio and, preferably, delivered through a single interface. Rich media dynamically changes over time and could respond to user interaction.
  • Rich media applications particularly in the Web services domain, include XML-based content such as Scalable Vector Graphics (SVG), Synchronized Multimedia Integration Language (SMIL) and Compound Documents Format (CDF).
  • SVG Scalable Vector Graphics
  • SMIL Synchronized Multimedia Integration Language
  • CDF Compound Documents Format
  • SVG (more specifically, version SVGT 1.2) is a language for describing two-dimensional graphics in XML.
  • SVG allows for three types of graphic objects: vector graphic shapes (e.g., paths consisting of straight lines and curves), multimedia (e.g., raster images, video), and text.
  • SVG drawings can be interactive (e.g., using a Document Object Model, or DOM, event model) and dynamic. Animations can be defined and triggered either declaratively (e.g., by embedding SVG animation elements in SVG content) or via scripting.
  • Sophisticated applications of SVG are possible by use of a supplemental scripting language which accesses the SVG Micro Document Object Model (uDOM), which provides complete access to all elements, attributes and properties.
  • uDOM SVG Micro Document Object Model
  • a rich set of event handlers can be assigned to any SVG graphical object. Because of its compatibility and leveraging of other Web standards (such as CDF, described below), features like scripting can be
  • SMIL (specifically, SMIL 2.0) enables simple authoring of interactive audiovisual presentations.
  • SMIL is typically used for rich media/multimedia presentations which integrate streaming audio and video with images, text or any other media type.
  • the CDF working group is producing recommendations on combining separate component languages (e.g., XML-based languages, elements and attributes from separate vocabularies), like XHTML, SVG, MathML, and SMIL, with a focus on user interface markups.
  • separate component languages e.g., XML-based languages, elements and attributes from separate
  • a growing number of mobile devices e.g., Nokia N93, N95
  • presentation of rich media content on a client device is scalable and interactive
  • challenges still exist in presentation when conforming to changes in screen orientation.
  • built-in advanced sensors in the phone such as an accelerometer, detect when the phone's screen is rotated and, in turn, generate events that notify the top-layer application framework.
  • This notification enables the application to respond to the orientation events to perform certain actions such as changing the appearance of the application (e.g., the application grid on the phone screen) or the content within the application by dispatching this event to the engine that is responsible for rendering or displaying the content.
  • these orientation events may be explicitly streamed from a remote terminal to the client and executed with or without event triggers.
  • the underlying engine for e.g. rich media engine
  • Event Identifier Namespace Description “screenOrientation0” urn:mpeg:mpeg4:laser:2005
  • the screen orientation has changed to typical ‘landscape’ orientation
  • screenOrientation90 urn:mpeg:mpeg4:laser:2005
  • the screen orientation has changed to typical ‘portrait’ orientation
  • screenOrientation180 urn:mpeg:mpeg4:laser:2005
  • the screen orientation has changed to inverted ‘landscape’ orientation
  • the screen orientation has changed to inverted ‘portrait’ orientation
  • the event identifiers are based on angles and, therefore, introduce ambiguity while mapping the underlying hardware/OS based events to these content based events.
  • the Symbian UI framework does not support inverted portrait and landscape modes.
  • the proposal does not define the mapping of the event identifiers to an event interface.
  • the proposal lacks the definition of event interface, which is crucial for obtaining information related to the screen orientation, such as screen width/height, and the location of the soft keys in response to the change in screen orientation.
  • U.S. Patent Publication No. 2002/0101439 titled “Method and Apparatus for Rotating an Image on a Display,” utilizes a three-dimensional rendering engine to rotate an image based on a user-selected or otherwise determined screen orientation.
  • a vertex coordinate transformation is defined for a rotated destination image.
  • the source image is used as a texture for texture mapping during rendering operation to produce rotated image.
  • this disclosure mainly concerns image manipulation via rotation and texture mapping to align with the desired screen orientation.
  • U.S. Pat. No. 6,597,384, titled, “Automatic Reorienting of Screen Orientation Using Touch Sensitive System,” discloses the automatic change of display by sensing a change in the direction of the display via touch sensors. This disclosure focuses primarily on the mapping of hardware events triggered by touch sensors to the actual orientation of the screen. It fails to address any application level functionality for handling such orientation and is restricted mainly to touch screens.
  • U.S. Pat. No. 5,910,797 titled, “Portable Data Processing Apparatus Provided with a Screen and a Gravitation-Controlled Sensor for Screen Orientation,” discloses a gravitation-controlled sensor in the screen for measuring a spatial orientation.
  • the disclosed apparatus has a programmed data processor under control of a predetermined range of spatial orientations imparting a non-stationary motion pattern to a predetermined selection among the objects.
  • the motion can be used in the way of a joystick.
  • This disclosure is related to hardware mechanisms for enabling screen orientation, rather than any application-level functionality, as needed for rich media content.
  • U.S. Pat. No. 5,661,632 titled, “Hand Held Computer with Dual Display Screen Orientation Capability Controlled by Toggle Switches Having First and Second Non-Momentary Positions,” discloses an architecture of a handheld computer having a generally rectangular housing on a front side wall of which a display screen is operatively mounted. A row of toggle switches is also mounted on the front housing side wall, adjacent the display screen, the toggle switches being operatively connected to the computer circuitry within the housing. One of the toggle switches is operative, via the internal computer circuitry, to selectively rotate, through an angle of 90 degrees, the orientation of data generated on the screen so that the screen data is in an upright viewing orientation relative to the user of the computer. Because the toggle switches are non-momentary, the orientation of the data generated on the screen is preserved while the handheld computer is turned off. Again, this disclosure concerns hardware rather than application level mechanisms for screen orientation.
  • One aspect of the invention relates to a method of displaying content on a device.
  • the method comprises detecting a screen orientation change event and triggering a screen orientation adjustment in a rich media engine.
  • the method further comprises determining if one or more applications are configured to accommodate a screen orientation change.
  • the screen orientation change event is a hardware change associated with the device.
  • the hardware change may be detected via sensors in the device.
  • the screen orientation change event is an input by the user.
  • the screen orientation change event is a signal received from a remote device.
  • the rich media engine includes a screenOrientation attribute.
  • the screenOrientation attribute may have a value selected from portrait, landscape or angle.
  • the rich media engine may further include a view attribute applicable to a subset of an application to be oriented.
  • the screenOrientation event or attribute may include attributes of screenwidth, screenHeight and screenAngle
  • the screen orientation adjustment repositions one or more soft keys.
  • a wireless device comprises a detection module adapted to detect a screen orientation change event and a screen orientation module adapted to trigger a screen orientation adjustment in a rich media engine.
  • a rich media interface comprises a ScreenOrientationEvent attribute adapted to change screen orientation of a screen of a device upon a triggering event and one or more screen dimension attributes defining the dimensions of the screen after changing orientation.
  • FIG. 1 is a picture illustrating an exemplary wireless device adapted to display content in portrait and landscape formats according to an embodiment of the present invention
  • FIG. 2 illustrates the display of content on a screen in a portrait format according to an embodiment of the present invention
  • FIG. 3 illustrates the display of content on a screen in a landscape format according to an embodiment of the present invention
  • FIGS. 4( a )-( d ) illustrate the display of content on a screen in various formats according to an embodiment of the present invention.
  • FIG. 5 is a flowchart illustrating a method of displaying content according to an embodiment of the present invention.
  • Embodiments of the present invention address the need for dynamic and interactive multimedia scenes (DIMS) applications to adapt to changes in screen orientation.
  • Embodiments of the present invention provide mechanisms that are both syntactic and functional to address the screen orientation problem for rich media content. Such mechanisms enable implicit (e.g., when the phone screen is rotated between landscape and portrait) and explicit (e.g., when the content itself includes a screen orientation event listener to a button that is triggered when clicked) adaptation of content to the required orientation modes.
  • embodiments of the invention provide (1) application based orientation modes for rich media content, (2) an event interface for these orientation modes, and (3) event management and processing of screen orientation events.
  • interactive mobile TV services are understood as the ability to provide a deterministic rendering and behavior of rich-media content, including audio-video content, text, images, XML-based content such as SVG, and TV and radio channels, together in the end-user interface.
  • the service must provide convenient navigation through the content in a single application or service and must allow synchronized interaction, such as voting and personalization (e.g., related menu or sub-menu, advertising and content in function of the end-user profile or service subscription).
  • voting and personalization e.g., related menu or sub-menu, advertising and content in function of the end-user profile or service subscription.
  • default applications run in portrait mode on hand-held devices. However, it may be optimal to view mobile TV in landscape mode for various reasons (e.g., displaying closed captioning, video content itself is in landscape mode, etc.).
  • a live enterprise data feed may include stock tickers that provide streaming of real-time quotes, live intra-day charts with technical indicators, news monitoring, weather alerts, charts, business updates, etc. Given how such data could dynamically vary in terms of what is shown and how much content is shown, the content may often toggle between several orientations.
  • certain rich media games allow for multimedia content to be presented with some game logic for entertainment purposes. Either based in user interaction or change in game logic, the orientation of the content may change.
  • the following may be used for generating screen-orientation events: (1) via device hardware events (e.g., rotation of the phone screen) and (2) explicit instructions, for example, sent from a remote terminal/server that would be triggered when certain events occur on the client.
  • the application could completely rely on the underlying application shell to take care of rotating the content based on the orientation of the screen.
  • automatic adaptation of the content by simple rotation is not always optimal.
  • Certain adjustments to the content presentation are often preferred for better viewing experience. For example, in FIG. 1 , the application grid on the N93 screen is readjusted from portrait to landscape to make efficient use of the screen.
  • the second case of adjusting orientation based on events would need explicit definition in the data as well.
  • an attribute called “screenOrientation” in the rich media content is provided.
  • this attribute takes one of three possible forms: specification of the portrait format, specification of the landscape format, or an angle in degrees. While the first two values correspond to existing modes on most devices, the third option can be user defined.
  • a “view” element for defining the scope of the screen orientation in rich media content is provided. While the “screenOrientation” attribute specifies the mode in which the rich media content has to be rendered, there still needs to be a way of specifying how much of the content has to be rendered in that mode.
  • a “view” is a subset of the rich media content that can be rendered and interacted within the application.
  • an application has at least one logical view. If the entire application is to rendered in a particular orientation, then views are not required. However, if a subset of the application is to be rendered in one mode then views are useful in differentiating the scope of the “screenOrientation” attribute applied.
  • a particular view can be associated with a screenOrientation event that would be triggered upon a certain condition being satisfied.
  • event types may include “screenOrientationPortrait,” “screenOrientationLandscape” and “screenOrientationManual” corresponding to the screenOrientation attribute values “portrait”, “landscape” and “orientation in degrees,” respectively.
  • the first two events i.e., portrait and landscape are the most commonly used orientations in current devices.
  • a screen orientation event interface for the rich media system.
  • the following is the definition of the screenOrientation interface that provides the contextual information related to screen orientation changes.
  • This interface is designed and based on DOM Level 3 events specification, and therefore can be used in any DOM based implementations. Note that this interface extends the base “event” interface, which contains the basic event information such as the event target, and event type information.
  • SOFTKEYS_BOTTOM indicates that the device soft keys at the bottom of the screen in the current screen orientation, as illustrated in FIG. 4 a.
  • SOFTKEYS_TOP indicates that the device soft keys are at the top of the screen in the current screen orientation, as illustrated in FIG. 4 b.
  • SOFTKEYS_LEFT indicates that the device soft keys are to the left of the screen in the current screen orientation, as illustrated in FIG. 4 c.
  • “SOFTKEYS_RIGHT” indicates that the device soft keys are to the right of the screen in the current screen orientation, as illustrated in FIG. 4 d.
  • “ScreenWidth” contains the new screen display or viewport width reflecting the new orientation. “ScreenHeight” contains the new screen display or viewport height reflecting the new orientation. “SoftKeysLayout” indicates the location of the device soft keys in response to the orientation change. The possible values are SOFTKEYS_LEFT, SOFTKEYS_RIGHT, SOFTKEYS_TOP, and SOFTKEYS_BOTTOM, as described above with reference to FIGS. 4( a )-( d ) .
  • the user agent or underlying rich media engine can adapt to the layout or visual representation of the content to current screen orientation.
  • the event management and processing of screen orientation events is provided.
  • the event management and processing of screen orientation events is based on the DOM Level 3 events processing model. This includes event creation, event dispatching, event registration and event handling.
  • Event creation deals with the creation of a ScreenOrientationEvent object which is an implementation of ScreenOrientationEvent interface. This event is directly mapped to the hardware event notification (e.g., using the Symbian UI framework) or through an event “message” from the server. During the event creation it is important that the relevant event attributes such as event target, event type, screenWidth, screenHeight attributes are set.
  • Event dispatching deals with dispatching the ScreenOrientation event to the appropriate target based on the event target.
  • the content author or application typically registers for the events if he/she chooses to listen to a particular event. For example, to listen to a ScreenOrientationEvent, a listener must be registered with the event target within the rich media engine. This can be done either declaratively using the ⁇ ev:listener> element or can be done by the implementation by attaching the “EventListener” object to the “EventTarget” object. This completes the registration process and is usually done prior to the event creation and dispatching phase.
  • Event handling relates to event processing or realization.
  • Event handlers contain the logic or code that must be executed based on a specific event listener and an event target. For example, when the ScreenOrientationEvent is dispatched to a particular event target as described in the previous stages, the event manager of the rich media engine checks the list of event listeners and invokes the appropriate event handler, thereby executing the logic that is present within the handler. And the handler can access all the information relevant to screen orientation as defined by the ScreenOrientationEvent object.
  • the following example illustrates the concept of screen orientation and the content adoption based on the orientation change.
  • the content is authored in SVG format.
  • the content is seen in the portrait mode (default), as illustrated in FIG. 2 .
  • the screen orientation change event is fired, either by the hardware or streamed from the server, this event gets dispatched to the rich media engine, where the ⁇ listener> elements get the notification.
  • the listener elements invoke the appropriate ⁇ handler> elements that, in this example, execute the associated scripts to change the layout of the content from ‘portrait’ mode to ‘landscape’ mode, as illustrated in FIG. 3 .
  • hardware-based orientation event triggers are mapped to software-based rich media screen events. While screenOrientation events can be triggered purely from the application software side (e.g., via a button click, menu selection, after a certain time duration, etc.), these events could also be triggered from a change in orientation of the phone screen. In order to address the latter scenario, the mapping from the underlying phone hardware event triggers to the rich media application based screenOrientation events needs to be specified. In order to explain the mapping, we take the example of Symbian, particularly the S60 application framework. Though the S60 application framework is not the only framework used on mobile phones, it is representative of a larger market of such adaptable screen devices.
  • the application will be displayed in the selected screen mode.
  • the application will no longer respond to hardware or software events which would normally affect screen mode. It will be permanently fixed to the selected screen mode.
  • the list of available screen modes are specific to that particular device. If an application wants a fixed orientation, it should use the scalable APIs in CEikAppUi. Particularly, a new enumeration and two APIs have been added to CEikAppUi:
  • class CSimpleAppUi public CQikAppUi ⁇ public: void ConstructL ( ); ⁇ CSimpleAppUi ( ); private: // Inherited from class CEikAppUi void HandleCommandL(TInt aCommand); void HandleScreenDeviceChangedL( ); . . .
  • HandleScreenDeviceChangedL( ) will be called by the system when the screen mode changes. To check which mode it is in (Flip open or Flip closed), the following code is inserted in the implementation of HandleScreendeviceChangedL( ):
  • the process flow is illustrated in FIG. 5 .
  • the embedded OS framework regularly or continuously polls for hardware orientation changes (block 510 ).
  • the OS framework determines whether the phone orientation has changed. As noted above, this may be achieved through use of sensors in the hardware. If no changes are detected, the OS framework continues to poll for hardware changes. On the other hand, if a change is detected, the process proceeds to block 530 and notifies the application framework.
  • the application framework is responsible for changing orientation
  • the process proceeds to block 550 , and the orientation change is applied to the rich media application. If the application framework is not responsible for changing orientation, the process proceeds to block 560 , and the rich media engine is notified of the orientation change.
  • the screenOrientation event is triggered and the necessary changes are applied to the content orientation.
  • the attribute values for screenOrientation can be extended for future orientation modes.
  • the screenOrientation event types can also be extended for future orientation modes.
  • the mapping of device hardware event triggers to rich media screenOrientation events are agnostic of the underlying embedded operating system.
  • a computer-readable medium may include removable and non-removable storage devices including, but not limited to, Read Only Memory (ROM), Random Access Memory (RAM), compact discs (CDs), digital versatile discs (DVD), etc.
  • program modules may include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types.
  • Computer-executable instructions, associated data structures, and program modules represent examples of program code for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represents examples of corresponding acts for implementing the functions described in such steps or processes.

Abstract

A method of displaying content on a device comprises detecting a screen orientation change event, determining if one or more applications are configured to accommodate a screen orientation change, and triggering a screen orientation adjustment in a rich media engine. The screen orientation change event may be a hardware change associated with the device. The hardware change may be detected via sensors in the device. The screen orientation change event may be an input by the user or a signal received from a remote device.

Description

CROSS-REFERENCE TO RELATED PATENT APPLICATIONS
This application claims priority under 35 U.S.C. §119(e) from U.S. Provisional Application Ser. No. 60/886,094, filed Jan. 22, 2007, incorporated herein by reference in its entirety.
BACKGROUND OF THE INVENTION
The present invention relates generally to the field of wireless devices and, more particularly, presentation and display of content on such devices.
Until recently, applications for mobile devices were text based with limited interactivity. However, as more wireless devices are equipped with color displays and more advanced graphics-rendering libraries, consumers will demand a rich media experience from all their wireless applications. Rich media content generally includes content that is graphically rich and contains compound (or multiple) media including graphics, text, video and audio and, preferably, delivered through a single interface. Rich media dynamically changes over time and could respond to user interaction. Rich media applications, particularly in the Web services domain, include XML-based content such as Scalable Vector Graphics (SVG), Synchronized Multimedia Integration Language (SMIL) and Compound Documents Format (CDF).
SVG (more specifically, version SVGT 1.2) is a language for describing two-dimensional graphics in XML. SVG allows for three types of graphic objects: vector graphic shapes (e.g., paths consisting of straight lines and curves), multimedia (e.g., raster images, video), and text. SVG drawings can be interactive (e.g., using a Document Object Model, or DOM, event model) and dynamic. Animations can be defined and triggered either declaratively (e.g., by embedding SVG animation elements in SVG content) or via scripting. Sophisticated applications of SVG are possible by use of a supplemental scripting language which accesses the SVG Micro Document Object Model (uDOM), which provides complete access to all elements, attributes and properties. A rich set of event handlers can be assigned to any SVG graphical object. Because of its compatibility and leveraging of other Web standards (such as CDF, described below), features like scripting can be done on XHTML and SVG elements simultaneously within the same Web page.
SMIL (specifically, SMIL 2.0) enables simple authoring of interactive audiovisual presentations. SMIL is typically used for rich media/multimedia presentations which integrate streaming audio and video with images, text or any other media type.
The CDF working group is producing recommendations on combining separate component languages (e.g., XML-based languages, elements and attributes from separate vocabularies), like XHTML, SVG, MathML, and SMIL, with a focus on user interface markups. When combining user interface markups, specific problems have to be resolved that are not addressed by the individual markups specifications, such as the propagation of events across markups, the combination of rendering or the user interaction model with a combined document. The CDF working group will address these types of problems.
In order to maximize the user experience while viewing rich media content, efficient delivery of such content of the client is important. However, optimal content presentation is also vital and is more challenging when viewing orientations have to be changed, either explicitly (e.g., declared in the content itself) or implicitly (e.g., via a change in the actual phone screen mode).
A growing number of mobile devices (e.g., Nokia N93, N95) have greater versatility in their form factors, particularly screen orientation and input modalities. While the presentation of rich media content on a client device is scalable and interactive, challenges still exist in presentation when conforming to changes in screen orientation. There is hence a need for the content or the layout of the content displayed on the screen to adapt to these orientation modes for delivering a good user experience.
Typically, built-in advanced sensors in the phone such as an accelerometer, detect when the phone's screen is rotated and, in turn, generate events that notify the top-layer application framework. This notification enables the application to respond to the orientation events to perform certain actions such as changing the appearance of the application (e.g., the application grid on the phone screen) or the content within the application by dispatching this event to the engine that is responsible for rendering or displaying the content. As an alternative, these orientation events may be explicitly streamed from a remote terminal to the client and executed with or without event triggers. However, in order for the underlying engine (for e.g. rich media engine) to understand these application events, it is required to provide mechanisms for enabling proper response to screen orientation events including:
    • a syntactic definition of the screen orientation modes in the rich media content;
    • a functional and syntactic definition of a rich media based event interface specific to these screen orientation modes; and
    • a methodology on how phone hardware based screen orientation events map to the application based event interface, the subsequent processing of these events and actual content realization by the rich media client engine.
Although the concept of different screen orientations has been prevalent for various types of user interface (UI) systems and devices, currently there is no systematic mechanism that defines the handling of screen orientation modes in the context of a rich media environment.
A preliminary contribution was submitted at a Third Generation Partnership Project (3GPP) SA4#41 meeting (ref: ftp://ftp.3gpp.org/TSG_SA/WG4_CODEC/TSGS4_41/Docs/S4-060640.zip) to address this problem. The proposal introduced four events based on angles:
Event Identifier Namespace Description
“screenOrientation0” urn:mpeg:mpeg4:laser:2005 The screen orientation has
changed to typical ‘landscape’
orientation
“screenOrientation90” urn:mpeg:mpeg4:laser:2005 The screen orientation has
changed to typical ‘portrait’
orientation
“screenOrientation180” urn:mpeg:mpeg4:laser:2005 The screen orientation has
changed to inverted ‘landscape’
orientation
“screenOrientation270” urn:mpeg:mpeg4:laser:2005 The screen orientation has
changed to inverted ‘portrait’
orientation
However, with this proposal, the event identifiers are based on angles and, therefore, introduce ambiguity while mapping the underlying hardware/OS based events to these content based events. For example, the Symbian UI framework does not support inverted portrait and landscape modes. Further, the proposal does not define the mapping of the event identifiers to an event interface. In addition, the proposal lacks the definition of event interface, which is crucial for obtaining information related to the screen orientation, such as screen width/height, and the location of the soft keys in response to the change in screen orientation.
Further, U.S. Patent Publication No. 2002/0101439, titled “Method and Apparatus for Rotating an Image on a Display,” utilizes a three-dimensional rendering engine to rotate an image based on a user-selected or otherwise determined screen orientation. A vertex coordinate transformation is defined for a rotated destination image. The source image is used as a texture for texture mapping during rendering operation to produce rotated image. However, this disclosure mainly concerns image manipulation via rotation and texture mapping to align with the desired screen orientation.
U.S. Pat. No. 6,597,384, titled, “Automatic Reorienting of Screen Orientation Using Touch Sensitive System,” discloses the automatic change of display by sensing a change in the direction of the display via touch sensors. This disclosure focuses primarily on the mapping of hardware events triggered by touch sensors to the actual orientation of the screen. It fails to address any application level functionality for handling such orientation and is restricted mainly to touch screens.
U.S. Pat. No. 5,910,797, titled, “Portable Data Processing Apparatus Provided with a Screen and a Gravitation-Controlled Sensor for Screen Orientation,” discloses a gravitation-controlled sensor in the screen for measuring a spatial orientation. The disclosed apparatus has a programmed data processor under control of a predetermined range of spatial orientations imparting a non-stationary motion pattern to a predetermined selection among the objects. The motion can be used in the way of a joystick. This disclosure is related to hardware mechanisms for enabling screen orientation, rather than any application-level functionality, as needed for rich media content.
U.S. Pat. No. 5,661,632, titled, “Hand Held Computer with Dual Display Screen Orientation Capability Controlled by Toggle Switches Having First and Second Non-Momentary Positions,” discloses an architecture of a handheld computer having a generally rectangular housing on a front side wall of which a display screen is operatively mounted. A row of toggle switches is also mounted on the front housing side wall, adjacent the display screen, the toggle switches being operatively connected to the computer circuitry within the housing. One of the toggle switches is operative, via the internal computer circuitry, to selectively rotate, through an angle of 90 degrees, the orientation of data generated on the screen so that the screen data is in an upright viewing orientation relative to the user of the computer. Because the toggle switches are non-momentary, the orientation of the data generated on the screen is preserved while the handheld computer is turned off. Again, this disclosure concerns hardware rather than application level mechanisms for screen orientation.
SUMMARY OF THE INVENTION
One aspect of the invention relates to a method of displaying content on a device. The method comprises detecting a screen orientation change event and triggering a screen orientation adjustment in a rich media engine.
In one embodiment, the method further comprises determining if one or more applications are configured to accommodate a screen orientation change.
In one embodiment, the screen orientation change event is a hardware change associated with the device. The hardware change may be detected via sensors in the device.
In one embodiment, the screen orientation change event is an input by the user.
In another embodiment, the screen orientation change event is a signal received from a remote device.
In one embodiment, the rich media engine includes a screenOrientation attribute. The screenOrientation attribute may have a value selected from portrait, landscape or angle. The rich media engine may further include a view attribute applicable to a subset of an application to be oriented. The screenOrientation event or attribute may include attributes of screenwidth, screenHeight and screenAngle
In one embodiment, the screen orientation adjustment repositions one or more soft keys.
In another aspect of the invention, a computer program product, embodied on a computer-readable medium, for displaying content on a device comprises computer code for detecting a screen orientation change event, and computer code for triggering a screen orientation adjustment in a rich media engine.
In another aspect of the invention, a wireless device comprises a detection module adapted to detect a screen orientation change event and a screen orientation module adapted to trigger a screen orientation adjustment in a rich media engine.
In another aspect of the invention, a rich media interface comprises a ScreenOrientationEvent attribute adapted to change screen orientation of a screen of a device upon a triggering event and one or more screen dimension attributes defining the dimensions of the screen after changing orientation.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a picture illustrating an exemplary wireless device adapted to display content in portrait and landscape formats according to an embodiment of the present invention;
FIG. 2 illustrates the display of content on a screen in a portrait format according to an embodiment of the present invention;
FIG. 3 illustrates the display of content on a screen in a landscape format according to an embodiment of the present invention;
FIGS. 4(a)-(d) illustrate the display of content on a screen in various formats according to an embodiment of the present invention; and
FIG. 5 is a flowchart illustrating a method of displaying content according to an embodiment of the present invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
Embodiments of the present invention address the need for dynamic and interactive multimedia scenes (DIMS) applications to adapt to changes in screen orientation. Embodiments of the present invention provide mechanisms that are both syntactic and functional to address the screen orientation problem for rich media content. Such mechanisms enable implicit (e.g., when the phone screen is rotated between landscape and portrait) and explicit (e.g., when the content itself includes a screen orientation event listener to a button that is triggered when clicked) adaptation of content to the required orientation modes. In particular, embodiments of the invention provide (1) application based orientation modes for rich media content, (2) an event interface for these orientation modes, and (3) event management and processing of screen orientation events.
There are several use cases where such screen orientation mechanisms described herein would be useful in rich media services. For example, interactive mobile TV services are understood as the ability to provide a deterministic rendering and behavior of rich-media content, including audio-video content, text, images, XML-based content such as SVG, and TV and radio channels, together in the end-user interface. The service must provide convenient navigation through the content in a single application or service and must allow synchronized interaction, such as voting and personalization (e.g., related menu or sub-menu, advertising and content in function of the end-user profile or service subscription). Typically, default applications run in portrait mode on hand-held devices. However, it may be optimal to view mobile TV in landscape mode for various reasons (e.g., displaying closed captioning, video content itself is in landscape mode, etc.).
Further, a live enterprise data feed may include stock tickers that provide streaming of real-time quotes, live intra-day charts with technical indicators, news monitoring, weather alerts, charts, business updates, etc. Given how such data could dynamically vary in terms of what is shown and how much content is shown, the content may often toggle between several orientations.
Further, certain rich media games allow for multimedia content to be presented with some game logic for entertainment purposes. Either based in user interaction or change in game logic, the orientation of the content may change.
In accordance with embodiments of the present invention, the following may be used for generating screen-orientation events: (1) via device hardware events (e.g., rotation of the phone screen) and (2) explicit instructions, for example, sent from a remote terminal/server that would be triggered when certain events occur on the client. In the first case, the application could completely rely on the underlying application shell to take care of rotating the content based on the orientation of the screen. However, automatic adaptation of the content by simple rotation is not always optimal. Certain adjustments to the content presentation are often preferred for better viewing experience. For example, in FIG. 1, the application grid on the N93 screen is readjusted from portrait to landscape to make efficient use of the screen. These adjustments of how the content is to be adapted have to be specified in the data itself. In addition, the second case of adjusting orientation based on events (e.g., a button click, after a duration of a certain period) would need explicit definition in the data as well. Thus, in accordance with an embodiment of the invention, an attribute called “screenOrientation” in the rich media content is provided. In one embodiment, this attribute takes one of three possible forms: specification of the portrait format, specification of the landscape format, or an angle in degrees. While the first two values correspond to existing modes on most devices, the third option can be user defined.
Attribute Value Namespace Description
“screenOrientation” Portrait dims:events Specified when the content needs to be
presented in “portrait” mode.
Landscape dims:events Specified when the content needs to be
presented in “landscape” mode.
Angle of dims:events Specified when the content needs to be
orientation presented in an orientation that is
in degrees inclined to the vertical viewing axis by
the specified angle.
In another embodiment, a “view” element for defining the scope of the screen orientation in rich media content is provided. While the “screenOrientation” attribute specifies the mode in which the rich media content has to be rendered, there still needs to be a way of specifying how much of the content has to be rendered in that mode.
For example:
<svg viewBox = “0 0 1000 1000” version =
“1.1”>
 <circle id = “s1” cx = “300” cy =
“300” r = “200” fill = “mediumorchid”
stroke = “black” stroke-width = “3”/>
 <ellipse id = “s2” cx = “350” cy =
“800” rx = “200” ry = “150”
fill=“lightskyblue” stroke = “black”
stroke-width = “3”/>
 <circle id = “s3” cx = “770” cy =
“750” r = “200” fill = “tomato” stroke =
“black” stroke-width = “3”/>
 <ellipse id = “s4” cx = “430” cy =
“600” rx = “200” ry = “150” transform =
“rotate(−45,100,100)” fill =
“mediumseagreen” stroke = “black”
stroke-width = ”3”/>
</svg>
Here, if the first circle and the first ellipse are to be rendered in landscape mode while the second circle and second ellipse are to be rendered normally in the default presentation mode, one would have to use the “screenOrientation” mode for every element that needs to be rendered in that mode. So in the above example, screenOrientation=“landscape” would have to be added to the first circle and first ellipse. While this may still be considered fine with just two elements, it is not optimal to add the same mode for large sets of elements. Thus, to address this issue, we introduce the concept of “views” in this invention.
A “view” is a subset of the rich media content that can be rendered and interacted within the application. By default, an application has at least one logical view. If the entire application is to rendered in a particular orientation, then views are not required. However, if a subset of the application is to be rendered in one mode then views are useful in differentiating the scope of the “screenOrientation” attribute applied.
Again, considering the above example, the svg content with views would be:
<svg viewBox = “0 0 1000 1000” version = “1.1”>
<view id = “1 ” screenOrientation =
“landscape”>
 <circle id = “s1” cx = “300” cy = “300” r = “200”
fill = “mediumorchid” stroke = “black” stroke-width =
“3”/>
 <ellipse id = “s2” cx = “350” cy = “800” rx = “200”
ry = “150” fill=“lightskyblue” stroke = “black” stroke-
width = “3”/>
</view>
 <circle id = “s3” cx = “770” cy = “750” r = “200”
fill = “tomato” stroke = “black” stroke-width = “3”/>
 <ellipse id = “s4” cx = “430” cy = “600” rx = “200”
ry = “150” transform = “rotate(−45,100,100)” fill =
“mediumseagreen” stroke = “black” stroke-width =
”3”/>
</svg>
Further, a particular view can be associated with a screenOrientation event that would be triggered upon a certain condition being satisfied.
In one embodiment of the invention, one or more event types are provided. For example, event types may include “screenOrientationPortrait,” “screenOrientationLandscape” and “screenOrientationManual” corresponding to the screenOrientation attribute values “portrait”, “landscape” and “orientation in degrees,” respectively. The first two events (i.e., portrait and landscape) are the most commonly used orientations in current devices.
In another embodiment, a screen orientation event interface for the rich media system is provided. The following is the definition of the screenOrientation interface that provides the contextual information related to screen orientation changes. This interface is designed and based on DOM Level 3 events specification, and therefore can be used in any DOM based implementations. Note that this interface extends the base “event” interface, which contains the basic event information such as the event target, and event type information.
interface ScreenOrientationEvent : Event
{
const unsigned short SOFTKEYS_LEFT = 1;
const unsigned short SOFTKEYS_RIGHT= 2;
const unsigned short SOFTKEYS_TOP = 3;
const unsigned short SOFTKEYS_BOTTOM= 4;
readonly attribute long screenWidth;
readonly attribute long screenHeight;
readonly attribute unsigned short softKeysLayout;
}
“SOFTKEYS_BOTTOM” indicates that the device soft keys at the bottom of the screen in the current screen orientation, as illustrated in FIG. 4 a.
“SOFTKEYS_TOP” indicates that the device soft keys are at the top of the screen in the current screen orientation, as illustrated in FIG. 4 b.
“SOFTKEYS_LEFT” indicates that the device soft keys are to the left of the screen in the current screen orientation, as illustrated in FIG. 4 c.
“SOFTKEYS_RIGHT” indicates that the device soft keys are to the right of the screen in the current screen orientation, as illustrated in FIG. 4 d.
“ScreenWidth” contains the new screen display or viewport width reflecting the new orientation. “ScreenHeight” contains the new screen display or viewport height reflecting the new orientation. “SoftKeysLayout” indicates the location of the device soft keys in response to the orientation change. The possible values are SOFTKEYS_LEFT, SOFTKEYS_RIGHT, SOFTKEYS_TOP, and SOFTKEYS_BOTTOM, as described above with reference to FIGS. 4(a)-(d).
For reference, provided below is the event interface as defined in DOM Level 3 events specification:
interface Event
  {
  readonly attribute EventTarget target;
  readonly attribute DOMString type;
  readonly attribute currentTarget;
  . . .
  . . .
. . .
  }
With the information provided by the screen orientation interface, the user agent or underlying rich media engine can adapt to the layout or visual representation of the content to current screen orientation.
In another embodiment, the event management and processing of screen orientation events is provided. According to an embodiment of the invention, the event management and processing of screen orientation events is based on the DOM Level 3 events processing model. This includes event creation, event dispatching, event registration and event handling.
Event creation deals with the creation of a ScreenOrientationEvent object which is an implementation of ScreenOrientationEvent interface. This event is directly mapped to the hardware event notification (e.g., using the Symbian UI framework) or through an event “message” from the server. During the event creation it is important that the relevant event attributes such as event target, event type, screenWidth, screenHeight attributes are set.
Event dispatching deals with dispatching the ScreenOrientation event to the appropriate target based on the event target.
With regard to the event registration, the content author or application typically registers for the events if he/she chooses to listen to a particular event. For example, to listen to a ScreenOrientationEvent, a listener must be registered with the event target within the rich media engine. This can be done either declaratively using the <ev:listener> element or can be done by the implementation by attaching the “EventListener” object to the “EventTarget” object. This completes the registration process and is usually done prior to the event creation and dispatching phase.
Event handling relates to event processing or realization. Event handlers contain the logic or code that must be executed based on a specific event listener and an event target. For example, when the ScreenOrientationEvent is dispatched to a particular event target as described in the previous stages, the event manager of the rich media engine checks the list of event listeners and invokes the appropriate event handler, thereby executing the logic that is present within the handler. And the handler can access all the information relevant to screen orientation as defined by the ScreenOrientationEvent object.
The following example illustrates the concept of screen orientation and the content adoption based on the orientation change. The content is authored in SVG format. When this content is loaded by the rich media client, the content is seen in the portrait mode (default), as illustrated in FIG. 2. Now, when the screen orientation change event is fired, either by the hardware or streamed from the server, this event gets dispatched to the rich media engine, where the <listener> elements get the notification. Upon this notification, the listener elements invoke the appropriate <handler> elements that, in this example, execute the associated scripts to change the layout of the content from ‘portrait’ mode to ‘landscape’ mode, as illustrated in FIG. 3.
<svg xmlns=“http://www.w3.org/2000/svg” version=“1.2”
baseProfile=“tiny”
 xmlns:ev=“http://www.w3.org/2001/xml-events”>
 xmlns:mynamespace=”http://www.example.com/
     <desc>An example of the screen orientation</desc>
 <rect xml:id=“icon1” x=“10” y=“10” width=“20” height=“10”
fill=“red”/>
 <rect xml:id=“icon3” x=“10” y=“30” width=“20” height=“10”
fill=“green”/>
 <rect xml:id=“icon5” x=“10” y=“50” width=“20” height=“10”
fill=“yellow”/>
 <rect xml:id=“icon2” x=“40” y=“10” width=“20” height=“10”
fill=“blue”/>
 <rect xml:id=“icon4” x=“40” y=“30” width=“20” height=“10”
fill=“purple”/>
 <rect xml:id=“icon6” x=“40” y=“50” width=“20” height=“10”
fill=“orange”/>
 <ev:listener ev:event=“mynamespace: screenOrientationLandscape ”
observer=“icon1” handler=“#theOrientationHandler”/>
  <ev:listener ev:event=“mynamespace: screenOrientationLandscape ”
observer=“icon2” handler=“#theOrientationHandler”/>
  <ev:listener ev:event=“mynamespace: screenOrientationLandscape ”
observer=“icon3” handler=“#theOrientationHandler”/>
  <ev:listener ev:event=“mynamespace: screenOrientationLandscape ”
observer=“icon4” handler=“#theOrientationHandler”/>
  <ev:listener ev:event=“mynamespace: screenOrientationLandscape ”
observer=“icon5” handler=“#theOrientationHandler”/>
  <ev:listener ev:event=“mynamespace: screenOrientationLandscape ”
observer=“icon6” handler=“#theOrientationHandler”/>
 <handler xml:id=“theOrientationHandler”
type=“application/ecmascript”>
    var tarIcon = evt.target;
   if(tar.id == “icon3” && evt.screenWidth > 60 &&
evt.screenWidth < 100){
    tarIcon.setFloatTrait(“x”, 70);
tarIcon.setFloatTrait(“y”, 10); }
   if(tar.id == “icon4”){
    tarIcon.setFloatTrait(“x”, 10);
tarIcon.setFloatTrait(“y”, 30); }
   if(tar.id == “icon5”){
    tarIcon.setFloatTrait(“x”, 40);
tarIcon.setFloatTrait(“y”, 30); }
   if(tar.id == “icon6” && evt.screenWidth > 60 &&
evt.screenWidth < 100){
    tarIcon.setFloatTrait(“x”, 70);
tarIcon.setFloatTrait(“y”, 30); }
 </handler>
</svg>
In one embodiment, hardware-based orientation event triggers are mapped to software-based rich media screen events. While screenOrientation events can be triggered purely from the application software side (e.g., via a button click, menu selection, after a certain time duration, etc.), these events could also be triggered from a change in orientation of the phone screen. In order to address the latter scenario, the mapping from the underlying phone hardware event triggers to the rich media application based screenOrientation events needs to be specified. In order to explain the mapping, we take the example of Symbian, particularly the S60 application framework. Though the S60 application framework is not the only framework used on mobile phones, it is representative of a larger market of such adaptable screen devices.
In Symbian, applications would normally be expected to run in any screen orientation, portrait or landscape. The applications can, however, specify a fixed orientation that they require to work in. By default, applications are set to work in any orientation, so only applications that have a definitive orientation requirement should take any action.
Looking at the API documentation for aknscreenmode.h, one method in this library is:
static IMPORT_C void TAknScreenModes::SetAppUiScreenModeL
( CAknAppUiBase
* aAppUi , const TAknScreenMode & aMode) [static]
Here, the application will be displayed in the selected screen mode. The application will no longer respond to hardware or software events which would normally affect screen mode. It will be permanently fixed to the selected screen mode. The list of available screen modes are specific to that particular device. If an application wants a fixed orientation, it should use the scalable APIs in CEikAppUi. Particularly, a new enumeration and two APIs have been added to CEikAppUi:
    • enum CAknAppUiBase::TAppUiOrientation with possible values;
    • EAppUiOrientationUnspecified This is the default rotation setting for an application, and it should be used by nearly all applications.
    • EAppUiOrientationPortrait Use a portrait screen rotation for this application.
      • This should only be used when an application specifically wants portrait rotation. The device will select an appropriate portrait rotation, if one is available.
    • EAppUiOrientationLandscape Use a landscape screen rotation for this application.
      • This should only be used when an application specifically wants landscape rotation. The device will select an appropriate landscape rotation, if one is available.
    • EAppUiOrientationAutomatic Use the normal device screen rotation set for this application.
      • Both portrait and landscape screen rotations are possible. The application rotation follows device screen rotation.
// Gets the application screen orientation.
mk:@MSITStore:C:\Symbian\9.2\S60_3rd_FP1_4\S60Doc\S60_
CPP_SDK_3rd_FP1_APIRefGuide.chm::/classCAknAppUiBase.html-
ba93469386810cee1e427ba6c4c55c3a#ba93469386810cee1e427
ba6c4c55c3a
TAppUiOrientation CEikAppUi::Orientation( ) const;
// Sets the application screen orientation.
mk:@MSITStore:C:\Symbian\9.2\S60_3rd_FP1_4\S60Doc\S60_
CPP_SDK_3rd_FP1_APIRefGuide.chm::/classCAknAppUiBase.html-
ba93469386810cee1e427ba6c4c55c3a#ba93469386810cee1e427
ba6c4c55c3a
void CEikAppUi::SetOrientationL(TAppUiOrientation
aOrientation);
The above methods can be invoked in AppUI::HandleScreenDeviceChangedL( )
An example is provided below. Here, we are trying to detect a flip-state by checking the current screen mode. Each mode has a different size, and one or more possible rotations/orientations. For the actual mode change detection, you need to implement the method HandleScreenDeviceChangedL( ) on the UI class. A sample header file definition based on the standard HelloWorld example is illustrated below.
class CSimpleAppUi : public CQikAppUi
  {
public:
  void ConstructL ( );
   ~CSimpleAppUi ( );
private:
   // Inherited from class CEikAppUi
   void HandleCommandL(TInt aCommand);
   void HandleScreenDeviceChangedL( );
. . .
HandleScreenDeviceChangedL( ) will be called by the system when the screen mode changes. To check which mode it is in (Flip open or Flip closed), the following code is inserted in the implementation of HandleScreendeviceChangedL( ):
CWsScreenDevice& screenDevice = *(iCoeEnv->ScreenDevice( ));
TBool isFlipOpen = (screenDevice.CurrentScreenMode( ) == 0)
?
  ETrue : EFalse;
Finally, it is ensured that the window server notifies your application when the screen mode changes. To do that, the following is added to the ConstructL( ) method:
RWindowGroup& rootWinGroup = iCoeEnv->RootWin( );
rootWinGroup.EnableScreenChangeEvents( )
This would, in turn, notify the rich media engine, and the corresponding screenOrientation event would be set in the rich media environment for either a specific view in the content or to entire document.
The process flow is illustrated in FIG. 5. The embedded OS framework regularly or continuously polls for hardware orientation changes (block 510). At block 520, the OS framework determines whether the phone orientation has changed. As noted above, this may be achieved through use of sensors in the hardware. If no changes are detected, the OS framework continues to poll for hardware changes. On the other hand, if a change is detected, the process proceeds to block 530 and notifies the application framework. At block 540, if the application framework is responsible for changing orientation, the process proceeds to block 550, and the orientation change is applied to the rich media application. If the application framework is not responsible for changing orientation, the process proceeds to block 560, and the rich media engine is notified of the orientation change. At block 570, the screenOrientation event is triggered and the necessary changes are applied to the content orientation.
In other embodiments, the attribute values for screenOrientation can be extended for future orientation modes. Further, the screenOrientation event types can also be extended for future orientation modes. The mapping of device hardware event triggers to rich media screenOrientation events are agnostic of the underlying embedded operating system.
Various embodiments described herein are described in the general context of method steps or processes, which may be implemented in one embodiment by a computer program product, embodied in a computer-readable medium, including computer-executable instructions, such as program code, executed by computers in networked environments. A computer-readable medium may include removable and non-removable storage devices including, but not limited to, Read Only Memory (ROM), Random Access Memory (RAM), compact discs (CDs), digital versatile discs (DVD), etc. Generally, program modules may include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. Computer-executable instructions, associated data structures, and program modules represent examples of program code for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represents examples of corresponding acts for implementing the functions described in such steps or processes.
Software and web implementations of various embodiments can be accomplished with standard programming techniques with rule-based logic and other logic to accomplish various database searching steps or processes, correlation steps or processes, comparison steps or processes and decision steps or processes. It should be noted that the words “component” and “module,” as used herein and in the following claims, is intended to encompass implementations using one or more lines of software code, and/or hardware implementations, and/or equipment for receiving manual inputs.
The foregoing description of embodiments has been presented for purposes of illustration and description. The foregoing description is not intended to be exhaustive or to limit embodiments of the present invention to the precise form disclosed, and modifications and variations are possible in light of the above teachings or may be acquired from practice of various embodiments. The embodiments discussed herein were chosen and described in order to explain the principles and the nature of various embodiments and its practical application to enable one skilled in the art to utilize the present invention in various embodiments and with various modifications as are suited to the particular use contemplated. The features of the embodiments described herein may be combined in all possible combinations of methods, apparatus, modules, systems, and computer program products.

Claims (23)

What is claimed is:
1. A method comprising:
detecting, by a processor, a screen orientation change event of a screen,
wherein the screen orientation change event is associated with triggering an angle of orientation of the screen,
wherein the triggered angle of orientation has an angle orientation value other than an angle orientation value of a portrait or landscape orientation;
notifying an application framework of the detected screen orientation change event;
retrieving an adjustment of the screen orientation from a rich media engine, one or more applications associated with the application framework, or an operating system associated with the application framework based on the detection of the screen orientation change event, said one or more applications being associated with the rich media engine,
wherein the rich media engine is configured to provide a rendering in a multimedia format of rich-media content in an end-user interface,
wherein the rich media engine includes a screenOrientation attribute,
wherein the screenOrientation attribute includes a screenWidth and/or screenHeight defining one or more screen dimension attributes of the screen,
wherein the screenOrientation attribute includes an attribute value, based on the angle orientation value, specifying the angle of orientation of the screen, other than an angle of portrait or landscape orientation, in rendering the rich-media content in the end-user interface,
determining if the one or more applications or the operating system are configured to accommodate a screen orientation change,
if the application framework is in control of changing the angle of orientation of the screen, then applying the angle of orientation of the screen to the rich media engine,
if the application framework is not in control of changing the angle of orientation of the screen, then notifying the rich media engine of the screen orientation change event,
wherein the screenOrientation attribute specifies a view mode which determines a subset of the rich-media content rendered in the end-user interface; and
adapting a layout and/or visual representation of at least part of the rich-media content in the end-user interface based on the adjustment of the screen orientation.
2. The method of claim 1, wherein the rich media engine is further configured to provide navigation through and/or synchronized interaction with the rendering of the rich-media content in the end-user interface,
wherein the rich media content includes at least one of audio content, video content, text, images, XML-based content and combinations thereof,
wherein the multimedia format is a dynamic and interactive multimedia scene (DIMS) format.
3. The method of claim 2, wherein the screen orientation change event is a hardware change detected via sensors in the device.
4. The method of claim 1, wherein the screen orientation change event is an input by the user.
5. The method of claim 1, wherein the screen orientation change event is a signal received from a remote device.
6. The method of claim 1,
wherein the value of the angle of orientation of the screen is user-defined,
wherein the ScreenOrientation attribute also includes one or more attribute values of a portrait format and/or a landscape format in rendering the rich-media content in the end-user interface.
7. The method of claim 1, wherein the adjustment of the screen orientation repositions one or more soft keys.
8. A non-transitory computer-readable storage medium carrying one or more sequences of one or more instructions which, when executed by one or more processors, cause an apparatus to at least perform the following steps:
detect a screen orientation change event of a screen,
wherein the screen orientation change event is associated with triggering an angle of orientation of the screen,
wherein the triggered angle of orientation has an angle orientation value other than an angle orientation value of a portrait or landscape orientation;
notify an application framework of the detected screen orientation change event;
retrieve an adjustment of the screen orientation from a rich media engine, one or more applications associated with the application framework, or an operating system associated with the application framework based on the detection of the screen orientation change event, said one or more applications being associated with the rich media engine,
wherein the rich media engine is configured to provide a rendering in a multimedia format of rich-media content in an end-user interface,
wherein the rich media engine includes a screenOrientation attribute,
wherein the screenOrientation attribute includes a screenWidth and/or screenHeight defining one or more screen dimension attributes of the screen,
wherein the screenOrientation attribute includes an attribute value, based on the angle orientation value, specifying the angle of orientation of the screen, other than an angle of portrait or landscape orientation, in rendering the rich-media content in the end-user interface,
determine if the one or more applications or the operating system are configured to accommodate a screen orientation change,
if the application framework is in control of changing the angle of orientation of the screen, then applying the angle of orientation of the screen to the rich media engine,
if the application framework is not in control of changing the angle of orientation of the screen, then notifying the rich media engine of the screen orientation change event,
wherein the screenOrientation attribute specifies a view mode which determines a subset of the rich-media content rendered in the end-user interface,
adapt a layout and/or visual representation of at least part of the rich-media content in the end-user interface based on the adjustment of the screen orientation.
9. The non-transitory computer-readable storage medium of claim 8, wherein the rich media engine is further configured to provide navigation through and/or synchronized interaction with the rendering of the rich-media content in the end-user interface,
wherein the rich media content includes at least one of audio content, video content, text, images, XML-based content and combinations thereof,
wherein the multimedia format is a dynamic and interactive multimedia scene (DIMS) format.
10. The non-transitory computer-readable storage medium of claim 9, wherein the screen orientation change event is a hardware change detected via sensors in the device.
11. The non-transitory computer-readable storage medium of claim 8, wherein the screen orientation change event is an input by the user.
12. The non-transitory computer-readable storage medium of claim 8, wherein the screen orientation change event is a signal received from a remote device.
13. The non-transitory computer-readable storage medium of claim 8,
wherein the value of the angle of orientation of the screen is user-defined,
wherein the ScreenOrientation attribute also includes one or more attribute values of a portrait format and/or a landscape format in rendering the rich-media content in the end-user interface.
14. The non-transitory computer-readable storage medium of claim 8, wherein the adjustment of the screen orientation repositions one or more soft keys.
15. An apparatus comprising:
at least one processor; and
at least one memory including computer program code for one or more programs,
the at least one memory and the computer program code configured to, with the at least one processor, cause the apparatus to perform at least the following,
detect a screen orientation change event of a screen,
wherein the screen orientation change event is associated with triggering an angle of orientation of the screen,
wherein the triggered angle of orientation has an angle orientation value other than an angle orientation value of a portrait or landscape orientation;
notify an application framework of the detected screen orientation change event,
retrieve an adjustment of the screen orientation from a rich media engine, one or more applications associated with the application framework, or an operating system associated with the application framework based on the detection of the screen orientation change event, said one or more applications being associated with the rich media engine,
wherein the rich media engine is configured to provide a rendering in a multimedia format of rich-media content in an end-user interface,
wherein the rich media engine includes a screenOrientation attribute,
wherein the screenOrientation attribute includes a screenWidth and/or screenHeight defining one or more screen dimension attributes of the screen,
wherein the screenOrientation attribute includes an attribute value, based on the angle orientation value, specifying the angle of orientation of the screen, other than an angle of portrait or landscape orientation, in rendering the rich-media content in the end-user interface,
determine if the one or more applications or the operating system are configured to accommodate a screen orientation change,
if the application framework is in control of changing the angle of orientation of the screen, then applying the angle of orientation of the screen to the rich media engine,
if the application framework is not in control of changing the angle of orientation of the screen, then notifying the rich media engine of the screen orientation change event,
wherein the screenOrientation attribute specifies a view mode which determines a subset of the rich-media content rendered in the end-user interface,
adapt a layout and/or visual representation of at least part of the rich-media content in the end-user interface based on the adjustment of the screen orientation.
16. The apparatus of claim 15, wherein the rich media engine is further configured to provide navigation through and/or synchronized interaction with the rendering of the rich-media content in the end-user interface,
wherein the rich media content includes at least one of audio content, video content, text, images, XML-based content and combinations thereof,
wherein the multimedia format is a dynamic and interactive multimedia scene (DIMS) format.
17. The apparatus of claim 16, wherein the screen orientation change event is a hardware change detected via sensors in the device.
18. The apparatus of claim 15, wherein the screen orientation change event is an input by the user.
19. The apparatus of claim 15, wherein the screen orientation change event is a signal received from a remote device.
20. The apparatus of claim 15,
wherein the value of the angle of orientation of the screen is user-defined,
wherein the ScreenOrientation attribute also includes one or more attribute values of a portrait format and/or a landscape format in rendering the rich-media content in the end-user interface.
21. The apparatus of claim 15, wherein the adjustment of the screen orientation repositions one or more soft keys.
22. A rich media engine interface embodied on a non-transitory computer-readable storage medium comprising:
a ScreenOrientationEvent attribute adapted to adjust a screen orientation of a screen upon a screen orientation change event,
wherein the screen orientation change event is associated with triggering an angle of orientation of the screen,
wherein the triggered angle of orientation has an angle orientation value other than an angle orientation value of a portrait or landscape orientation;
one or more screen dimension attributes retrieved from one or more applications or an operating system defining dimensions of the screen after changing orientation,
wherein the rich media engine interface is configured to provide a rendering in a multimedia format of rich-media content,
wherein the screenOrientationEvent attribute includes a screenWidth and/or screenHeight defining one or more screen dimension attributes of the screen,
wherein the screenOrientationEvent attribute includes a value of a specification of an angle of orientation format, other than an angle of portrait or landscape orientation, rendering the rich-media content in the screen,
if an application framework is in control of changing an angle of orientation of the screen, then applying the angle of orientation of the screen to the rich media engine associated with the rich media engine interface,
if the application framework is not in control of changing the angle of orientation of the screen, then notifying the rich media engine of the screen orientation change event,
wherein the screenOrientation attribute specifies a view mode which determines a subset of the rich-media content rendered in the end-user interface.
23. The rich media engine interface of claim 22, further comprising:
one or more soft key attributes defining positions of one or more soft keys on the screen,
wherein the screenOrientationEvent attribute specifies a view mode which determines a subset of the rich-media content rendered in the end-user interface,
wherein the rich media engine interface is further configured to provide navigation through and/or synchronized interaction with the rendering of the rich-media content,
wherein the rich media content includes at least one of audio content, video content, text, images, XML-based content and combinations thereof,
wherein the multimedia format is a dynamic and interactive multimedia scene (DIMS) format.
US12/018,119 2007-01-22 2008-01-22 System and method for screen orientation in a rich media environment Active 2032-02-04 US9813531B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/018,119 US9813531B2 (en) 2007-01-22 2008-01-22 System and method for screen orientation in a rich media environment

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US88609407P 2007-01-22 2007-01-22
US12/018,119 US9813531B2 (en) 2007-01-22 2008-01-22 System and method for screen orientation in a rich media environment

Publications (2)

Publication Number Publication Date
US20080209442A1 US20080209442A1 (en) 2008-08-28
US9813531B2 true US9813531B2 (en) 2017-11-07

Family

ID=39644162

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/018,119 Active 2032-02-04 US9813531B2 (en) 2007-01-22 2008-01-22 System and method for screen orientation in a rich media environment

Country Status (2)

Country Link
US (1) US9813531B2 (en)
WO (1) WO2008090516A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170286715A1 (en) * 2016-03-31 2017-10-05 Ca, Inc. Prevention of shoulder surfing
US10775838B2 (en) 2019-01-11 2020-09-15 Disney Enterprises, Inc. Mounted displays that autorotate to match content on display

Families Citing this family (50)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7312785B2 (en) 2001-10-22 2007-12-25 Apple Inc. Method and apparatus for accelerated scrolling
US7333092B2 (en) 2002-02-25 2008-02-19 Apple Computer, Inc. Touch pad for handheld device
US20070152977A1 (en) 2005-12-30 2007-07-05 Apple Computer, Inc. Illuminated touchpad
US7499040B2 (en) 2003-08-18 2009-03-03 Apple Inc. Movable touch pad with added functionality
US8059099B2 (en) 2006-06-02 2011-11-15 Apple Inc. Techniques for interactive input to portable electronic devices
US7495659B2 (en) 2003-11-25 2009-02-24 Apple Inc. Touch pad for handheld device
EP1779228B1 (en) 2004-08-16 2012-04-11 Apple Inc. A method of increasing the spatial resolution of touch sensitive devices
US20070152983A1 (en) 2005-12-30 2007-07-05 Apple Computer, Inc. Touch pad with symbols based on mode
US8743060B2 (en) 2006-07-06 2014-06-03 Apple Inc. Mutual capacitance touch sensing device
US8022935B2 (en) 2006-07-06 2011-09-20 Apple Inc. Capacitance sensing electrode with integrated I/O mechanism
US9360967B2 (en) 2006-07-06 2016-06-07 Apple Inc. Mutual capacitance touch sensing device
US7795553B2 (en) * 2006-09-11 2010-09-14 Apple Inc. Hybrid button
US8274479B2 (en) 2006-10-11 2012-09-25 Apple Inc. Gimballed scroll wheel
US8482530B2 (en) 2006-11-13 2013-07-09 Apple Inc. Method of capacitively sensing finger position
WO2009032898A2 (en) 2007-09-04 2009-03-12 Apple Inc. Compact input device
US8683378B2 (en) 2007-09-04 2014-03-25 Apple Inc. Scrolling techniques for user interfaces
US9851800B1 (en) 2007-11-05 2017-12-26 Sprint Communications Company L.P. Executing computing tasks based on force levels
US8416198B2 (en) 2007-12-03 2013-04-09 Apple Inc. Multi-dimensional scroll wheel
US8125461B2 (en) * 2008-01-11 2012-02-28 Apple Inc. Dynamic input graphic display
US8820133B2 (en) 2008-02-01 2014-09-02 Apple Inc. Co-extruded materials and methods
US20090207134A1 (en) * 2008-02-14 2009-08-20 Netgear Inc. Remote control apparatus with integrated positional responsive alphabetic keyboard
US9454256B2 (en) 2008-03-14 2016-09-27 Apple Inc. Sensor configurations of an input device that are switchable based on mode
US20100056127A1 (en) * 2008-08-27 2010-03-04 John Osborne System and Method for Application Layer Widgets for Mobile Devices
US8816967B2 (en) 2008-09-25 2014-08-26 Apple Inc. Capacitive sensor having electrodes arranged on the substrate and the flex circuit
US20100083163A1 (en) * 2008-09-29 2010-04-01 Yahoo! Inc. Methods and Systems for Optimizing Webpage Content Based on a Screen Orientation of a Device
EP2175345A1 (en) * 2008-10-07 2010-04-14 Research in Motion Limited A method and handheld electronic device having a graphic user interface with efficient orientation sensor use
KR20100040545A (en) * 2008-10-10 2010-04-20 삼성전자주식회사 Apparatus and method for providing user interface based structured rich media data
US8395590B2 (en) 2008-12-17 2013-03-12 Apple Inc. Integrated contact switch and touch sensor elements
US8719701B2 (en) * 2009-01-02 2014-05-06 Apple Inc. Identification of guides and gutters of a document
US9354751B2 (en) 2009-05-15 2016-05-31 Apple Inc. Input device with optimized capacitive sensing
US8872771B2 (en) 2009-07-07 2014-10-28 Apple Inc. Touch sensing device having conductive nodes
US8493346B2 (en) * 2009-12-31 2013-07-23 International Business Machines Corporation Morphing touchscreen keyboard interface
US8819576B2 (en) * 2011-05-09 2014-08-26 Blackberry Limited Systems and methods for facilitating an input to an electronic device
US9311426B2 (en) * 2011-08-04 2016-04-12 Blackberry Limited Orientation-dependent processing of input files by an electronic device
KR20150013591A (en) * 2012-04-26 2015-02-05 퀄컴 인코포레이티드 Altering attributes of content that is provided in a portion of a display area based on detected inputs
US9357272B2 (en) 2012-08-03 2016-05-31 Intel Corporation Device orientation capability exchange signaling and server adaptation of multimedia content in response to device orientation
WO2014062730A1 (en) 2012-10-15 2014-04-24 Famous Industries, Inc. Efficient manipulation of surfaces in multi-dimensional space using energy agents
US9772889B2 (en) * 2012-10-15 2017-09-26 Famous Industries, Inc. Expedited processing and handling of events
US9501171B1 (en) 2012-10-15 2016-11-22 Famous Industries, Inc. Gesture fingerprinting
US10877780B2 (en) 2012-10-15 2020-12-29 Famous Industries, Inc. Visibility detection using gesture fingerprinting
US10908929B2 (en) 2012-10-15 2021-02-02 Famous Industries, Inc. Human versus bot detection using gesture fingerprinting
US10592064B2 (en) * 2013-09-17 2020-03-17 Amazon Technologies, Inc. Approaches for three-dimensional object display used in content navigation
US20150109532A1 (en) * 2013-10-23 2015-04-23 Google Inc. Customizing mobile media captioning based on mobile media rendering
CN106462180B (en) 2014-04-30 2020-01-03 惠普发展公司,有限责任合伙企业 Screen orientation adjustment
US10139998B2 (en) 2014-10-08 2018-11-27 Weebly, Inc. User interface for editing web content
CN105302445B (en) * 2015-11-12 2019-07-23 小米科技有限责任公司 Graphic user interface method for drafting and device
US9723258B2 (en) * 2015-12-30 2017-08-01 Roku, Inc. Orientation adjustment for casting videos
US11620976B2 (en) 2020-06-09 2023-04-04 Meta Platforms Technologies, Llc Systems, devices, and methods of acoustic echo cancellation based on display orientation
US11340861B2 (en) 2020-06-09 2022-05-24 Facebook Technologies, Llc Systems, devices, and methods of manipulating audio data based on microphone orientation
US11586407B2 (en) 2020-06-09 2023-02-21 Meta Platforms Technologies, Llc Systems, devices, and methods of manipulating audio data based on display orientation

Citations (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1124175A2 (en) 2000-02-08 2001-08-16 Nokia Corporation Display apparatus
US6348928B1 (en) * 1998-11-13 2002-02-19 Lg Electronics Inc. Apparatus for automatically rotating visual display unit and method therefor
US20020021278A1 (en) * 2000-07-17 2002-02-21 Hinckley Kenneth P. Method and apparatus using multiple sensors in a device with a display
US20020078158A1 (en) * 2000-08-28 2002-06-20 Brown Scott T. E-mail messaging system and method for enhanced rich media delivery
US20020091615A1 (en) * 2001-01-09 2002-07-11 Salvani Joseph M. Transaction communication system
US20030203747A1 (en) * 2002-04-26 2003-10-30 Nec Corporation Foldable portable telephone having a display portion selectively put into a lengthwise state or an oblong state and a pair of front camera portions
US20030236836A1 (en) * 2002-03-21 2003-12-25 Borthwick Ernest Mark System and method for the design and sharing of rich media productions via a computer network
US20040103373A1 (en) * 2002-11-19 2004-05-27 Wei Coach K. System and method for stateful web-based computing
US20040257337A1 (en) * 2003-04-04 2004-12-23 Canon Kabushiki Kaisha Display control device and method, and display system
WO2005039157A1 (en) * 2003-10-22 2005-04-28 Sanyo Electric Co., Ltd. Mobile telephone apparatus, display method, and program
US20060046848A1 (en) * 2004-08-31 2006-03-02 Nintendo Co., Ltd., Game apparatus, storage medium storing a game program, and game control method
US20060101162A1 (en) * 2004-11-08 2006-05-11 Opera Software Asa Method and device for adapting a computer program to a device on which it is installed
US7113194B2 (en) * 2001-01-30 2006-09-26 Ati International Srl Method and apparatus for rotating an image on a display
US20060238384A1 (en) * 2005-01-05 2006-10-26 Richard Hess System and method for portable communication device in an aircraft
US20070004451A1 (en) * 2005-06-30 2007-01-04 C Anderson Eric Controlling functions of a handheld multifunction device
US20070044028A1 (en) * 2004-04-01 2007-02-22 Dunn Michael H Virtual flip chart method and apparatus
US20070111750A1 (en) * 2003-10-22 2007-05-17 Josef Stohr Mobile communication terminal with multi orientation user interface
US20070232336A1 (en) * 2006-04-04 2007-10-04 Samsung Electronics Co., Ltd. Apparatus and method for automatic display control in mobile terminal
WO2007115983A1 (en) * 2006-04-10 2007-10-18 Streamezzo Method for restitution of at least one multimedia scene
US20080146339A1 (en) * 2006-12-14 2008-06-19 Arlen Lynn Olsen Massive Multiplayer Online Sports Teams and Events
US20090055415A1 (en) * 2007-08-24 2009-02-26 Microsoft Corporation Dynamic and versatile notepad
US20090137272A1 (en) * 2005-11-22 2009-05-28 Kei Okuda Mobile phone
US7746343B1 (en) * 2005-06-27 2010-06-29 Google Inc. Streaming and interactive visualization of filled polygon data in a geographic information system
US7844301B2 (en) * 2005-10-14 2010-11-30 Lg Electronics Inc. Method for displaying multimedia contents and mobile communications terminal capable of implementing the same
US7941271B2 (en) * 2006-11-21 2011-05-10 Microsoft Corporation Displaying images related to a requested path
US8036347B1 (en) * 2006-10-18 2011-10-11 Avaya Inc. Method and apparatus providing additional information to an interactive voice response (IVR) system user

Patent Citations (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6348928B1 (en) * 1998-11-13 2002-02-19 Lg Electronics Inc. Apparatus for automatically rotating visual display unit and method therefor
EP1124175A2 (en) 2000-02-08 2001-08-16 Nokia Corporation Display apparatus
US20020021278A1 (en) * 2000-07-17 2002-02-21 Hinckley Kenneth P. Method and apparatus using multiple sensors in a device with a display
US20020078158A1 (en) * 2000-08-28 2002-06-20 Brown Scott T. E-mail messaging system and method for enhanced rich media delivery
US20020091615A1 (en) * 2001-01-09 2002-07-11 Salvani Joseph M. Transaction communication system
US7113194B2 (en) * 2001-01-30 2006-09-26 Ati International Srl Method and apparatus for rotating an image on a display
US20030236836A1 (en) * 2002-03-21 2003-12-25 Borthwick Ernest Mark System and method for the design and sharing of rich media productions via a computer network
US20030203747A1 (en) * 2002-04-26 2003-10-30 Nec Corporation Foldable portable telephone having a display portion selectively put into a lengthwise state or an oblong state and a pair of front camera portions
US20040103373A1 (en) * 2002-11-19 2004-05-27 Wei Coach K. System and method for stateful web-based computing
US20040257337A1 (en) * 2003-04-04 2004-12-23 Canon Kabushiki Kaisha Display control device and method, and display system
WO2005039157A1 (en) * 2003-10-22 2005-04-28 Sanyo Electric Co., Ltd. Mobile telephone apparatus, display method, and program
US20070265031A1 (en) * 2003-10-22 2007-11-15 Sandy Electric Co., Ltd. Mobile Phone, Display Method, and Computer Program
US20070111750A1 (en) * 2003-10-22 2007-05-17 Josef Stohr Mobile communication terminal with multi orientation user interface
US20070044028A1 (en) * 2004-04-01 2007-02-22 Dunn Michael H Virtual flip chart method and apparatus
US20060046848A1 (en) * 2004-08-31 2006-03-02 Nintendo Co., Ltd., Game apparatus, storage medium storing a game program, and game control method
US20060101162A1 (en) * 2004-11-08 2006-05-11 Opera Software Asa Method and device for adapting a computer program to a device on which it is installed
US20060238384A1 (en) * 2005-01-05 2006-10-26 Richard Hess System and method for portable communication device in an aircraft
US7746343B1 (en) * 2005-06-27 2010-06-29 Google Inc. Streaming and interactive visualization of filled polygon data in a geographic information system
US20070004451A1 (en) * 2005-06-30 2007-01-04 C Anderson Eric Controlling functions of a handheld multifunction device
US7844301B2 (en) * 2005-10-14 2010-11-30 Lg Electronics Inc. Method for displaying multimedia contents and mobile communications terminal capable of implementing the same
US20090137272A1 (en) * 2005-11-22 2009-05-28 Kei Okuda Mobile phone
US20070232336A1 (en) * 2006-04-04 2007-10-04 Samsung Electronics Co., Ltd. Apparatus and method for automatic display control in mobile terminal
WO2007115983A1 (en) * 2006-04-10 2007-10-18 Streamezzo Method for restitution of at least one multimedia scene
US20090305743A1 (en) * 2006-04-10 2009-12-10 Streamezzo Process for rendering at least one multimedia scene
US8036347B1 (en) * 2006-10-18 2011-10-11 Avaya Inc. Method and apparatus providing additional information to an interactive voice response (IVR) system user
US7941271B2 (en) * 2006-11-21 2011-05-10 Microsoft Corporation Displaying images related to a requested path
US20080146339A1 (en) * 2006-12-14 2008-06-19 Arlen Lynn Olsen Massive Multiplayer Online Sports Teams and Events
US20090055415A1 (en) * 2007-08-24 2009-02-26 Microsoft Corporation Dynamic and versatile notepad

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
3GPP TSG-SA4#41, Tdoc S4-060640, Screen orientation management in DIMS, Nov. 6-10, 2006.
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or The Declaration; International Search Report for corresponding International Application No. PCT/IB2008/050225, dated Jun. 12, 2008, 5 pages.
W3C, "Scalable Vector Graphics (SVG) 1.0 Specificaiton", Sep. 2001, pp. 1-617. *
Wick, "Changing Screen Orientation Programmatically", Oct. 2004, pp. 1-7. *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170286715A1 (en) * 2016-03-31 2017-10-05 Ca, Inc. Prevention of shoulder surfing
US10380377B2 (en) * 2016-03-31 2019-08-13 Ca, Inc. Prevention of shoulder surfing
US10775838B2 (en) 2019-01-11 2020-09-15 Disney Enterprises, Inc. Mounted displays that autorotate to match content on display

Also Published As

Publication number Publication date
US20080209442A1 (en) 2008-08-28
WO2008090516A1 (en) 2008-07-31

Similar Documents

Publication Publication Date Title
US9813531B2 (en) System and method for screen orientation in a rich media environment
KR102364443B1 (en) Display apparatus for displaying and method thereof
US8255825B2 (en) Content aware adaptive display
US11025919B2 (en) Client-based adaptive streaming of nonlinear media
US9628744B2 (en) Display apparatus and control method thereof
US20140337792A1 (en) Display apparatus and user interface screen providing method thereof
US9836437B2 (en) Screencasting for multi-screen applications
US20150227224A1 (en) User terminal device and displaying method thereof
CN112533021B (en) Display method and display equipment
WO2019072096A1 (en) Interactive method, device, system and computer readable storage medium in live video streaming
US20120210205A1 (en) System and method for using an application on a mobile device to transfer internet media content
US20090305743A1 (en) Process for rendering at least one multimedia scene
CN112514398A (en) Method and apparatus for marking user interactions on overlays for omnidirectional content and grouping overlays to a background
WO2021179359A1 (en) Display device and display picture rotation adaptation method
US20080046817A1 (en) Hierarchical state programming with a markup language
KR20160135435A (en) Display device and controlling method thereof
US20140173516A1 (en) Display apparatus and method of providing user interface thereof
US20090140977A1 (en) Common User Interface Structure
US20120044138A1 (en) METHOD AND APPARATUS FOR PROVIDING USER INTERACTION IN LASeR
CN112399263A (en) Interaction method, display device and mobile terminal
WO2023011058A1 (en) Display device, communication terminal, and projected-screen image dynamic display method
JP4849756B2 (en) Method and apparatus for generating a video window with parameters for determining position and scaling factor
US20130167075A1 (en) Managing Display Areas
CN112771878B (en) Method, client and server for processing media data
US20170094367A1 (en) Text Data Associated With Separate Multimedia Content Transmission

Legal Events

Date Code Title Description
AS Assignment

Owner name: NOKIA CORPORATION,FINLAND

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SETLUR, VIDYA;CHITTURI, SURESH;SIGNING DATES FROM 20080313 TO 20080331;REEL/FRAME:020928/0371

Owner name: NOKIA CORPORATION, FINLAND

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SETLUR, VIDYA;CHITTURI, SURESH;SIGNING DATES FROM 20080313 TO 20080331;REEL/FRAME:020928/0371

AS Assignment

Owner name: SISVEL INTERNATIONAL S.A., LUXEMBOURG

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NOKIA CORPORATION;REEL/FRAME:028379/0241

Effective date: 20120419

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4