WO2015026099A1 - 디스플레이 장치가 화면을 디스플레이 하는 방법 및 그 디스플레이 장치 - Google Patents
디스플레이 장치가 화면을 디스플레이 하는 방법 및 그 디스플레이 장치 Download PDFInfo
- Publication number
- WO2015026099A1 WO2015026099A1 PCT/KR2014/007540 KR2014007540W WO2015026099A1 WO 2015026099 A1 WO2015026099 A1 WO 2015026099A1 KR 2014007540 W KR2014007540 W KR 2014007540W WO 2015026099 A1 WO2015026099 A1 WO 2015026099A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- button
- touch screen
- touch input
- displayed
- display device
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04886—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04842—Selection of displayed objects or displayed text elements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04845—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04803—Split screen, i.e. subdividing the display area or the window area into separate subareas
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04806—Zoom, i.e. interaction techniques or interactors for controlling the zooming operation
Definitions
- the present invention relates to a method for displaying a screen and a display device of the present invention, and more particularly, to a method and a device for displaying at least one application execution screen.
- a mobile device eg, mobile phone, smartphone or tablet PC
- a desktop computer or touch screen may have at least one display device.
- the user may divide and use the screen of the display device according to the working environment.
- the page up button or the page down button on the keyboard may be used to move up or down the web page.
- the mouse cursor can be used to select a scroll bar on the side of the web page to move up or down the web page.
- a mobile device has a relatively small screen size that can be output, and has a limitation on an input method.
- Mobile devices have difficulty in splitting and using the screen.
- various applications are provided in a mobile device to satisfy a consumer's desire.
- the mobile device is manufactured in a portable size, there are limitations on the size of a display and a provided user interface (UI). Accordingly, there is a inconvenience for a user in executing a plurality of applications in a mobile device. For example, in a mobile device, an execution screen of an application is displayed in the entire area of the display to execute one application. Then, to execute another application, the other application must be executed after terminating the displayed application. In other words, in order to execute various applications in a mobile device, it is cumbersome to repeat execution and termination processes for each application.
- a display method and a display apparatus for easily controlling a plurality of application execution windows while displaying a plurality of application execution windows on one display included in the display device are provided.
- a display device having a touch screen displays a screen, the method comprising the steps of displaying a button on the touch screen, and the position where the button is displayed Dividing the touch screen into a plurality of areas based on the received data, receiving a touch input for moving a displayed button, acquiring an inclination value of a line connecting a start point and an end point of the touch input, The method may include selecting an area corresponding to an inclination value from the area and moving a button to a predetermined position included in the selected area.
- a user interface that allows a user to more intuitively control a plurality of application execution windows may be provided.
- FIG. 1 is a block diagram illustrating a display apparatus according to an exemplary embodiment.
- FIGS. 2A through 2E are conceptual views illustrating a method of displaying an application execution window according to an embodiment of the present invention.
- 3A to 3E are conceptual views illustrating a method of moving a position of a displayed button according to one embodiment of the present invention.
- FIG. 4 is a flowchart illustrating a process of displaying a button according to an embodiment of the present invention.
- FIG. 5 is a flowchart illustrating a process of recognizing a fling gesture according to an embodiment of the present invention.
- 6A through 6C are conceptual diagrams for describing a method of determining the direction of fling.
- FIG. 7 is a block diagram schematically illustrating a configuration of a display apparatus according to an exemplary embodiment.
- FIG. 8 is a conceptual diagram illustrating a method of recognizing a fling gesture by a display apparatus according to another exemplary embodiment.
- FIG. 9 is a conceptual diagram illustrating a method of recognizing a fling gesture by a display apparatus according to another exemplary embodiment.
- 10 to 14 are exemplary views illustrating an embodiment in which the display apparatus is a wearable device.
- a display apparatus having a touch screen displays a screen, including displaying a button on the touch screen, and dividing the touch screen into a plurality of areas based on a position at which the button is displayed.
- Receiving a touch input for moving the displayed button acquiring an inclination value of a line connecting the start point and the end point of the touch input, and selecting an area corresponding to the inclination value from the plurality of divided regions And moving the button to a predetermined position included in the selected area.
- the obtaining of the inclination value may include acquiring an acceleration value of the movement of the button and acquiring the inclination value when the acceleration value is greater than or equal to a preset threshold.
- the predetermined position may be included in the edge area of the touch screen.
- the step of dividing the touch screen into the plurality of areas may be performed by separating the touch screen with a line segment connecting a plurality of preset points of the edge area of the touch screen from the position where the button is displayed.
- the method may include dividing into a plurality of regions.
- the obtaining of the slope value may be characterized by determining the coordinates of the touch input before the preset number as the start point of the touch input from the end point of the touch input.
- a display apparatus may include a touch screen for displaying a button and a controller for dividing the touch screen into a plurality of areas based on a position at which the button is displayed, and the touch screen moves the displayed button.
- Receives a touch input and the control unit obtains the inclination value of the line connecting the start point and the end point of the touch input based on the touch input, selects the area corresponding to the inclination value from the plurality of divided areas, included in the selected area
- the touch screen may be controlled to move and display the button to a predetermined position.
- the controller may acquire the acceleration value of the movement of the button, and control the touch screen to display the button at a predetermined position when the acceleration value is greater than or equal to a preset threshold.
- the predetermined position may be included in the edge area of the touch screen.
- control unit divides the touch screen into a plurality of areas by a line segment connecting a plurality of preset points of the edge area of the touch screen from the position where the button is displayed. Can be.
- the controller may determine the coordinates of the touch input before the preset number as the start point of the touch input from the end point of the touch input.
- a program for executing the above-described method on a computer may be recorded on a computer-readable recording medium.
- touch refers to not only the case where the touch device directly contacts the display device, but also the display device detects the access of the user's body (eg, a finger). It may include input.
- FIG. 1 is a block diagram illustrating a display apparatus according to an exemplary embodiment.
- FIG. 1 is only a display device according to an exemplary embodiment, and the display device according to the present invention may include fewer or more components than those shown in FIG. 1.
- the display apparatus 100 may be connected to an external device (not shown) using the mobile communication module 120, the sub communication module 130, and the connector 165.
- the external device may include at least one of another device (not shown), a mobile phone (not shown), a smart phone (not shown), a tablet PC (not shown), and a server (not shown).
- the display apparatus 100 includes a touch screen 190 and a touch screen controller 195.
- the display apparatus 100 may include a controller 110, a mobile communication module 120, a sub communication module 130, a multimedia module 140, a camera module 150, a GPS module 155, and an input / output module ( 160, a sensor module 170, a storage unit 175, and a power supply unit 180.
- the sub communication module 130 includes at least one of a wireless LAN module 131 and a short range communication module 132
- the multimedia module 140 includes a broadcast communication module 141, an audio play module 142, and a video play module.
- 143 may include at least one.
- the camera module 150 includes at least one of the first camera 151 and the second camera 152, and the input / output module 160 includes a button 161, a microphone 162, a speaker 163, and vibrations. It may include at least one of the motor 164, the connector 165 and the keypad 166.
- the controller 110 may store a signal or data input from the outside of the ROM 112 and the display apparatus 100, in which the CPU 111, a control program for controlling the display apparatus 100 is stored, or the display apparatus 100.
- the RAM 113 may be used as a storage area for a task performed at.
- the CPU 111 may include a plurality of processors such as single core, dual core, triple core, or quad core.
- the CPU 111, the ROM 112, and the RAM 113 may be connected to each other through an internal bus.
- the controller 110 may include a mobile communication module 120, a sub communication module 130, a multimedia module 140, a camera module 150, a GPS module 155, an input / output module 160, a sensor module 170, and a storage module.
- the unit 175, the power supply unit 180, the touch screen 190, and the touch screen controller 195 may be controlled.
- the mobile communication module 120 may allow the display apparatus 100 to be connected to an external device through mobile communication using at least one (one or more) antennas (not shown) under the control of the controller 110. .
- the mobile communication module 120 may be connected to a mobile phone (not shown), a smartphone (not shown), a tablet PC (not shown), or another device (not shown) having a phone number input to the display device 100.
- a wireless signal for transmitting a voice call, a video call, a text message (SMS) or a multimedia message (MMS) may be transmitted / received.
- the sub communication module 130 may include at least one of the WLAN module 131 and the short range communication module 132.
- the WLAN module 131 may be included, only the local area communication module 132 may be included, or both the WLAN module 131 and the local area communication module 132 may be included.
- the WLAN module 131 may connect to the Internet at a place where a wireless access point (AP) (not shown) is installed under the control of the controller 110.
- the WLAN module 131 may support a WLAN standard IEEE802.11x of the Institute of Electrical and Electronics Engineers (IEEE).
- the short range communication module 132 may wirelessly perform short range communication between the display apparatus 100 and an image forming apparatus (not shown) under the control of the controller 110.
- the short range communication method may include a Bluetooth, an infrared data association (IrDA), and a Zig-bee method.
- the display apparatus 100 may include at least one of the mobile communication module 120, the wireless LAN module 131, and the short range communication module 132 according to performance.
- the multimedia module 140 may include a broadcast communication module 141, an audio play module 142, or a video play module 143.
- the broadcast communication module 141 broadcasts a broadcast signal (for example, a TV broadcast signal, a radio broadcast signal or a data broadcast signal) transmitted from a broadcast station through a broadcast communication antenna (not shown) under the control of the controller 110. Additional information (eg, an electric program guide (EPG) or an electric service guide (ESG)) may be received.
- EPG electric program guide
- ESG electric service guide
- the audio reproduction module 142 may reproduce the digital audio file stored or received under the control of the controller 110.
- the video play module 143 may play a digital video file stored or received under the control of the controller 110.
- the video play module 143 may play a digital audio file.
- the multimedia module 140 may include an audio play module 142 and a video play module 143 except for the broadcast communication module 141.
- the audio playback module 142 or the video playback module 143 of the multimedia module 140 may be included in the controller 100.
- the camera module 150 may include at least one of the first camera 151 and the second camera 152 for capturing a still image or a video under the control of the controller 110.
- the first camera 151 or the second camera 152 may include an auxiliary light source (not shown) that provides the amount of light required for imaging.
- the first camera 151 may be disposed on the front surface of the display apparatus 100, and the second camera 152 may be disposed on the rear surface of the display apparatus 100.
- the first camera 151 and the second camera 152 may be disposed adjacent to each other (eg, larger than 1 cm and smaller than 8 cm) to capture a 3D still image or a 3D video.
- the GPS module 155 receives radio waves from a plurality of GPS satellites (not shown) on the Earth's orbit, and uses a time of arrival from the GPS satellites (not shown) to the display apparatus 100. The position of the display apparatus 100 can be calculated.
- the input / output module 160 may include at least one of a plurality of buttons 161, a microphone 162, a speaker 163, a vibration motor 164, a connector 165, and a keypad 166.
- Button 161 may be formed on the front, side or rear of the housing of the display device 100, the power / lock button (not shown), volume button (not shown) menu button, home button, return It may include at least one of a button (back button) and the search button 161.
- the microphone 162 may generate an electrical signal by receiving a voice or sound under the control of the controller 110.
- the speaker 163 displays sound corresponding to various signals of the mobile communication module 120, the sub communication module 130, the multimedia module 140, or the camera module 150 under the control of the controller 110. ) Can be output to the outside.
- the speaker 163 may output a sound corresponding to a function performed by the display apparatus 100.
- One or more speakers 163 may be formed at appropriate positions or positions of the housing of the display apparatus 100.
- the vibration motor 164 may convert an electrical signal into mechanical vibration under the control of the controller 110. For example, when the voice call is received from another device (not shown) in the vibration mode, the vibration motor 164 may operate. The vibration motor 164 may operate in response to a user's touch operation on the touch screen 190 and a continuous movement of a touch input on the touch screen 190.
- the connector 165 may be used as an interface for connecting the display apparatus 100 and an external device (not shown) or a power source (not shown). Under the control of the controller 110, data stored in the storage unit 175 of the display apparatus 100 may be transmitted to an external device (not shown) or an external device (not shown) through a wired cable connected to the connector 165. You can receive data from. Power may be input from a power source (not shown) or charged with a battery (not shown) through a wired cable connected to the connector 165.
- the keypad 166 may receive a key input from the user for the control of the display apparatus 100.
- the keypad 166 includes a physical keypad (not shown) formed on the display device 100 or a virtual keypad (not shown) displayed on the touch screen 190. Physical capacities (not shown) formed in the display apparatus 100 may be excluded depending on the performance or structure of the display apparatus 100.
- the sensor module 170 includes at least one sensor that detects a state of the display apparatus 100.
- the sensor module 170 may include a proximity sensor that detects whether the user approaches the display device 100, an illuminance sensor (not shown), or a display device that detects an amount of light around the display device 100. It may include a motion sensor (not shown) that detects an operation of the display device 100 (eg, rotation of the display device 100, acceleration or vibration applied to the display device 100).
- the sensor of the sensor module 170 may be added or deleted according to the performance of the display apparatus 100.
- the storage unit 175 may control the mobile communication module 120, the sub communication module 130, the multimedia module 140, the camera module 150, the GPS module 155, and the input / output module under the control of the controller 110.
- the input / output signal or data corresponding to the operation of the sensor module 170 and the touch screen 190 may be stored.
- the storage unit 175 may store a control program and applications for controlling the display apparatus 100 or the controller 110.
- the term “storage unit” may include a storage unit 175, a ROM 112 in the controller 110, a RAM 113, or a memory card (not shown) mounted in the display apparatus 100.
- the storage unit may include a nonvolatile memory, a volatile memory, a hard disk drive (HDD), or a solid state drive (SSD).
- the power supply unit 180 may supply power to at least one battery (not shown) disposed in the housing of the display apparatus 100 under the control of the controller 110. In addition, the power supply unit 180 may supply power input from an external power source (not shown) to each unit of the display apparatus 100 through a wired cable connected to the connector 165.
- the touch screen 190 may output a user interface corresponding to various services to the user.
- the touch screen 190 may transmit an analog signal corresponding to at least one touch input to the user interface to the touch screen controller 195.
- the touch screen 190 may receive at least one touch input through a user's body (eg, a finger) or a touchable input means (eg, a stylus pen).
- the touch screen 190 may receive continuous movement of at least one touch input.
- the touch screen 190 may transmit an analog signal corresponding to continuous movement of the touch input to the touch screen controller 195.
- the touch input is not limited to an input through contact between the touch screen 190 and the user's body or touchable input means, and a non-contact input (for example, the distance between the touch screen 190 and the body is 1 mm or less).
- the detectable interval on the touch screen 190 may be changed according to the performance or structure of the display apparatus 100.
- the touch screen 190 may be implemented by, for example, a resistive method, a capacitive method, an infrared method, or an ultrasonic wave method.
- the touch screen controller 195 may convert an analog signal received from the touch screen 190 into a digital signal (for example, X coordinates and Y coordinates) and transmit the converted analog signal to the controller 110.
- the controller 110 may control the touch screen 190 by using a digital signal received from the touch screen controller 195.
- the controller 110 may select an application execution icon (not shown) displayed on the touch screen 190 or execute an application in response to a touch input.
- the touch screen controller 195 may be included in the touch screen 190 or the controller 110.
- FIG. 2A through 2E are conceptual views illustrating a method of displaying an application execution window according to an embodiment of the present invention.
- the display device 200 may be implemented in various forms such as a TV or a medical data display device as well as the mobile device described above with reference to FIG. 1.
- the display apparatus 200 may set the plurality of regions 201 to 204 on the touch screen.
- the region set by the display apparatus 200 is referred to as a first region 201, a second region 202, a third region 203, and a fourth region 204. do.
- an area in which the first area 201 and the third area 203 are combined is referred to as a fifth area (not shown).
- the region in which the second region 202 and the fourth region 204 are combined is referred to as a sixth region (not shown).
- the region in which the first region 201 and the second region 202 are combined is referred to as a seventh region.
- the control unit may include a first boundary line 211 that divides the first area 201 and the second area 202, and a second boundary line that divides the third area 203 and the fourth area 204. 212, a third boundary line 213 that separates the first area 201 and the third area 204, and a fourth boundary line 214 that separates the second area 202 and the fourth area 204 from each other. Can be.
- the first boundary line 211 and the second boundary line 212 may form one line segment.
- the third boundary line 213 and the fourth boundary line 214 may form one line segment.
- the first boundary line 211 to the fourth boundary line 214 are not necessarily displayed, but may be lines virtually set in the display apparatus 200.
- the controller may set the first area 201 to the fourth area 204 so as not to overlap each other. For example, as illustrated in FIG. 2A, the controller (not shown) may set the first area 201 on the upper left side.
- the controller may set the second region 202 on the upper right side.
- the controller (not shown) may set the third region 203 on the lower left side.
- the controller (not shown) may set the fourth region 204 on the lower right side.
- the controller may set the first boundary line 211 and the second boundary line 212 to divide the screen from side to side.
- the third boundary line 213 and the fourth boundary line 214 may be set to divide the screen up and down.
- the touch screen may display the button 210 at a point where the first boundary line 211, the second boundary line 212, the third boundary line 213, and the fourth boundary line 214 meet.
- the button 210 may be a function key for changing the size of an area displaying an execution window of an application or entering an operation mode for controlling the execution window.
- the controller may control the touch screen to display an execution window of the application in each of the plurality of regions.
- the controller may control the touch screen to display an execution window of an application in each of the plurality of areas 201 to 204 as shown in FIGS. 2B to 2K.
- an execution window of an application means a window displaying an execution screen of an application.
- Objects related to the application may be displayed on the execution screen of the application.
- the object may be implemented in various forms such as text, figure, icon, button, check box, photo, video, web, map, and the like.
- a predetermined function or event may be performed on the object.
- the object may be referred to as a view, depending on the operating system. For example, at least one of a capture button for capturing the execution window for controlling the display of the execution window, a minimize button for minimizing the execution window, a full screen button for maximizing the display of the execution window, and an exit button for exiting the execution window may be displayed. Can be.
- the controller may set an edge area 220 for the touch screen of the display apparatus 200.
- the edge area 220 refers to an area within a predetermined distance from the edge of the touch screen.
- the touch screen may receive a touch input from the user 1 for moving the button 210 displayed on the touch screen of the display apparatus 200. For example, as a touch input for dragging the button 210 is recognized from the user 1, the display apparatus 200 may change the position of the button 210 of the button 210 to a position where the touch input is recognized. .
- the controller may reset and display the areas 201 to 204 and the boundary lines 211 to 214 on which the execution window of the application is displayed.
- the display apparatus 200 may not display areas in which execution windows of some applications are displayed on the touch screen. That is, among the regions 201 to 204 illustrated in FIG. 2B, the region 201 in which the execution window of the A application is displayed and the region 203 in which the execution window of the D application is displayed may not be displayed.
- the display apparatus 200 may divide the screen based on the boundary line 214.
- the display apparatus 200 may display the execution window 202 of the B application and the execution window 204 of the C application on each of the divided screens.
- the display apparatus 200 may move the button 210 to the edge area according to the direction of the fling gesture.
- a fling gesture refers to an input in which a point where a user contacts a touch screen moves with an acceleration greater than or equal to a preset threshold.
- the fling gesture may include information about the direction.
- the controller (not shown) may determine a direction from the start point of the touch input constituting the fling gesture toward the end point of the touch input as the direction of the fling gesture.
- the display apparatus 200 may display the button 210 on the left edge of the touch screen as shown in FIG. 2C.
- the display apparatus 200 may divide a screen based on the position of the button 210 and display an execution window of an application in each of the divided regions.
- the button 210 may be moved in any direction as shown in FIG. 2D as well as up, down, left, and right.
- the display apparatus 200 may display the button 210 at the upper left corner of the touch screen as shown in FIG. 2E.
- the display device may display the execution screen 204 of the C application on the entire area of the touch screen.
- 3A to 3E are conceptual views illustrating a method of moving a position of a displayed button according to one embodiment of the present invention.
- the display apparatus 300 may display a button 310 on a touch screen.
- the controller (not shown) included in the display apparatus 300 may display execution windows of the plurality of applications as illustrated in FIGS. 2A to 2E based on the location of the button 310.
- the controller may divide the touch screen into a plurality of regions 341 to 348 based on the position where the button 310 is displayed, separately from the region where the execution window of the application is displayed.
- the touch screen may be divided into a plurality of areas 341 to 348 by a line segment connecting a plurality of preset points in the edge area of the touch screen from the position displayed on the button 310. have.
- the controller (not shown) included in the display apparatus 300 may include a fling gesture when the acceleration value of the movement of the button is greater than or equal to a preset threshold by a fling gesture input to the touch screen.
- the slope value of may be obtained.
- the inclination value of the fling gesture refers to the inclination value of the line connecting the start point and the end point of the touch input. Based on the obtained inclination value, the controller (not shown) may determine the direction 330 of the fling gesture.
- the controller may control the touch screen to move the button 310 to one of the predetermined positions 351 to 358 based on the direction 330 of the fling gesture.
- the predetermined positions 351 to 358 may be included in an edge region (not shown).
- the predetermined positions 351 to 358 may correspond to the positions 351 and 355 corresponding to the x-axis coordinates on the touch screen on which the button 310 is displayed within an edge region (not shown).
- Positions 353 and 357 corresponding to y-axis coordinates on the touch screen on which 310 is displayed and positions 352, 354, 356, and 358 corresponding to corners of the touch screen may be included.
- the x-axis and the y-axis may vary depending on the embodiment.
- the controller may divide the touch screen into a plurality of regions 341 to 348 based on the position of the button 310 and the position of the button 310. According to an embodiment of the present invention, the controller (not shown) may determine predetermined positions 351 to 358 determined according to the position of the button 310 in the edge area of the touch screen from the position where the button 310 is displayed.
- the touch screen may be divided into a plurality of regions 341 to 348 based on the line segments connecting the. That is, the region may be divided by a boundary line between the middle point of the first point 351 and the second point 352 and the point where the button 310 is displayed.
- the region may be divided by a middle line between the second point 352 and the third point 353 and the point where the button 310 is displayed. In addition, the region may be divided by a middle line between the third point 353 and the fourth point 354 and the point where the button 310 is displayed. In addition, the region may be divided by a middle line between the fourth point 354 and the fifth point 355 and the point where the button 310 is displayed. In addition, the region may be divided by a middle line between the fifth point 355 and the sixth point 356 and the point where the button 310 is displayed. In addition, the region may be divided by a middle line between the sixth point 356 and the seventh point 357 and the point where the button 310 is displayed.
- an area may be divided by a middle line between the seventh point 357 and the eighth point 358 and the point where the button 310 is displayed.
- the region may be divided by a middle line between the eighth point 358 and the first point 351 and the point where the button 310 is displayed.
- the regions 341 to 348 illustrated in FIGS. 3A to 3E are exemplifications for describing the present disclosure.
- the controller may divide the touch screen by another method.
- the touch screen may be divided into more or fewer regions instead of eight regions of the controller (not shown).
- the controller may acquire an inclination value of a line connecting the start point and the end point of the fling gesture. That is, the controller (not shown) may obtain a slope value from the difference between the x coordinate and the y coordinate value between the start point and the end point of the fling gesture.
- the inclination value may be referred to as the inclination value of the touch input or the inclination value of the fling zester.
- the controller may select a region corresponding to the inclination value from the divided regions 341 to 348.
- the controller may select an area 346 including the direction 330 of the fling gesture.
- the controller may move the button 310 to a predetermined position 356 included in the selected area 346.
- the controller may control the touch screen to display the button 310 at the moved position.
- the controller may select the area 347 including the direction 331 of the fling gesture.
- the button 310 may be moved to a predetermined position 357 included in the controller (not shown) selected area 347.
- the controller may control the touch screen to display the button 310 at the moved position.
- the display apparatus 300 moves the button 310 to one of predetermined positions 351 to 358 included in an edge area, thereby executing execution windows 201 to 204 of the application shown in FIG. 2A. Some of the execution windows may not be displayed. Accordingly, some of the execution windows 201 to 204 of the application illustrated in FIG. 2A may be enlarged and displayed.
- the controller may track the position of the touch input continuously input to the touch screen or the position of the button 310. As the position of the touch input or the position of the button 310 input to the touch screen is moved, the controller (not shown) may acquire an acceleration value. Thereafter, the controller (not shown) may determine that the touch input received through the touch screen is a fling gesture when the acquired acceleration value is greater than or equal to a preset threshold. For example, when the user moves the button 310 and moves more than 50 pixels per second while the coordinates of the button 310 are stationary, the controller (not shown) indicates that the received touch input is a fling gesture. You can judge.
- the controller may acquire an inclination value of the fling gesture and move the button 310 to a predetermined position.
- the acceleration is referred to herein, but the acceleration may be variously changed according to the embodiment in terms of speed, displacement, or vector.
- FIG. 4 is a flowchart illustrating a process of displaying a button according to an embodiment of the present invention.
- the display device may display a button on the touch screen (S410) and divide the touch screen into a plurality of areas based on the displayed position of the button (S420).
- the display device may receive a touch input for moving a button through the touch screen (S430).
- the touch input for moving the button may be a drag gesture in which the touch recognition position moves from the position where the button is displayed while the touch recognition is maintained.
- the display device may acquire a tilt value of the touch input (S440).
- the display device may select an area corresponding to the inclination value obtained in step S440 among the areas divided in step S420 (S450).
- the display device may move the button to a preset position included in the area selected in operation S450 (S460).
- FIG. 5 is a flowchart illustrating a process of recognizing a fling gesture according to an embodiment of the present invention.
- the display device may acquire an acceleration value of the movement of the button (S510).
- the acceleration value of the movement of the button may obtain the acceleration value by calculating the coordinate value of the touch input (S510).
- the display device may determine whether the acceleration value is greater than or equal to a preset threshold value (S520). If the acceleration value is greater than or equal to a preset threshold, the display device may determine that the touch input is a fling gesture. When the acceleration value is greater than or equal to a preset threshold value, the display device may acquire a tilt value of the touch input (that is, a tilt value of the fling gesture) as in steps S440 through S460. In addition, the display device may move the button based on the tilt value.
- 6A through 6C are conceptual diagrams for describing a method of determining the direction of fling.
- the display apparatus 600 may obtain the inclination value of the fling gesture from the start point and the end point of the touch input 610 input by the user on the touch screen.
- the inclination value of the fling gesture indicates the direction 630-1 of the fling gesture recognized by the display apparatus 600.
- the display apparatus 600 may determine the coordinate of the touch input before the preset number as the start point of the touch input from the end point pk of the touch input.
- the touch input 610 may include k coordinates p1 to pk received through the touch screen.
- the control unit (not shown) included in the display apparatus 600 includes a preset number? From the end point pk of the touch input. N for? The previous coordinate (pk-n) may be determined as the starting point of the touch input.
- the display apparatus 600 may obtain a direction of a fling gesture in accordance with a user's intention by determining a coordinate (pk-n) before the preset number from the end point pk as a start point of the touch input. have.
- FIG. 7 is a block diagram schematically illustrating a configuration of a display apparatus according to an exemplary embodiment.
- the display apparatus 700 includes a touch screen 710 displaying a button and a controller 720 dividing the touch screen 710 into a plurality of areas.
- the touch screen 710 receives a touch input for moving the displayed button.
- the controller 720 may acquire an inclination value of a line connecting the start point and the end point of the touch input based on the touch input.
- the controller 720 may select an area corresponding to an inclination value from among the plurality of divided areas.
- the controller 720 may control the touch screen to move and display the button to a predetermined position included in the selected area.
- the controller 720 may acquire the acceleration value of the movement of the button displayed on the touch screen 710. If the acceleration value is greater than or equal to a preset threshold, the controller 720 may control the touch screen 710 to display a button at a predetermined position.
- the predetermined position may be included in an edge area of the touch screen 710.
- the controller 720 divides the touch screen 710 into a plurality of areas by a line segment connecting a plurality of preset points of an edge area of the touch screen 710 from the position where the button is displayed. can do.
- the controller 720 may determine the coordinate of the touch input before the preset number as the start point of the touch input from the end point of the touch input.
- the display device may include various types of touch screens.
- the display device may include a touch screen having an n-shape (n is an integer of 3 or more).
- the display device may divide the touch screen into 2 * n areas based on the position of the button, and move the position of the button according to the direction of the fling gesture.
- FIG. 8 is a conceptual diagram illustrating a method of recognizing a fling gesture by a display apparatus according to another exemplary embodiment.
- the display device includes the pentagonal touch screen 800
- the display device recognizes a fling gesture.
- the display device may include the touch screen 800 in ten areas 820-1, 820-2, 820-3, 820-4, 820-5, and 820-. 6, 820-7, 820-8, 820-9, and 820-10).
- the display device may determine the direction 830 of the fling gesture based on the button 810.
- the display device determines the location of the button 810 in ten divided areas 820-1, 820-2, 820-3, 820-4, 820-5, and 820-.
- 6, 820-7, 820-8, 820-9, and 820-10 may be moved according to an area including the direction 830 of the determined fling gesture.
- the display device may move the position of the button 810 to the point 842-1. .
- the display device may move the position of the button 810 to the point 841-1.
- the display device may move the position of the button 810 to the point 843-1.
- the display device may move the position of the button 810 to the point 841-2.
- the display device may move the position of the button 810 to the point 842-2.
- the display device may move the position of the button 810 to the point 841-3.
- the display device may move the position of the button 810 to the point 842-3.
- the display device may move the position of the button 810 to the point 841-4.
- the display device may move the position of the button 810 to the point 842-4.
- FIG. 8 is a diagram for describing an exemplary embodiment, wherein the number of regions in which the touch screen 800 is divided and the points 81-1 to 841-5 and 842-1 to 842-5 to which the button 810 is to be moved are illustrated. The location may vary depending on the embodiment.
- FIG. 9 is a conceptual diagram illustrating a method of recognizing a fling gesture by a display apparatus according to another exemplary embodiment.
- the display device includes the circular touch screen 900, the display device recognizes a fling gesture.
- the display device may determine the direction 930 of the fling gesture.
- the display device may determine a point 940 on the outside of the display device based on the line segment 920 extending in the direction 930 of the fling gesture determined from the position of the button 910.
- the display device may move the position of the button 910 to the determined point 940.
- FIG. 9 is for describing some embodiments, but is not limited thereto.
- FIGS. 10 to 14 are exemplary views illustrating an embodiment in which the display apparatus is a wearable device.
- FIGS. 10 to 12 are exemplary views illustrating an operation of the display apparatus 1000 having a circular touch screen.
- the display apparatus 1000 may display a button 1010 on a touch screen.
- the display apparatus 1000 may display a user interface or an application execution screen on divided areas of the touch screen based on the button 1010.
- the user 1 may input a fling gesture for the button 1010 through the touch screen of the display apparatus 1000.
- the display apparatus 1000 may determine a point 1040 on the outside of the touch screen based on the line segment 1020 extending in the direction 1030 of the fling gesture.
- the display apparatus 1000 may move the position of the button 1010 to the point 1040 of FIG. 11.
- the display apparatus 1000 may divide the touch screen based on the position of the moved button 1010 and display a user interface or an application execution screen on each of the divided regions.
- FIG. 13 and 14 illustrate exemplary operations of the display apparatus 1300 having a rectangular touch screen.
- the display device 1300 may include a rectangular touch screen.
- the display device 1300 may display the button 1310 on the touch screen.
- the display device 1300 may display a user interface or an application execution screen on divided regions of the touch screen based on the button 1310.
- the display apparatus 1300 may move the position of the button 1310 based on the direction of the fling gesture.
- the display device 1300 displays a user interface or an application execution screen on the entire area of the touch screen based on the position of the moved button 1310 or displays a user interface or application execution screen on each of the divided areas on the touch screen. Etc. can be displayed.
- the display apparatus may provide a user interface for easily adjusting an area in which execution windows of a plurality of applications are displayed to the user according to the user's intention.
- Computer readable media can be any available media that can be accessed by a computer and includes both volatile media such as RAM and nonvolatile media such as ROM, removable and non-removable media.
- Computer readable media may include both computer storage media and communication media.
- Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data.
- Communication media typically includes computer readable instructions, data structures, program modules, or other data in a modulated data signal, or other transmission mechanism, and includes any information delivery media.
- the computer storage medium may be implemented with a ROM, RAM, flash memory, CD, DVD, magnetic disk or magnetic tape.
- Displaying a button on the touch screen dividing the touch screen into a plurality of areas based on the position at which the button is displayed, receiving a touch input for moving the displayed button, and starting and ending points of the touch input. Acquiring an inclination value of a line segment connecting the same, selecting an area corresponding to the inclination value among the plurality of divided regions, and moving a button to a predetermined position included in the selected region.
- the program may be recorded in a storage medium according to an embodiment of the present invention.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
Claims (11)
- 터치스크린을 구비하는 디스플레이 장치가 화면을 디스플레이 방법에 있어서,상기 터치스크린 상에, 버튼을 표시하는 단계;상기 버튼이 표시되는 위치에 기초하여, 상기 터치스크린을 복수의 영역으로 분할하는 단계;상기 표시된 버튼을 이동시키는 터치 입력을 수신하는 단계;상기 터치 입력의 시작점 및 종료점을 잇는 선분의 기울기 값을 획득하는 단계;상기 분할된 복수의 영역 중에서, 상기 기울기 값에 상응하는 영역을 선택하는 단계; 및상기 선택된 영역에 포함된 소정의 위치(position)로 상기 버튼을 이동시키는 단계를 포함하는, 디스플레이 방법.
- 제 1 항에 있어서,상기 기울기 값을 획득하는 단계는,상기 버튼의 움직임의 가속도 값을 획득하는 단계; 및상기 가속도 값이 기 설정된 임계값 이상인 경우, 상기 기울기 값을 획득하는 단계를 포함하는, 디스플레이 방법.
- 제 1 항에 있어서,상기 소정의 위치는,상기 터치스크린의 에지(edge) 영역에 포함되는, 디스플레이 방법.
- 제 1 항에 있어서,상기 터치스크린을 상기 복수의 영역으로 분할하는 단계는,상기 버튼이 표시된 위치로부터 상기 터치스크린의 에지(edge) 영역의 기 설정된 복수의 지점을 잇는 선분으로 상기 터치스크린을 복수의 영역으로 분할하는 단계를 포함하는, 디스플레이 방법.
- 제 1 항에 있어서,상기 기울기 값을 획득하는 단계는,상기 터치 입력의 상기 종료점으로부터 기 설정된 개수 이전의 상기 터치 입력의 좌표를 상기 터치 입력의 시작점으로 결정하는 것을 특징으로 하는, 디스플레이 방법.
- 버튼을 표시하는 터치스크린; 및상기 버튼이 표시되는 위치에 기초하여, 상기 터치스크린을 복수의 영역으로 분할하는 제어부; 를 포함하고,상기 터치스크린은, 상기 표시된 버튼을 이동시키는 터치 입력을 수신하며,상기 제어부는, 상기 터치 입력에 기초하여 상기 터치 입력의 시작점 및 종료점을 잇는 선분의 기울기 값을 획득하고, 상기 분할된 복수의 영역 중에서 상기 기울기 값에 상응하는 영역을 선택하며, 상기 선택된 영역에 포함된 소정의 위치(position)로 상기 버튼을 이동시켜 표시하도록 상기 터치스크린을 제어하는 것을 특징으로 하는, 디스플레이 장치.
- 제 6 항에 있어서,상기 제어부는,상기 버튼의 움직임의 가속도 값을 획득하고, 상기 가속도 값이 기 설정된 임계값 이상인 경우, 상기 버튼을 상기 소정의 위치에 표시되도록 상기 터치스크린을 제어하는, 디스플레이 장치.
- 제 6 항에 있어서,상기 소정의 위치는,상기 터치스크린의 에지(edge) 영역에 포함되는, 디스플레이 장치.
- 제 6 항에 있어서,상기 제어부는,상기 버튼이 표시된 위치로부터 상기 터치스크린의 에지(edge) 영역의 기 설정된 복수의 지점을 잇는 선분으로 상기 터치스크린을 복수의 영역으로 분할하는 것을 특징으로 하는, 디스플레이 장치.
- 제 6 항에 있어서,상기 제어부는,상기 터치 입력의 상기 종료점으로부터 기 설정된 개수 이전의 상기 터치 입력의 좌표를 상기 터치 입력의 시작점으로 결정하는, 디스플레이 장치.
- 제 1 항의 방법을 컴퓨터에서 실행시키기 위한 프로그램을 기록한 컴퓨터로 읽을 수 있는 기록매체.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020167004571A KR102191972B1 (ko) | 2013-08-22 | 2014-08-13 | 디스플레이 장치가 화면을 디스플레이 하는 방법 및 그 디스플레이 장치 |
US14/912,783 US10564843B2 (en) | 2013-08-22 | 2014-08-13 | Display device and method of displaying screen on said display device |
US16/730,284 US11042294B2 (en) | 2013-08-22 | 2019-12-30 | Display device and method of displaying screen on said display device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR20130099928 | 2013-08-22 | ||
KR10-2013-0099928 | 2013-08-22 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/912,783 A-371-Of-International US10564843B2 (en) | 2013-08-22 | 2014-08-13 | Display device and method of displaying screen on said display device |
US16/730,284 Continuation US11042294B2 (en) | 2013-08-22 | 2019-12-30 | Display device and method of displaying screen on said display device |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015026099A1 true WO2015026099A1 (ko) | 2015-02-26 |
Family
ID=52483833
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2014/007540 WO2015026099A1 (ko) | 2013-08-22 | 2014-08-13 | 디스플레이 장치가 화면을 디스플레이 하는 방법 및 그 디스플레이 장치 |
Country Status (3)
Country | Link |
---|---|
US (2) | US10564843B2 (ko) |
KR (1) | KR102191972B1 (ko) |
WO (1) | WO2015026099A1 (ko) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106155516A (zh) * | 2015-04-28 | 2016-11-23 | 阿里巴巴集团控股有限公司 | 操作按钮展现方法及设备 |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9652124B2 (en) | 2014-10-31 | 2017-05-16 | Microsoft Technology Licensing, Llc | Use of beacons for assistance to users in interacting with their environments |
USD779556S1 (en) * | 2015-02-27 | 2017-02-21 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with an icon |
DE102017113202A1 (de) * | 2017-06-15 | 2018-12-20 | Borgward Trademark Holdings Gmbh | Bildschirmaufteilungsverfahren, zugehörige Einrichtung und zugehöriges Fahrzeug |
EP3667484A1 (en) * | 2017-08-22 | 2020-06-17 | Samsung Electronics Co., Ltd. | Electronic device and control method thereof |
JP7093690B2 (ja) * | 2018-07-05 | 2022-06-30 | フォルシアクラリオン・エレクトロニクス株式会社 | 情報制御装置、及び表示変更方法 |
USD924907S1 (en) * | 2019-01-31 | 2021-07-13 | Salesforce.Com, Inc. | Display screen or portion thereof with graphical user interface |
USD913319S1 (en) * | 2019-01-31 | 2021-03-16 | Salesforce.Com, Inc. | Display screen or portion thereof with graphical user interface |
JP7331465B2 (ja) * | 2019-05-29 | 2023-08-23 | セイコーエプソン株式会社 | 表示装置の制御方法、及び表示装置 |
WO2021137319A1 (ko) * | 2019-12-31 | 2021-07-08 | 엘지전자 주식회사 | 애플리케이션과 관련된 컨텐츠를 표시하는 전자 장치 및 그 제어 방법 |
WO2022014740A1 (ko) * | 2020-07-15 | 2022-01-20 | 엘지전자 주식회사 | 이동 단말기 및 그 제어 방법 |
US11703990B2 (en) * | 2020-08-17 | 2023-07-18 | Microsoft Technology Licensing, Llc | Animated visual cues indicating the availability of associated content |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20070001771A (ko) * | 2005-06-29 | 2007-01-04 | 정순애 | 화면정보 제어방법 |
US20070101289A1 (en) * | 2005-10-27 | 2007-05-03 | Awada Faisal M | Maximizing window display area using window flowing |
KR20110063409A (ko) * | 2011-05-11 | 2011-06-10 | 주식회사 케이티테크 | 화면 분할 표시 장치 및 방법 |
US20110175930A1 (en) * | 2010-01-19 | 2011-07-21 | Hwang Inyong | Mobile terminal and control method thereof |
US20120244913A1 (en) * | 2009-10-30 | 2012-09-27 | Cellco Partnership D/B/A Verizon Wireless | Flexible home page layout for mobile devices |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5265159B2 (ja) * | 2007-09-11 | 2013-08-14 | 株式会社バンダイナムコゲームス | プログラム及びゲーム装置 |
JP4683110B2 (ja) * | 2008-10-17 | 2011-05-11 | ソニー株式会社 | 表示装置、表示方法及びプログラム |
US8302026B2 (en) | 2008-11-28 | 2012-10-30 | Microsoft Corporation | Multi-panel user interface |
US20110210850A1 (en) * | 2010-02-26 | 2011-09-01 | Phuong K Tran | Touch-screen keyboard with combination keys and directional swipes |
JP5665391B2 (ja) * | 2010-07-02 | 2015-02-04 | キヤノン株式会社 | 表示制御装置及び表示制御装置の制御方法 |
JP5269851B2 (ja) | 2010-09-27 | 2013-08-21 | 富士フイルム株式会社 | 画像編集装置及びその画像編集方法並びにそのプログラム |
KR20130064458A (ko) * | 2011-12-08 | 2013-06-18 | 삼성전자주식회사 | 복수 개의 영역으로 구분된 화면을 디스플레이하는 디스플레이 장치 및 그 방법 |
US9507512B1 (en) * | 2012-04-25 | 2016-11-29 | Amazon Technologies, Inc. | Using gestures to deliver content to predefined destinations |
-
2014
- 2014-08-13 WO PCT/KR2014/007540 patent/WO2015026099A1/ko active Application Filing
- 2014-08-13 US US14/912,783 patent/US10564843B2/en active Active
- 2014-08-13 KR KR1020167004571A patent/KR102191972B1/ko active IP Right Grant
-
2019
- 2019-12-30 US US16/730,284 patent/US11042294B2/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20070001771A (ko) * | 2005-06-29 | 2007-01-04 | 정순애 | 화면정보 제어방법 |
US20070101289A1 (en) * | 2005-10-27 | 2007-05-03 | Awada Faisal M | Maximizing window display area using window flowing |
US20120244913A1 (en) * | 2009-10-30 | 2012-09-27 | Cellco Partnership D/B/A Verizon Wireless | Flexible home page layout for mobile devices |
US20110175930A1 (en) * | 2010-01-19 | 2011-07-21 | Hwang Inyong | Mobile terminal and control method thereof |
KR20110063409A (ko) * | 2011-05-11 | 2011-06-10 | 주식회사 케이티테크 | 화면 분할 표시 장치 및 방법 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106155516A (zh) * | 2015-04-28 | 2016-11-23 | 阿里巴巴集团控股有限公司 | 操作按钮展现方法及设备 |
Also Published As
Publication number | Publication date |
---|---|
US10564843B2 (en) | 2020-02-18 |
US20200142586A1 (en) | 2020-05-07 |
KR102191972B1 (ko) | 2020-12-16 |
US11042294B2 (en) | 2021-06-22 |
KR20160045715A (ko) | 2016-04-27 |
US20160196057A1 (en) | 2016-07-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2015026099A1 (ko) | 디스플레이 장치가 화면을 디스플레이 하는 방법 및 그 디스플레이 장치 | |
WO2014017790A1 (en) | Display device and control method thereof | |
WO2013176515A1 (en) | Multiple display method with multiple communication terminals, machine-readable storage medium and communication terminal | |
WO2013103275A1 (en) | Method and apparatus for implementing multi-vision system by using multiple portable terminals | |
WO2014193101A1 (en) | Method and apparatus for controlling screen display using environmental information | |
WO2014017722A1 (en) | Display device for executing multiple applications and method for controlling the same | |
WO2014088310A1 (en) | Display device and method of controlling the same | |
WO2015030488A1 (en) | Multi display method, storage medium, and electronic device | |
WO2014129862A1 (en) | Method for controlling display of multiple objects depending on input related to operation of mobile terminal, and mobile terminal therefor | |
WO2018021696A1 (en) | Image display apparatus and method of displaying image | |
WO2013073908A1 (en) | Apparatus with touch screen for preloading multiple applications and method of controlling the same | |
WO2014088343A1 (en) | Display device and method of controlling the same | |
WO2015046809A1 (en) | Method for displaying previews in a widget | |
WO2014196840A1 (en) | Portable terminal and user interface method in portable terminal | |
WO2015174612A1 (ko) | 이동단말기 및 그 제어방법 | |
WO2013073890A1 (en) | Apparatus including a touch screen under a multi-application environment and controlling method thereof | |
WO2014189223A1 (en) | Apparatus, method and computer readable recording medium for displaying thumbnail image of panoramic photo | |
WO2014157894A1 (en) | Display apparatus displaying user interface and method of providing the user interface | |
WO2015005732A1 (en) | Method of sharing electronic document and devices for the same | |
WO2013191488A1 (en) | Apparatus including a touch screen and screen change method thereof | |
WO2015005628A1 (en) | Portable device for providing combined ui component and method of controlling the same | |
WO2014030956A1 (en) | Apparatus for uploading contents, user terminal apparatus for downloading contents, server, contents sharing system and their contents sharing method | |
WO2015174632A1 (en) | Method of controlling mobile terminal using fingerprint recognition and mobile terminal using the same | |
WO2014163333A1 (ko) | 사용자 인터페이스 표시 방법 및 장치 | |
WO2015020288A1 (en) | Display apparatus and the method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14838086 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14912783 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 20167004571 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 14838086 Country of ref document: EP Kind code of ref document: A1 |