WO2014059886A1 - 获取图像的方法和装置 - Google Patents

获取图像的方法和装置 Download PDF

Info

Publication number
WO2014059886A1
WO2014059886A1 PCT/CN2013/084868 CN2013084868W WO2014059886A1 WO 2014059886 A1 WO2014059886 A1 WO 2014059886A1 CN 2013084868 W CN2013084868 W CN 2013084868W WO 2014059886 A1 WO2014059886 A1 WO 2014059886A1
Authority
WO
WIPO (PCT)
Prior art keywords
filter
instruction
image
gesture
gesture operation
Prior art date
Application number
PCT/CN2013/084868
Other languages
English (en)
French (fr)
Inventor
李�诚
郑晖
Original Assignee
腾讯科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 腾讯科技(深圳)有限公司 filed Critical 腾讯科技(深圳)有限公司
Publication of WO2014059886A1 publication Critical patent/WO2014059886A1/zh
Priority to US14/688,123 priority Critical patent/US9674425B2/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04845Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04847Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/62Control of parameters via user interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/631Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/2621Cameras specially adapted for the electronic generation of special effects during image pickup, e.g. digital cameras, camcorders, video cameras having integrated special effects capability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/2628Alteration of picture size, shape, position or orientation, e.g. zooming, rotation, rolling, perspective, translation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders

Definitions

  • the present invention relates to the field of video processing, and in particular, to a method and apparatus for acquiring an image. Background of the invention
  • a filter is an image processing technique that can quickly modify or beautify an image taken by a user.
  • Various filter operations can be performed by installing an application in a terminal device such as a mobile terminal or a computer.
  • the application provides multiple filters and pre-sets the parameters for each filter. After the user selects a filter, you can directly view the filter effect displayed by the terminal device, that is, the video or picture processed by the filter.
  • the embodiment of the present application provides a method and apparatus for acquiring an image, which can perform filtering processing on an image according to a user's gesture, thereby obtaining a rich filter effect.
  • a method of acquiring an image comprising the steps of:
  • An apparatus for acquiring an image comprising:
  • Receiving unit a gesture operation for receiving an input, the operation information acting on an image acquired by the camera;
  • An obtaining unit configured to determine a filter parameter according to a current filter type and the gesture operation;
  • an execution unit configured to perform, according to the filter parameter, a filter operation corresponding to the filter type on the image;
  • Output unit Used to display in real time the image acquired after the filter operation is performed.
  • the terminal device can receive a gesture operation, determine a filter parameter according to the current filter type and the gesture operation, and perform a filter operation corresponding to the current filter type according to the filter parameter.
  • the current filter type may be at least one filter type preset setting, or may be at least one filter type determined according to selection information input by the user. Therefore, when the filter operation is performed, the filter operation can be adjusted in real time according to the user's gesture operation, thereby obtaining a rich filter effect.
  • FIG. 1 is a flowchart of an implementation of a method for acquiring an image according to an embodiment of the present application.
  • FIGS. 2a, 2b, and 2c are schematic diagrams showing effects of adjusting a vortex filter by a rotation and sliding gesture according to an embodiment of the present application.
  • 3a, 3b, and 3c are schematic diagrams showing effects of adjusting a vortex filter by a rotation and zoom gesture according to an embodiment of the present application.
  • FIG. 4 is a flowchart of an implementation of a method for acquiring an image according to an embodiment of the present application.
  • 5a and 5b are schematic diagrams showing an effect of adjusting a fisheye filter by a zoom gesture according to an embodiment of the present application.
  • 6a, 6b, and 6c are schematic diagrams showing effects of controlling a color value of a color filter by a sliding gesture according to an embodiment of the present application.
  • 7a and 7b are schematic diagrams showing effects of changing a color by using a sliding gesture according to an embodiment of the present application.
  • FIG. 8 is a schematic structural diagram of a graphics system of an OpenGL ES 2.0 for acquiring an image according to an embodiment of the present application.
  • FIG. 9 is a schematic diagram of real-time rendering implemented by combining hardware and software provided by an embodiment of the present application.
  • FIG. 10 is a schematic flowchart of a picture processing provided by an embodiment of the present application.
  • FIG. 11 is a structural block diagram of an apparatus for acquiring an image according to an embodiment of the present application.
  • FIG. 12 is a structural block diagram of an apparatus for performing picture processing according to an embodiment of the present application. Mode for carrying out the invention
  • a gesture operation for an image is received, a filter parameter is obtained according to the gesture operation, and a filter operation is performed according to the filter parameter, thereby obtaining a rich filter effect.
  • FIG. 1 is a flowchart showing an implementation process of acquiring an image provided by an embodiment of the present application. As shown in Figure 1, the method includes the following steps.
  • step S101 the terminal device receives an input gesture operation for an image acquired through a camera.
  • the image acquired by the camera is a video or a picture.
  • the terminal device receives the gesture operation through the input device.
  • the input device may be a touch screen.
  • the user inputs a gesture operation through the touch screen. Since the touch screen supports rich gesture operations, for mobile terminal devices, For example, mobile phones, PADs, etc., users can enter gestures flexibly and conveniently through the touch screen.
  • a gesture operation input by a user through a touch screen may correspond to a slide instruction, a stretch instruction, a rotation instruction, a click or double-click instruction, and the like.
  • the input device may be at least one of a keyboard or a mouse or a trackball.
  • the user input gesture operation may correspond to a drag instruction, a click instruction, a double-click instruction, a right-click instruction, a rotation instruction, and the like.
  • step S102 the terminal device determines a filter parameter of the filter operation according to the current filter type and the gesture operation.
  • the terminal device may adopt at least one filter type preset as the current filter type, such as at least one of a fisheye filter, a color filter, and a vortex filter.
  • the terminal device can also determine the current filter type according to the user's selection. For example, the terminal device first provides the user with an alternate plurality of filter types, receives one or more filter types selected by the user, and uses one or more filter types selected by the user as the current filter type.
  • the filter parameters obtainable according to the gesture operation include at least one or more of an operation command, a filter intensity, a filter color, a filter size, and a filter transparency.
  • the filter parameter may further include filter coordinate information.
  • the operation commands corresponding to the same gesture operation are also different.
  • the sliding command, the stretching command, and the rotation command input through the touch screen correspond to different operation commands when different types of filters are used.
  • the correspondence between the filter type, the gesture operation, and the operation instruction is pre-stored in the terminal device.
  • the terminal device may also receive and store a filter type set by the user, a correspondence relationship between the gesture operation and the operation instruction. For example, the user sets some favorite gesture operations corresponding to commonly used operation instructions.
  • the operation instruction corresponding to the gesture operation may be a slide instruction, a stretch instruction, a rotation instruction, a click instruction, or a double-click instruction.
  • the operation instruction corresponding to the gesture operation may be a drag instruction, a click instruction, a double-click instruction, a right-click instruction, a rotation instruction, or the like.
  • filter parameters such as filter strength, filter color, filter size, filter transparency, etc.
  • filter parameters may also be determined based on the filter type and the magnitude of the gesture operation.
  • the amplitude of the gesture operation can be obtained according to the filter coordinate information.
  • the filter parameters may be determined based on the stretched length of the gesture operation or the size of the zoom.
  • the filter intensity in the black and white filter such as the depth of black
  • the filter color can be determined based on the magnitude of the slide.
  • the filter type is Fisheye Filter and the gesture is Stretch
  • the filter size can be determined based on the extent of the stretch.
  • the position touched by the finger is filter coordinate information.
  • the input device is at least one of a mouse, a keyboard, and a trackball
  • the position of the pointer during the gesture operation is filter coordinate information.
  • the filter coordinate information may also be a preset position at which the filter operation is performed by default.
  • step S103 according to the filter parameter, a filter operation corresponding to the current filter type is performed to obtain a filter effect of the image.
  • FIGS. 2a, 2b, and 2c are schematic diagrams showing effects of adjusting a vortex filter by a rotation and sliding gesture according to an embodiment of the present application.
  • Figure 2a is an image screen obtained after conventional vortex filter processing.
  • Figure 2b shows the filter effect after the stretch command is executed with the preset filter type vortex filter.
  • the gesture operation input by the user through the touch screen is as shown in FIG. 2b.
  • the vortex filter of the gesture operation corresponds to the stretching instruction.
  • the filter emphasis can be determined, that is, the vortex is enlarged.
  • the multiple based on the position of the finger in the gesture operation, determines the position at which the vortex filter is executed. In the picture
  • FIG. 2b the vortex portion corresponding to the gesture operation is enlarged according to the stretching instruction.
  • Fig. 2c shows the effect after the stretch command is executed in the case where the preset filter type is a vortex filter, and then the slide command is executed.
  • the vortex in Figure 2b is moved to the position corresponding to the sliding command.
  • FIG. 3a, FIG. 3b, and FIG. 3c are schematic diagrams showing the effect of changing the vortex filter by rotating and scaling the hand posture according to an embodiment of the present application.
  • Figure 3a is an image of a vortex filter operation performed.
  • the user can change the intensity of the vortex filter operation in Figure 3a by entering the rotation gesture as shown in Figure 3b.
  • the position where the intensity changes is the position of the rotation gesture, and the degree of the intensity change is determined according to the magnitude of the rotation gesture, thereby obtaining an effect diagram as shown in FIG. 3b.
  • the user can zoom in on the vortex portion in Figure 3b by entering the Stretch gesture. Among them, the degree of magnification is determined according to the magnitude of the stretching hand, so that the effect illustrated in Fig. 3c is obtained on the basis of Fig. 3b.
  • the parameters of the filter operation can be flexibly changed according to the input gesture operation.
  • different users input different instructions to obtain different filter effects.
  • different users input the same command, but the order of inputting the same command is different, and different filter effects are also obtained.
  • step S104 the image obtained after the execution of the filter operation is displayed in real time through the output device.
  • a plurality of filter operations can be performed in step S103, and a final filter effect can be obtained.
  • the final filter effect is displayed in real time in this step.
  • the filter effect acquired in step S103 is displayed through the display screen.
  • an image can be displayed by a touch screen that receives a gesture operation.
  • smart phones, tablets, etc. can provide a touch screen function, receive gesture operations, and display images processed by the filter.
  • it can also be used for other devices such as desktop computers.
  • the devices are merely exemplified herein, and the present invention is not limited to the present application.
  • the method of acquiring an image described in the present application can be applied to other devices in which the head performs screen presentation.
  • FIG. 4 is a flowchart showing an implementation process of acquiring an image provided by an embodiment of the present application. As shown in FIG. 4, the method includes the following steps.
  • step S401 the terminal device receives the input selection information, and determines the current filter type based on the selection information.
  • one or more filter types can be determined as the current filter type, thereby enriching the filter effect obtained after the filter processing.
  • step S402 the terminal device receives an input gesture operation that acts on the image acquired by the camera.
  • step S403 the terminal device determines a filter operation filter parameter according to the current filter type and the gesture operation.
  • step S401 Since the filter type is selected in step S401 in this embodiment, the filter type selected in step S401 is directly used in step S403.
  • step S404 a filter operation corresponding to the filter type is performed according to the filter parameters.
  • the filter parameters include one or more of filter intensity, filter color, filter size, and filter transparency.
  • the filter parameters may further include filter coordinate information, and the filter operation corresponding to the filter parameters is performed at a position corresponding to the filter coordinate information in the image.
  • the filter parameters may be acquired according to an operation amplitude value and/or a gesture operation type of the gesture operation. For example, stretching length or scaling based on gestures
  • the size corresponds to the corresponding filter parameter, or corresponds to the corresponding filter parameter according to the operation instruction such as the zoom instruction, the smoothing instruction, the stretching instruction, and the like.
  • the filter coordinate information may be acquired at the position of the image according to the gesture operation, and the position of the filter effect to be converted may be flexibly obtained, or may be set to a fixed number of positions, which is not limited herein.
  • the correspondence between the gesture operation and the filter operation determining a filter operation corresponding to the gesture operation in the current filter type; according to the filter type and the filter parameter , Perform a filter operation on the image.
  • the filter parameters required for the filter operation may be determined according to the operation amplitude value of the gesture operation. For example, when the gesture input through the touch screen is a sliding operation, the filter parameters are determined according to the distance of the sliding distance, for example, determining the depth of black in the black and white filter. Therefore, the amplitude value of the gesture operation changes the relevant parameters in the filter operation, thereby further improving the diversity of the filter operation and obtaining a richer filter effect.
  • the following is an example of a filter effect obtained by performing a filter operation after receiving a gesture operation input through a touch screen in a fisheye filter mode.
  • FIG. 5a and 5b are schematic diagrams showing an effect of adjusting a fisheye filter by a zoom gesture according to an embodiment of the present application.
  • the original picture before the filter operation is performed as shown in Fig. 5a.
  • the size of the QQ figurine in Figure 5a matches the actual size.
  • Figure 5b is an effect diagram obtained after the filter operation is performed.
  • the operation command corresponding to the gesture operation is a zoom command, which can adjust the degree of distortion of the fisheye filter.
  • the fisheye filter effect corresponding to the zoom level is executed according to the degree of zoom indicated by the gesture operation. For example, the gesture operation indicates a slight zoom, and a slight distortion is performed at the position corresponding to the filter coordinate information in the image.
  • FIGS. 6a, 6b, and 6c are schematic diagrams showing effects of controlling a color value filtered by a color filter by a sliding gesture according to an embodiment of the present application.
  • the filter filter operation first determine the hand in the swipe gesture The color interval corresponding to the potential expansion amplitude filters the color other than the color interval and performs gradation processing.
  • the swipe gesture the red vase, the green book cover and the blue magazine inner page, the non-displayed items and other images are grayscaled in the figures of Figs. 6a, 6b, and 6c, respectively.
  • FIG. 7a and 7b are schematic diagrams showing an effect of changing a color by a sliding gesture according to an embodiment of the present application.
  • Figure 7a shows the original image without the filter operation. After receiving the swipe gesture input through the touch screen, the proportion of the color in Fig. 7a is changed, thereby adjusting the color display of the screen to realize the filter effect of the Washington filter.
  • step S405 the image obtained after the filter operation is performed is displayed in real time through the output device.
  • a plurality of filter operations can be performed in step S404, and a final filter effect can be obtained.
  • the final filter effect is displayed in real time in this step.
  • the filter effect acquired in step S403 is displayed through the display screen.
  • an image can be displayed by a touch screen that receives a gesture operation.
  • smartphones, tablets, etc. can provide a touch screen function, receive gesture operations, and display images processed by the filter.
  • it can also be used for other devices such as desktop computers.
  • the devices are merely exemplified herein, and the present invention is not limited thereto.
  • the present application can be applied only to the specific devices described above, but any other device that has a camera for screen presentation can be applied. A method of acquiring an image as described herein.
  • the output device uses the OpenGL ES technology to display the screen obtained after the filter operation in real time through a Graphic Processing Unit (GPU).
  • GPU Graphic Processing Unit
  • the processing process that maximizes the time spent by the filter is given to
  • OpenGL is done, creating and writing a shader shader in the shader.
  • OpenGL compiles the shader and links it to the program.
  • the original program acquires image data, and transforms each pixel of the image through a shader shader placed in the OpenGL pipeline flow.
  • the whole process is parallel, and the shader language is used to program and process the transformation, and the terminal GPU is used to perform the calculation, which greatly improves the calculation.
  • FIG. 10 is a schematic flowchart of a picture processing provided by an embodiment of the present application. As shown in FIG. 10, the method includes the following steps.
  • step S1001 the terminal device receives an input gesture operation, which is used on the picture to be processed.
  • step S1002 the terminal device determines a filter parameter of the filter operation according to the current filter type and the gesture operation.
  • step S1003 the terminal device operates on the filter corresponding to the filter type according to the filter parameter.
  • the filter parameter may include filter coordinate information, and the filter operation corresponding to the filter parameter is performed at a position corresponding to the filter coordinate information in the picture.
  • the filter parameters include at least one of a filter type, a filter intensity, a filter color, a filter size, and a filter transparency.
  • the filter parameters may be acquired according to the operation amplitude value and/or the gesture operation type of the gesture operation, and the filter coordinate information may be acquired at the position of the image according to the gesture operation.
  • the operation instruction corresponding to the gesture operation includes at least one of a slide instruction, a stretch instruction, a rotation instruction, a click instruction, and a double-click instruction.
  • step S1004 the terminal device displays a picture acquired after the filter operation is performed.
  • the method includes: receiving input selection information, and determining a current filter type according to the selection information.
  • one or more filter types may be determined as the current filter type, thereby enriching the filter effect obtained after the filter processing.
  • the steps in the embodiment of the present application are basically the same as those in the foregoing embodiment, except that the object processed in this embodiment is a picture.
  • the terminal device can display or save the filtered picture.
  • FIG. 11 is a schematic structural diagram of an apparatus for acquiring an image according to an embodiment of the present application.
  • the apparatus for acquiring an image includes a receiving unit 1101, an obtaining unit 1102, an executing unit 1103, and an output unit 1104.
  • the receiving unit 1101 is for receiving an input gesture operation, the gesture operation acting on an image acquired by the camera.
  • the acquisition unit 1102 is configured to determine a filter operation filter parameter according to the current filter type and the gesture operation.
  • the execution unit 1103 is configured to perform a filter operation corresponding to the filter type according to the filter parameter.
  • the output unit 1104 is for displaying an image performing the filter operation in real time through an output device.
  • the receiving unit 1101 obtains a filter parameter by the obtaining unit 1102.
  • the filter parameters include at least one or more of an operation command, a filter type, a filter intensity, a filter color, a filter size, and a filter transparency.
  • the operation instruction can be obtained according to the type of the gesture operation; the filter parameters such as the filter intensity, the filter color, the filter size, and the filter transparency can be obtained according to the operation amplitude value of the gesture operation.
  • the operation command corresponding to the gesture operation includes one or more of a slide instruction, a stretch instruction, a rotation instruction, a single-click instruction, and a double-click instruction.
  • the filter parameters may further include filter coordinate information. Said The filter coordinate information is acquired at the position of the image in accordance with the gesture operation. The filter operation corresponding to the filter parameter is performed at a position corresponding to the filter coordinate information.
  • the apparatus for acquiring an image may further include a filter selection unit 1105.
  • the filter selection unit 1105 is configured to receive the input selection information, and determine the current filter type based on the selection information. This gives the user the flexibility to choose the filter type for more filter effects.
  • FIG. 12 is a structural block diagram of an apparatus for performing picture processing according to an embodiment of the present application. As shown in FIG. 12, the apparatus for image processing includes a receiving module 1201, an obtaining module 1202, an executing module 1203, and an output module 1204.
  • the receiving module 1201 is configured to receive an input gesture operation, the gesture operation acting on a picture to be processed acquired by the camera.
  • the acquisition module 1202 is configured to determine a filter operation filter parameter based on the current filter type and the gesture operation.
  • the executing module 1203 is configured to perform a filter operation corresponding to the filter type on the to-be-processed picture according to the filter parameter.
  • the output module 1204 is for displaying a picture acquired after the filter operation is performed.
  • the filter parameters include at least one of the following: operational command, filter strength, filter color, filter size, filter transparency.
  • the operation instruction may be obtained according to the type of the gesture operation; the filter parameters such as the filter intensity, the filter color, the filter size, and the filter transparency may be obtained according to the operation amplitude value of the gesture operation.
  • the operation instruction corresponding to the gesture operation includes at least one of a slide instruction, a stretch instruction, a rotation instruction, a click instruction, and a double-click instruction.
  • the filter parameters further include filter coordinate information.
  • the filter coordinate information is acquired at a position of the image according to the gesture operation.
  • the filter parameter corresponds to The filter operation can be performed at a position corresponding to the filter coordinate information.
  • the embodiment of the present application may further include a filter selection module.
  • the filter selection module 1205 is configured to receive input selection information and determine a current filter type based on the selection information. This gives the user the flexibility to choose the filter type for more filter effects.
  • the terminal device may receive a gesture operation, determine a filter parameter according to the current filter type and the gesture operation, and perform a filter operation corresponding to the current filter type according to the filter parameter.
  • the current filter type may be at least one filter type preset setting, or may be at least one filter type determined according to selection information input by the user. Therefore, when the filter operation is performed, the filter operation can be adjusted in real time according to the user's gesture operation, thereby obtaining a rich filter effect.
  • the methods and apparatus provided by the embodiments of the present invention may be implemented by hardware, or computer readable instructions, or a combination of hardware and computer readable instructions.
  • the computer readable instructions used in this embodiment are stored by a plurality of processors in a readable storage medium, such as a hard disk, a CD-ROM,
  • DVD DVD, CD, floppy disk, tape, RAM, ROM or other suitable storage device.
  • computer readable instructions may be replaced by specific hardware, such as custom integrated circuits, gate arrays, FPGAs, PLDs, and computers with specific functions, and the like.
  • Embodiments of the present invention provide a computer readable storage medium for storing instructions for causing a computer to perform the methods described herein.
  • the system or device provided by this embodiment has a storage medium in which computer readable program code is stored for implementing the functions of any of the above embodiments, and these systems or devices (or CPUs or MPUs) can read and Execute the program code stored on the storage medium.
  • the program code read from the storage medium can implement any of the above embodiments, and thus the program code and the storage medium storing the program code are a part of the technical solution.
  • Storage media for providing program code include floppy disks, hard disks, magneto-optical disks, optical disks (eg, CD-ROM, CD-R, CD-RW, DVD-ROM, DVD-RAM, DVD-RW, DVD+RW), disks, Flash card, ROM, etc.
  • the program code can also be downloaded from the server computer via the communication network.
  • an operation at least partially implemented by the program code may be implemented by an operating system running on a computer, thereby implementing the technical solution of any of the above embodiments, wherein the computer is executed based on the program code. instruction.
  • program code in the storage medium is written to the memory, wherein the memory is located in an expansion board inserted in the computer or in an expansion unit connected to the computer.
  • the CPU in the expansion board or the expansion unit performs at least part of the operation based on the program code according to the instruction, thereby implementing the technical solution of any of the above embodiments.

Abstract

本申请提供了一种获取图像的方法、装置,该方法包括:接收输入的手势操作,所述操作信息作用在通过摄像头获取的图像上;根据当前滤镜类型和所述手势操作确定滤镜参数;根据所述滤镜参数,在所述图像上执行所述滤镜类型对应的滤镜操作;实时显示执行所述滤镜操作后获取的图像。

Description

获取图像的方法和装置
技术领域
本发明涉及视频处理领域, 尤其涉及一种获取图像的方法和装置。 发明背景
滤镜是一种可以快速修改或者美化用户所拍摄的图像的图像处理技 术。 在移动终端或者计算机等终端设备中可以通过安装应用程序进行各 种不同的滤镜操作。 通常, 应用程序提供多个滤镜, 并预先设定每个滤 镜的参数。 当用户选择了某个滤镜之后, 可以直接察看终端设备显示的 滤镜效果, 即经过该滤镜处理后的视频或者图片。
但是, 由于滤镜的参数是预先设定的, 每个视频或者图片只能获得 一种与该滤镜对应的滤镜效果。 因此, 由于滤镜的类型和数量的限制, 通过现有滤镜操作获得的滤镜效果比较单一。 发明内容
本申请实施例提供一种获取图像的方法和装置, 能够根据用户的手 势操作对图像进行滤镜处理, 从而获得的丰富的滤镜效果。
一种获取图像的方法, 所述方法包括下述步骤:
接收输入的手势操作, 所述操作信息作用在通过摄像头获取的图像 上;
根据当前滤镜类型和所述手势操作确定滤镜参数;
根据所述滤镜参数, 在所述图像上执行所述滤镜类型对应的滤镜操 作;
实时显示执行所述滤镜操作后获取的图像。 一种获取图像的装置, 所述装置包括:
接收单元: 用于接收输入的手势操作, 所述操作信息作用在通过摄 像头获取的图像上;
获取单元: 用于根据当前滤镜类型和所述手势操作确定滤镜参数; 执行单元: 用于根据所述滤镜参数, 在所述图像上执行所述滤镜类 型对应的滤镜操作;
输出单元: 用于实时显示执行所述滤镜操作后获取的图像。
可见, 根据本申请实施例, 终端设备可以接收手势操作, 根据当前 滤镜类型和该手势操作确定滤镜参数, 并根据该滤镜参数执行该当前滤 镜类型对应的滤镜操作。 其中, 当前滤镜类型可以是预设设置的至少一 个滤镜类型, 也可以是根据用户输入的选择信息确定的至少一个滤镜类 型。 因此, 在进行滤镜操作时, 能够实时地根据用户的手势操作对滤镜 操作进行调整, 从而获得的丰富的滤镜效果。 附图简要说明
图 1是本申请实施例提供的获取图像的方法的实现流程图。
图 2a、 2b、 2c为本申请实施例提供的通过旋转和滑动手势调整漩涡 滤镜的效果示意图。
图 3a、 3b、 3c为本申请实施例提供的通过旋转和缩放手势调整漩涡 滤镜的效果示意图。
图 4是本申请实施例提供的获取图像的方法的实现流程图。
图 5a、 5b为本申请实施例提供的通过缩放手势调整鱼眼滤镜的效果 示意图。
图 6a、 6b、 6c为本申请实施例提供的通过滑动手势控制滤色滤镜的 颜色值的效果示意图。 图 7a、 7b为本申请实施例提供的利用滑动手势改变色彩的效果示意 图。
图 8是本申请实施例提供的获取图像的 OpenGL ES 2.0的图形系统 结构示意图。
图 9 为本申请实施例提供的通过软硬件结合实现的实时渲染示意 图。
图 10为本申请实施例提供的图片处理的流程示意图。
图 11为本申请实施例提供的获取图像的装置的结构框图。
图 12是本申请实施例提供的图片处理的装置的结构框图。 实施本发明的方式
为了使本发明的技术方案及优点更加清楚明白, 以下结合附图及实 施例, 对本发明进行进一步详细说明。 应当理解, 此处所描述的具体实 施例仅仅用以解释本发明, 并不用于限定本发明。
在本申请实施例中, 接收针对图像的手势操作, 根据该手势操作获 取滤镜参数, 并根据该滤镜参数执行滤镜操作, 从而获得丰富的滤镜效 果。
图 1示出了本申请实施例提供的获取图像的方法实现流程。 如图 1 所示, 该方法包括如下步骤。
在步骤 S101中, 终端设备接收输入的手势(gesture )操作, 该手势 操作针对通过摄像头 (camera )获取的图像。
根据本申请实施例, 通过摄像头获取的图像为视频或者图片。
在本步骤中, 终端设备通过输入设备接收该手势操作。
根据本申请的一个实施例, 输入设备可以是触摸屏。 用户通过触摸 屏输入手势操作。由于触摸屏支持的手势操作丰富,对于移动终端设备, 如手机、 PAD等设备, 用户可以通过触摸屏灵活方便地输入手势操作。 例如, 用户通过触摸屏输入的手势操作可以对应滑动指令、 拉伸指令、 旋转指令、 单击或双击指令等。
根据本申请的一个实施例, 输入设备可以是键盘或者鼠标或者轨迹 球中的至少一种。 此时, 用户输入手势操作可以对应拖动指令、 单击指 令、 双击指令、 右击指令、 旋转指令等。
在步骤 S102中,终端设备根据当前滤镜类型和该手势操作确定滤镜 操作的滤镜参数。
根据本申请的一个实施例, 该终端设备可以采用预置的至少一个滤 镜类型作为当前滤镜类型, 如鱼眼滤镜、 滤色滤镜、 漩涡滤镜中的至少 一个。 另外, 该终端设备也可以根据用户的选择确定当前滤镜类型。 例 如, 终端设备首先向用户提供备选的多个滤镜类型, 接收用户选择的一 个或多个滤镜类型, 并将用户选择的一个或多个滤镜类型作为当前滤镜 类型。
根据本申请的一个实施例, 根据手势操作可以获得的滤镜参数至少 包括操作指令, 滤镜强度、 滤镜颜色、 滤镜大小、 滤镜透明度中的一种 或者多种。 根据本申请的一个实施例, 该滤镜参数还可以包括滤镜坐标 信息。
根据本申请的一个实施例, 当采用的滤镜类型不同时, 相同手势操 作对应的操作指令也不同。例如通过触摸屏输入的滑动指令、拉伸指令、 旋转指令在采用不同类型的滤镜时, 对应不同的操作指令。 根据本申请 一个实施例, 终端设备中预先存储滤镜类型, 手势操作和操作指令的对 应关系。 根据本申请一个实施例, 终端设备也可以接收并存储用户设定 的滤镜类型, 手势操作和操作指令的对应关系。 例如, 用户设定一些喜 欢用的手势操作对应常用的操作指令。 当输入设备是触摸屏时,手势操作对应的操作指令可以为滑动指令、 拉伸指令、 旋转指令、 单击指令或双击指令等。 当输入设备键盘或者鼠 标或者轨迹球中的至少一种时, 手势操作对应的操作指令可以为拖动指 令、 单击指令、 双击指令、 右击指令、 旋转指令等。
根据本申请的一个实施例, 还可以根据滤镜类型和手势操作的幅度 确定滤镜参数, 例如滤镜强度、 滤镜颜色、 滤镜大小、 滤镜透明度等。 其中手势操作的幅度可以根据滤镜坐标信息获取。
根据本申请的一个实施例, 可以根据手势操作的拉伸长度或者缩放 的大小确定滤镜参数。 例如, 当滤镜类型为黑白滤镜, 手势操作为滑动 时,可以根据滑动的幅度确定黑白滤镜中滤镜强度,例如黑色的深浅度。 当滤镜类型为华盛顿滤镜, 手势为滑动时, 可以根据滑动的幅度确定滤 镜颜色。 当滤镜类型为鱼眼滤镜, 手势为拉伸时, 可以根据拉伸的幅度 确定滤镜大小。
当输入设备为触摸屏时, 手指触摸的位置为滤镜坐标信息。 当输入 设备为鼠标, 键盘和轨迹球中的至少一个时, 手势操作过程中指针的位 置为滤镜坐标信息。 另外, 根据本申请的一个实施例, 滤镜坐标信息也 可以是预先设定的位置, 默认在该预定的位置进行滤镜操作。
在步骤 S103中,根据所述滤镜参数,执行所述当前滤镜类型对应的 滤镜操作, 获取所述图像的滤镜效果。
图 2a、 2b、 2c为本申请实施例提供的通过旋转和滑动手势调整漩涡 滤镜的效果示意图。
图 2a为在常规漩涡滤镜处理后获得的图像画面。 图 2b为在预置的 滤镜类型为旋涡滤镜的情况下, 执行了拉伸命令后的滤镜效果。 用户通 过触摸屏输入的手势操作如图 2b所示。 该手势操作的漩涡滤镜中对应 拉伸指令。 根据该手势操作的幅度可以确定滤镜强调, 即漩涡被放大的 倍数, 根据该手势操作中手指的位置, 确定执行漩涡滤镜的位置。 在图
2b中, 根据拉伸指令放大手势操作对应的漩涡部分。 图 2c为在预置的 滤镜类型为旋涡滤镜的情况下, 执行了拉伸命令后的, 接着又执行了滑 动命令后的效果。 在图 2c中, 图 2b中的漩涡被移动到了滑动命令对应 的位置。
图 3a、 图 3b、 图 3c所示为本申请实施例提供的通过旋转和缩放手 势改变漩涡滤镜的效果示意图。 图 3a为执行了漩涡滤镜操作的图像。用 户通过输入如图 3b中的旋转手势, 可以改变图 3a中漩涡滤镜操作的强 度。 其中, 强度改变的位置为旋转手势的位置, 强度改变的程度根据旋 转手势的幅度确定, 从而得到如图 3b 所示效果图。 用户通过输入拉伸 手势, 可以将图 3b 中的漩涡部分放大。 其中, 放大的程度根据拉伸手 势的幅度确定, 从而在图 3b的基础上得到图 3c述示效果。
因此, 根据本申请实施例, 在进行滤镜操作的过程中, 可以根据输 入的手势操作灵活改变滤镜操作的参数, 对于同一个画面, 不同的使用 者输入不同指令会得到不同的滤镜效果。 或者不同的使用者输入相同的 指令, 但输入相同指令的顺序不同, 也会得到不同的滤镜效果。
在步骤 S104中,将执行所述滤镜操作后获得的图像通过输出设备进 行实时显示。
根据本申请实施例,可以在步骤 S103中可以执行多个滤镜操作,并 获得最终的滤镜效果。 在本步骤中将最终的滤镜效果进行实时显示。
在本步骤中,步骤 S103中获取的滤镜效果通过显示屏进行显示。例 如可以通过接收手势操作的触摸屏显示图像。在实际应用中,智能手机、 平板电脑等都可以提供触摸屏功能, 接收手势操作, 显示滤镜处理后的 图像。 当然也可以用于其它诸如台式电脑等设备, 在此对设备仅为举例 说明, 并不能以此局限于本申请仅能应用于上述具体设备, 但凡有摄像 头进行画面呈现的其它设备, 均可以应用本申请所述的获取图像的方 法。
在本实施例中, 通过接收包括图面坐标的手势操作, 在预置的滤镜 类型下执行不同的滤镜操作, 使滤镜的变换方式更加丰富, 得到经过滤 镜效果处理后的画面也更加丰富。
图 4示出了本申请实施例提供的获取图像的方法实现流程。 如图 4 所示, 该方法包括如下步骤。
在步骤 S401中,终端设备接收输入的选择信息,根据该选择信息确 定当前滤镜类型。
根据本申请实施例, 根据该选择信息, 可以确定一个或多个滤镜类 型作为当前滤镜类型, 从而丰富滤镜处理后获得的滤镜效果。
在步骤 S402中,终端设备接收输入的手势操作,该手势操作作用在 通过摄像头获取的图像上。
在步骤 S403中,终端设备根据当前滤镜类型和该手势操作确定滤镜 操作滤镜参数。
由于本实施例中在步骤 S401选择有滤镜类型, 因此, 步骤 S403中 直接采用步骤 S401中选择的滤镜类型。
在步骤 S404中,根据所述滤镜参数,执行所述滤镜类型对应的滤镜 操作。
在本实施例中, 滤镜参数包括滤镜强度、 滤镜颜色、 滤镜大小、 滤 镜透明度中的一种或者多种。
所述滤镜参数还可以包括滤镜坐标信息, 所述滤镜参数对应的滤镜 操作在所述图像中滤镜坐标信息对应的位置上执行。
根据本申请一个实施例, 滤镜参数可以根据所述手势操作的操作幅 度值和 /或手势操作类型获取。 例如, 根据手势操作的拉伸长度或者缩放 的大小对应相应的滤镜参数,或者根据操作指令如缩放指令、平滑指令、 拉伸指令等对应相应的滤镜参数。
所述滤镜坐标信息可以根据所述手势操作在所述图像的位置获取, 以灵活的获取需要进行变换的滤镜效果的位置, 也可以设定为固定的几 个位置, 在此不作限定。
在本步骤中, 根据预先存储的滤镜类型, 手势操作和滤镜操作的对 应关系, 确定在当前滤镜类型中与该手势操作对应的滤镜操作; 根据所 述滤镜类型及滤镜参数, 对该图像进行滤镜操作。
根据本申请一个实施例, 在确定滤镜操作后, 可以根据手势操作的 操作幅度值, 确定该滤镜操作所需要执行的滤镜参数。 例如, 当通过触 摸屏输入的手势操作为滑动操作时, 根据滑动的距离的大小, 确定滤镜 参数, 例如确定黑白滤镜中黑色的深浅度。 因此, 通过手势操作的幅度 值改变滤镜操作中的相关参数, 从而进一步提高了滤镜操作的多样性, 获得更丰富的滤镜效果。
下面结合图片举例说明在鱼眼滤镜模式下, 接收通过触摸屏输入的 手势操作后, 通过执行滤镜操作获得的滤镜效果。
图 5a、 5b为本申请实施例提供的通过缩放手势调整鱼眼滤镜的效果 示意图。 在图 5a所示执行滤镜操作前的原始画面。 图 5a中的 QQ公仔 的大小与实际相符。 图 5b为执行了滤镜操作后获得的效果图。 如图 5b 所示, 手势操作对应的操作指令为缩放指令, 可以对鱼眼滤镜的扭曲程 度进行调节。 根据手势操作指示的缩放程度, 执行与该缩放程度对应级 别的鱼眼滤镜效果。 例如, 手势操作指示了轻微的缩放, 在图像中滤镜 坐标信息对应的位置执行轻微的扭曲操作。
图 6a、 6b、 6c为本申请实施例提供的通过滑动手势控制滤色滤镜过 滤的颜色值的效果示意图。 在滤色滤镜操作中, 首先确定滑动手势中手 势展开幅度对应的颜色区间, 将除该颜色区间之外的颜色过滤并进行灰 度处理。通过滑动手势,在图 6a、 6b、 6c图中的分别显示了红色的花瓶, 绿色的书籍封面和蓝色的杂志内页, 不显示的物品和其它图像作灰度处 理。
图 7a、 7b为本申请实施例提供的通过滑动手势改变色彩的效果示意 图。 图 7a为未进行滤镜操作的原始图像。接收了通过触摸屏输入的滑动 手势后, 改变图 7a中颜色所占的比例, 从而调整画面的颜色显示, 实现 华盛顿滤镜的滤镜效果。
在步骤 S405中,将执行所述滤镜操作后获得的图像通过输出设备进 行实时显示。
根据本申请实施例,可以在步骤 S404中可以执行多个滤镜操作,并 获得最终的滤镜效果。 在本步骤中将最终的滤镜效果进行实时显示。
在本步骤中,步骤 S403中获取的滤镜效果通过显示屏进行显示。例 如可以通过接收手势操作的触摸屏显示图像。在实际应用中,智能手机、 平板电脑等都可以提供触摸屏功能, 接收手势操作, 显示滤镜处理后的 图像。 当然也可以用于其它诸如台式电脑等设备, 在此对设备仅为举例 说明, 并不能以此局限于本申请仅能应用于上述具体设备, 但凡有摄像 头进行画面呈现的其它设备, 均可以应用本申请所述的获取图像的方 法。
根据本申请一个实施例, 输出设备利用 OpenGL ES技术, 通过图形 处理器( Graphic Processing Unit, GPU ) 实时显示滤镜操作后获得的画 面。
如图 8所示, 大部分图形系统都可以比作工厂中的装配线 (Assemble line)或者称为管道 (Pipeline)。前一道的输出作为下道工序的输入。主 CPU 发出一个绘图指令, 然后可能由硬件部件完成坐标变换, 裁剪, 添加颜 色或是材质, 最后在屏幕上显示出来。
根据本申请一个实施例, 将滤镜消耗资源时间最大的处理工序交予
OpenGL完成, 在着色器里创建并编写 shader着色程序, 如图 9所示, OpenGL编译着色 shader并链接到程序 program上。 原程序获取图像数 据, 通过置于 OpenGL管道流程的 shader着色器对图片每个像素来进行 变换处理, 整个过程是并行的, 用 shader语言来编程处理变换, 依赖终 端 GPU来进行运算, 大大提高了滤镜处理的效率。
图 10为本申请实施例提供的图片处理的流程示意图。如图 10所示, 该方法包括以下步骤。
在步骤 S1001中, 终端设备接收输入的手势操作, 所述手势操作作 用在待处理的图片上。
在步骤 S1002中, 终端设备根据当前滤镜类型和所述手势操作确定 滤镜操作的滤镜参数。
在步骤 S1003中, 终端设备根据所述滤镜参数, 在所述滤镜类型对 应的滤镜操作。
其中, 所述滤镜参数可以包括滤镜坐标信息, 所述滤镜参数对应的 滤镜操作在所述图片中滤镜坐标信息对应的位置上执行。
所述滤镜参数包括滤镜类型、 滤镜强度、 滤镜颜色、 滤镜大小和滤 镜透明度中的至少一个。
根据本申请一个实施例, 滤镜参数可以根据所述手势操作的操作幅 度值和 /或手势操作类型获取,滤镜坐标信息可以根据所述手势操作在所 述图像的位置获取。 所述手势操作对应的操作指令包括滑动指令、 拉伸 指令、 旋转指令、 单击指令、 双击指令的至少其中之一。
在步骤 S1004中, 终端设备显示执行所述滤镜操作后获取的图片。 根据本申请一个实施例, 在终端设备接收输入的手势操作之前, 还 包括: 接收输入的选择信息, 根据该选择信息确定当前的滤镜类型。 根 据本申请实施例, 根据该选择信息, 可以确定一个或多个滤镜类型作为 当前滤镜类型, 从而丰富滤镜处理后获得的滤镜效果。
本申请实施例中所述步骤与上述实施例的处理过程基本相同, 区别 之处在于本实施例所处理的对象为图片。 在本实施例中, 终端设备可以 显示或者保存经过滤镜处理后的图片。
图 11示出了本申请实施例提供的获取图像的装置结构示意图。如图 11所示, 该获取图像的装置包括接收单元 1101、 获取单元 1102、 执行 单元 1103、 输出单元 1104。
接收单元 1101用于接收输入的手势操作,所述手势操作作用在通过 摄像头获取的图像上。
获取单元 1102 用于根据当前滤镜类型和该手势操作确定滤镜操作 滤镜参数。
执行单元 1103用于根据所述滤镜参数,执行所述滤镜类型对应的滤 镜操作。
输出单元 1104 用于通过输出设备实时显示执行所述滤镜操作的图 像。
根据本申请一个实施例,接收单元 1101接收到该手势操作后, 由获 取单元 1102 获取滤镜参数。 所述滤镜参数至少包括操作指令, 滤镜类 型、 滤镜强度、 滤镜颜色、 滤镜大小、 滤镜透明度中的一种或多种。 其 中, 操作指令可以根据手势操作的类型获取; 滤镜强度、 滤镜颜色、 滤 镜大小、 滤镜透明度等滤镜参数可以根据所述手势操作的操作幅度值获 取。 手势操作对应的操作指令包括滑动指令、 拉伸指令、 旋转指令、 单 击指令、 双击指令中的一种或多种。
根据本申请一个实施例, 滤镜参数还可以包括滤镜坐标信息。 所述 滤镜坐标信息根据所述手势操作在所述图像的位置获取。 所述滤镜参数 对应的滤镜操作在所述滤镜坐标信息对应的位置上执行。
根据本申请一个实施例, 该获取图像的装置还可以包括滤镜选择单 元 1105。 该滤镜选择单元 1105用于接收输入的选择信息, 根据所述选 择信息确定当前滤镜类型。 从而使用户可以灵活选择滤镜类型, 实现更 多的滤镜效果。
上述装置与方法实施例相对应, 在此不作重复赘述。
图 12示出了本申请实施例提供的图片处理的装置的结构框图。如图 12所示, 该图片处理的装置包括接收模块 1201、 获取模块 1202、 执行 模块 1203、 输出模块 1204。
接收模块 1201用于接收输入的手势操作,所述手势操作作用在通过 摄像头获取的待处理的图片上。
获取模块 1202 用于根据当前滤镜类型和该手势操作确定滤镜操作 滤镜参数。
执行模块 1203用于根据所述滤镜参数,在所述待处理的图片上执行 滤镜类型对应的滤镜操作。
输出模块 1204用于显示执行所述滤镜操作后获取的图片。
所述滤镜参数包括以下各项的至少其中之一: 操作指令、滤镜强度、 滤镜颜色、 滤镜大小、 滤镜透明度。 其中, 操作指令可以根据手势操作 的类型获取; 滤镜强度、 滤镜颜色、 滤镜大小、 滤镜透明度等滤镜参数 可以根据所述手势操作的操作幅度值获取。 所述手势操作对应的操作指 令包括滑动指令、 拉伸指令、 旋转指令、 单击指令、 双击指令的至少其 中之一。
根据本申请一个实施例, 滤镜参数还包括滤镜坐标信息。 所述滤镜 坐标信息根据所述手势操作在所述图像的位置获取。 所述滤镜参数对应 的滤镜操作可以在所述滤镜坐标信息对应的位置上执行。
根据本申请一个实施例, 本申请实施例还可包括滤镜选择模块
1205。 该滤镜选择模块 1205 用于接收输入的选择信息, 根据该选择信 息确定当前滤镜类型。 从而使用户可以灵活选择滤镜类型, 实现更多的 滤镜效果。
本实施例为方法实施例对应的系统实施方式, 在此不作重复表述。 综上, 在本发明实施例中, 终端设备可以接收手势操作, 根据当前 滤镜类型和该手势操作确定滤镜参数, 并根据该滤镜参数执行该当前滤 镜类型对应的滤镜操作。 其中, 当前滤镜类型可以是预设设置的至少一 个滤镜类型, 也可以是根据用户输入的选择信息确定的至少一个滤镜类 型。 因此, 在进行滤镜操作时, 能够实时地根据用户的手势操作对滤镜 操作进行调整, 从而获得的丰富的滤镜效果。
本发明实施例提供的方法和设备可以由硬件、 或计算机可读指令、 或者硬件和计算机可读指令的结合来实现。 本实施例中使用的计算机可 读指令由多个处理器存储在可读存储介质中, 例如硬盘、 CD-ROM、
DVD, 光盘、 软盘、 磁带、 RAM、 ROM或其它合适的存储设备。 或者, 至少部分计算机可读指令可以由具体硬件替换, 例如, 定制集成线路、 门阵列、 FPGA、 PLD和具体功能的计算机等等。
本发明实施例提供了计算机可读存储介质, 用于存储指令使得计算 机执行本文所述的方法。 具体地, 本实施例提供的系统或设备都具有存 储介质, 其中存储了计算机可读程序代码, 用于实现上述任意实施例的 功能, 并且这些系统或设备 (或 CPU或 MPU ) 能够读取并且执行存储 在存储介质中的程序代码。
在这种情况下, 从存储介质中读取的程序代码可以实现上述任一实 施例, 因此该程序代码和存储该程序代码的存储介质是技术方案的一部 用于提供程序代码的存储介质包括软盘、 硬盘、 磁光盘、 光盘(例 如 CD-ROM、 CD-R, CD-RW、 DVD-ROM、 DVD-RAM、 DVD-RW, DVD+RW ), 磁盘、 闪存卡、 ROM等等。 可选地, 程序代码也可以通过 通信网络从 务器电脑上下载。
应该注意的是, 对于由计算机执行的程序代码, 至少部分由程序代 码实现的操作可以由运行在计算机上的操作系统实现, 从而实现上述任 一实施例的技术方案, 其中该计算机基于程序代码执行指令。
另外, 存储介质中的程序代码被被写入存储器, 其中, 该存储器位 于插入在计算机中的扩展板中, 或者位于连接到计算机的扩展单元中。 在本实施例中,扩展板或扩展单元中的 CPU根据指令,基于程序代码执 行至少部分操作, 从而实现上述任一实施例的技术方案。
以上所述仅为本发明的较佳实施例而已, 并不用以限制本发明, 凡 在本发明的精神和原则之内所作的任何修改、 等同替换和改进等, 均应 包含在本发明的保护范围之内。

Claims

权利要求书
1、 一种获取图像的方法, 其特征在于, 所述方法包括下述步骤: 接收输入的手势操作, 所述操作信息作用在通过摄像头获取的图像 上;
根据当前滤镜类型和所述手势操作确定滤镜参数;
根据所述滤镜参数, 在所述图像上执行所述滤镜类型对应的滤镜操 作;
实时显示执行所述滤镜操作后获取的图像。
2、根据权利要求 1所述的方法, 其特征在于, 所述滤镜参数包括操 作指令, 以及滤镜强度、 滤镜颜色、 滤镜大小、 滤镜透明度中的至少一 个。
3、根据权利要求 2所述的方法, 其特征在于, 所述操作指令根据所 述手势操作的类型获取; 所述滤镜强度、 滤镜颜色、 滤镜大小、 滤镜透 明度中的至少一个根据所述手势操作的操作幅度值获取。
4、根据权利要求 1所述方法, 其特征在于, 所述接收输入的手势操 作之前, 还包括:
接收输入的选择信息;
根据所述选择信息确定所述当前滤镜类型。
5、根据权利要求 1所述的方法, 其特征在于, 所述滤镜参数包括滤 镜坐标信息。
6、根据权利要求 5所述的方法, 其特征在于, 所述滤镜坐标信息根 据所述手势操作在所述图像的位置获取。
7、根据权利要求 2所述的方法, 其特征在于, 所述操作指令包括滑 动指令、 拉伸指令、 旋转指令、 单击指令、 双击指令的至少其中之一。
8、 根据权利要求 1-7任一项所述的方法, 其特征在于, 所述通过摄 像头获取的图像包括视频或者图片。
9、 一种获取图像的装置, 其特征在于, 所述装置包括:
接收单元: 用于接收输入的手势操作, 所述操作信息作用在通过摄 像头获取的图像上;
获取单元: 用于根据当前滤镜类型和所述手势操作确定滤镜参数; 执行单元: 用于根据所述滤镜参数, 在所述图像上执行所述滤镜类 型对应的滤镜操作;
输出单元: 用于实时显示执行所述滤镜操作后获取的图像。
10、 根据权利要求 9所述的装置, 其特征在于, 所述滤镜参数包括 操作指令, 以及滤镜强度、 滤镜颜色、 滤镜大小、 滤镜透明度中的至少 一个。
11、根据权利要求 10所述的装置, 其特征在于, 所述操作指令根据 所述手势操作的类型获取; 所述滤镜强度、 滤镜颜色、 滤镜大小、 滤镜 透明度中的至少一个根据所述手势操作的操作幅度值获取。
12、 根据权利要求 9所述装置, 其特征在于, 所述装置还包括: 滤镜类型选择单元: 接收输入的选择信息, 根据所述选择信息确定 所述当前滤镜类型。
13、 根据权利要求 9所述的装置, 其特征在于, 所述滤镜参数包括 滤镜坐标信息.
14、 根据权利要求 9所述的装置, 其特征在于, 所述滤镜坐标信息 根据所述手势操作在所述图像的位置获取。
15、根据权利要求 10所述的装置, 其特征在于, 所述操作指令包括 滑动指令、拉伸指令、 旋转指令、 单击指令、 双击指令的至少其中之一。
16、 根据权利要求 9-15任一项所述的装置, 其特征在于, 所述通过 摄像头获取的图像包括视频或者图片。
PCT/CN2013/084868 2012-10-18 2013-10-09 获取图像的方法和装置 WO2014059886A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/688,123 US9674425B2 (en) 2012-10-18 2015-04-16 Image acquisition method and apparatus

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201210397363.4A CN103777852B (zh) 2012-10-18 2012-10-18 一种获取图像的方法、装置
CN201210397363.4 2012-10-18

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/688,123 Continuation US9674425B2 (en) 2012-10-18 2015-04-16 Image acquisition method and apparatus

Publications (1)

Publication Number Publication Date
WO2014059886A1 true WO2014059886A1 (zh) 2014-04-24

Family

ID=50487568

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2013/084868 WO2014059886A1 (zh) 2012-10-18 2013-10-09 获取图像的方法和装置

Country Status (3)

Country Link
US (1) US9674425B2 (zh)
CN (1) CN103777852B (zh)
WO (1) WO2014059886A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015169213A1 (zh) * 2014-05-07 2015-11-12 厦门美图之家科技有限公司 一种视频文件的滤镜处理方法

Families Citing this family (161)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8554868B2 (en) 2007-01-05 2013-10-08 Yahoo! Inc. Simultaneous sharing communication interface
EP2732383B1 (en) 2011-07-12 2018-04-04 Snap Inc. Methods and systems of providing visual content editing functions
US8972357B2 (en) 2012-02-24 2015-03-03 Placed, Inc. System and method for data collection to validate location data
US11734712B2 (en) 2012-02-24 2023-08-22 Foursquare Labs, Inc. Attributing in-store visits to media consumption based on data collected from user devices
US10155168B2 (en) 2012-05-08 2018-12-18 Snap Inc. System and method for adaptable avatars
WO2014031899A1 (en) 2012-08-22 2014-02-27 Goldrun Corporation Augmented reality virtual content platform apparatuses, methods and systems
KR102063915B1 (ko) 2013-03-14 2020-01-08 삼성전자주식회사 사용자 기기 및 그 동작 방법
US9705831B2 (en) 2013-05-30 2017-07-11 Snap Inc. Apparatus and method for maintaining a message thread with opt-in permanence for entries
US10439972B1 (en) 2013-05-30 2019-10-08 Snap Inc. Apparatus and method for maintaining a message thread with opt-in permanence for entries
CA2863124A1 (en) 2014-01-03 2015-07-03 Investel Capital Corporation User content sharing system and method with automated external content integration
US9628950B1 (en) 2014-01-12 2017-04-18 Investment Asset Holdings Llc Location-based messaging
US10082926B1 (en) 2014-02-21 2018-09-25 Snap Inc. Apparatus and method for alternate channel communication initiated through a common message thread
JP2015172836A (ja) * 2014-03-11 2015-10-01 キヤノン株式会社 表示制御装置及び表示制御方法
US9276886B1 (en) 2014-05-09 2016-03-01 Snapchat, Inc. Apparatus and method for dynamically configuring application component tiles
JP6349962B2 (ja) * 2014-05-27 2018-07-04 富士ゼロックス株式会社 画像処理装置およびプログラム
US9537811B2 (en) 2014-10-02 2017-01-03 Snap Inc. Ephemeral gallery of ephemeral messages
US9396354B1 (en) 2014-05-28 2016-07-19 Snapchat, Inc. Apparatus and method for automated privacy protection in distributed images
IL239237B (en) 2014-06-05 2018-12-31 Rotem Efrat Network document extension
US9113301B1 (en) 2014-06-13 2015-08-18 Snapchat, Inc. Geo-location based event gallery
CN105245948B (zh) * 2014-06-26 2019-02-05 北京新媒传信科技有限公司 视频处理方法和装置
CN105320506B (zh) * 2014-06-30 2018-11-23 中国移动通信集团公司 一种获取图像的方法及装置
US9225897B1 (en) 2014-07-07 2015-12-29 Snapchat, Inc. Apparatus and method for supplying content aware photo filters
CN105338403B (zh) * 2014-08-06 2020-03-03 腾讯科技(北京)有限公司 滤镜处理方法、装置及电子设备
CN104144297A (zh) * 2014-08-19 2014-11-12 诚迈科技(南京)股份有限公司 一种自动给拍摄的照片添加水印的系统及方法
CN104159032B (zh) * 2014-08-20 2018-05-29 广东欧珀移动通信有限公司 一种实时调整相机拍照美颜效果的方法及装置
US10055717B1 (en) 2014-08-22 2018-08-21 Snap Inc. Message processor with application prompts
US10423983B2 (en) 2014-09-16 2019-09-24 Snap Inc. Determining targeting information based on a predictive targeting model
US10824654B2 (en) 2014-09-18 2020-11-03 Snap Inc. Geolocation-based pictographs
US11216869B2 (en) 2014-09-23 2022-01-04 Snap Inc. User interface to augment an image using geolocation
US10284508B1 (en) 2014-10-02 2019-05-07 Snap Inc. Ephemeral gallery of ephemeral messages with opt-in permanence
US9015285B1 (en) 2014-11-12 2015-04-21 Snapchat, Inc. User interface for accessing media at a geographic location
US10311916B2 (en) 2014-12-19 2019-06-04 Snap Inc. Gallery of videos set to an audio time line
US9385983B1 (en) 2014-12-19 2016-07-05 Snapchat, Inc. Gallery of messages from individuals with a shared interest
US9754355B2 (en) 2015-01-09 2017-09-05 Snap Inc. Object recognition based photo filters
US11388226B1 (en) 2015-01-13 2022-07-12 Snap Inc. Guided personal identity based actions
US10133705B1 (en) 2015-01-19 2018-11-20 Snap Inc. Multichannel system
US9521515B2 (en) 2015-01-26 2016-12-13 Mobli Technologies 2010 Ltd. Content request by location
CN104967801B (zh) * 2015-02-04 2019-09-17 腾讯科技(深圳)有限公司 一种视频数据处理方法和装置
US10223397B1 (en) 2015-03-13 2019-03-05 Snap Inc. Social graph based co-location of network users
KR102217723B1 (ko) 2015-03-18 2021-02-19 스냅 인코포레이티드 지오-펜스 인가 프로비저닝
US9692967B1 (en) 2015-03-23 2017-06-27 Snap Inc. Systems and methods for reducing boot time and power consumption in camera systems
US10135949B1 (en) 2015-05-05 2018-11-20 Snap Inc. Systems and methods for story and sub-story navigation
US9881094B2 (en) 2015-05-05 2018-01-30 Snap Inc. Systems and methods for automated local story generation and curation
CN105068822A (zh) * 2015-07-09 2015-11-18 厦门美图网科技有限公司 一种实时滤镜的在线配置方法、系统及拍摄终端
US10993069B2 (en) 2015-07-16 2021-04-27 Snap Inc. Dynamically adaptive media content delivery
CN106406505A (zh) * 2015-07-28 2017-02-15 北京金山安全软件有限公司 一种图片滤镜效果的编辑方法及其系统
US10817898B2 (en) 2015-08-13 2020-10-27 Placed, Llc Determining exposures to content presented by physical objects
CN105224177B (zh) * 2015-09-30 2019-03-01 努比亚技术有限公司 一种移动终端及应用图标的重绘方法
US9652896B1 (en) 2015-10-30 2017-05-16 Snap Inc. Image based tracking in augmented reality systems
US10474321B2 (en) 2015-11-30 2019-11-12 Snap Inc. Network resource location linking and visual content sharing
US9984499B1 (en) 2015-11-30 2018-05-29 Snap Inc. Image and point cloud based tracking and in augmented reality systems
CN105357451B (zh) * 2015-12-04 2019-11-29 Tcl集团股份有限公司 基于滤镜特效的图像处理方法及装置
US10354425B2 (en) 2015-12-18 2019-07-16 Snap Inc. Method and system for providing context relevant media augmentation
US10285001B2 (en) 2016-02-26 2019-05-07 Snap Inc. Generation, curation, and presentation of media collections
US10679389B2 (en) 2016-02-26 2020-06-09 Snap Inc. Methods and systems for generation, curation, and presentation of media collections
US11023514B2 (en) 2016-02-26 2021-06-01 Snap Inc. Methods and systems for generation, curation, and presentation of media collections
CN105741228B (zh) * 2016-03-11 2019-03-12 腾讯科技(深圳)有限公司 图形处理方法及装置
US10339365B2 (en) 2016-03-31 2019-07-02 Snap Inc. Automated avatar generation
US11900418B2 (en) 2016-04-04 2024-02-13 Snap Inc. Mutable geo-fencing system
US10805696B1 (en) 2016-06-20 2020-10-13 Pipbin, Inc. System for recording and targeting tagged content of user interest
US11044393B1 (en) 2016-06-20 2021-06-22 Pipbin, Inc. System for curation and display of location-dependent augmented reality content in an augmented estate system
US11876941B1 (en) 2016-06-20 2024-01-16 Pipbin, Inc. Clickable augmented reality content manager, system, and network
US11201981B1 (en) 2016-06-20 2021-12-14 Pipbin, Inc. System for notification of user accessibility of curated location-dependent content in an augmented estate
US10638256B1 (en) 2016-06-20 2020-04-28 Pipbin, Inc. System for distribution and display of mobile targeted augmented reality content
US10334134B1 (en) 2016-06-20 2019-06-25 Maximillian John Suiter Augmented real estate with location and chattel tagging system and apparatus for virtual diary, scrapbooking, game play, messaging, canvasing, advertising and social interaction
US11785161B1 (en) 2016-06-20 2023-10-10 Pipbin, Inc. System for user accessibility of tagged curated augmented reality content
US10430838B1 (en) 2016-06-28 2019-10-01 Snap Inc. Methods and systems for generation, curation, and presentation of media collections with automated advertising
US9681265B1 (en) 2016-06-28 2017-06-13 Snap Inc. System to track engagement of media items
US10733255B1 (en) 2016-06-30 2020-08-04 Snap Inc. Systems and methods for content navigation with automated curation
US10348662B2 (en) 2016-07-19 2019-07-09 Snap Inc. Generating customized electronic messaging graphics
CN109804411B (zh) 2016-08-30 2023-02-17 斯纳普公司 用于同时定位和映射的系统和方法
US10432559B2 (en) 2016-10-24 2019-10-01 Snap Inc. Generating and displaying customized avatars in electronic messages
EP3535756B1 (en) 2016-11-07 2021-07-28 Snap Inc. Selective identification and order of image modifiers
CN106791016A (zh) * 2016-11-29 2017-05-31 努比亚技术有限公司 一种拍照方法及终端
US10203855B2 (en) 2016-12-09 2019-02-12 Snap Inc. Customized user-controlled media overlays
US11616745B2 (en) 2017-01-09 2023-03-28 Snap Inc. Contextual generation and selection of customized media content
US10454857B1 (en) 2017-01-23 2019-10-22 Snap Inc. Customized digital avatar accessories
US10915911B2 (en) 2017-02-03 2021-02-09 Snap Inc. System to determine a price-schedule to distribute media content
US11250075B1 (en) 2017-02-17 2022-02-15 Snap Inc. Searching social media content
US10319149B1 (en) 2017-02-17 2019-06-11 Snap Inc. Augmented reality anamorphosis system
US10074381B1 (en) 2017-02-20 2018-09-11 Snap Inc. Augmented reality speech balloon system
US10565795B2 (en) 2017-03-06 2020-02-18 Snap Inc. Virtual vision system
US10523625B1 (en) 2017-03-09 2019-12-31 Snap Inc. Restricted group content collection
US10581782B2 (en) 2017-03-27 2020-03-03 Snap Inc. Generating a stitched data stream
US10582277B2 (en) 2017-03-27 2020-03-03 Snap Inc. Generating a stitched data stream
US11170393B1 (en) 2017-04-11 2021-11-09 Snap Inc. System to calculate an engagement score of location based media content
US10387730B1 (en) 2017-04-20 2019-08-20 Snap Inc. Augmented reality typography personalization system
CN110945555A (zh) 2017-04-27 2020-03-31 斯纳普公司 社交媒体平台上用户位置的区域级别表示
US10212541B1 (en) 2017-04-27 2019-02-19 Snap Inc. Selective location-based identity communication
US11893647B2 (en) 2017-04-27 2024-02-06 Snap Inc. Location-based virtual avatars
US10467147B1 (en) 2017-04-28 2019-11-05 Snap Inc. Precaching unlockable data elements
US10803120B1 (en) 2017-05-31 2020-10-13 Snap Inc. Geolocation based playlists
CN107368603A (zh) * 2017-07-28 2017-11-21 上海鋆创信息技术有限公司 一种图片滤镜处理的方法及装置、存储介质和终端
CN107608506B (zh) * 2017-09-01 2021-01-26 北京小米移动软件有限公司 图片处理方法及装置
US11475254B1 (en) 2017-09-08 2022-10-18 Snap Inc. Multimodal entity identification
US10740974B1 (en) 2017-09-15 2020-08-11 Snap Inc. Augmented reality system
CN108307088B (zh) * 2017-10-09 2019-07-26 腾讯科技(深圳)有限公司 一种图像处理方法、装置、系统和存储介质
US10499191B1 (en) 2017-10-09 2019-12-03 Snap Inc. Context sensitive presentation of content
US10573043B2 (en) 2017-10-30 2020-02-25 Snap Inc. Mobile-based cartographic control of display content
US11265273B1 (en) 2017-12-01 2022-03-01 Snap, Inc. Dynamic media overlay with smart widget
US11017173B1 (en) 2017-12-22 2021-05-25 Snap Inc. Named entity recognition visual context and caption data
US10678818B2 (en) 2018-01-03 2020-06-09 Snap Inc. Tag distribution visualization system
US11507614B1 (en) 2018-02-13 2022-11-22 Snap Inc. Icon based tagging
US10885136B1 (en) 2018-02-28 2021-01-05 Snap Inc. Audience filtering system
US10979752B1 (en) 2018-02-28 2021-04-13 Snap Inc. Generating media content items based on location information
US10327096B1 (en) 2018-03-06 2019-06-18 Snap Inc. Geo-fence selection system
KR102574151B1 (ko) 2018-03-14 2023-09-06 스냅 인코포레이티드 위치 정보에 기초한 수집가능한 항목들의 생성
US11163941B1 (en) 2018-03-30 2021-11-02 Snap Inc. Annotating a collection of media content items
US10219111B1 (en) 2018-04-18 2019-02-26 Snap Inc. Visitation tracking system
US10896197B1 (en) 2018-05-22 2021-01-19 Snap Inc. Event detection system
CN110662102B (zh) * 2018-06-29 2021-11-09 武汉斗鱼网络科技有限公司 滤镜渐变效果展示方法、存储介质、设备及系统
US10679393B2 (en) 2018-07-24 2020-06-09 Snap Inc. Conditional modification of augmented reality object
US10997760B2 (en) 2018-08-31 2021-05-04 Snap Inc. Augmented reality anthropomorphization system
US10698583B2 (en) 2018-09-28 2020-06-30 Snap Inc. Collaborative achievement interface
US10778623B1 (en) 2018-10-31 2020-09-15 Snap Inc. Messaging and gaming applications communication platform
CN111199519B (zh) * 2018-11-16 2023-08-22 北京微播视界科技有限公司 特效包的生成方法和装置
CN109523465B (zh) * 2018-11-21 2023-06-23 维沃移动通信有限公司 一种滤镜处理方法和终端
CN109462727B (zh) * 2018-11-23 2022-01-25 维沃移动通信有限公司 一种滤镜调整方法及移动终端
CN109361867B (zh) * 2018-11-23 2021-06-22 维沃移动通信有限公司 一种滤镜处理方法及移动终端
US10939236B1 (en) 2018-11-30 2021-03-02 Snap Inc. Position service to determine relative position to map features
US11199957B1 (en) 2018-11-30 2021-12-14 Snap Inc. Generating customized avatars based on location information
US11032670B1 (en) 2019-01-14 2021-06-08 Snap Inc. Destination sharing in location sharing system
US10939246B1 (en) 2019-01-16 2021-03-02 Snap Inc. Location-based context information sharing in a messaging system
US11294936B1 (en) 2019-01-30 2022-04-05 Snap Inc. Adaptive spatial density based clustering
US11972529B2 (en) 2019-02-01 2024-04-30 Snap Inc. Augmented reality system
US10936066B1 (en) 2019-02-13 2021-03-02 Snap Inc. Sleep detection in a location sharing system
US10838599B2 (en) 2019-02-25 2020-11-17 Snap Inc. Custom media overlay system
US10964082B2 (en) 2019-02-26 2021-03-30 Snap Inc. Avatar based on weather
US10852918B1 (en) 2019-03-08 2020-12-01 Snap Inc. Contextual information in chat
US11868414B1 (en) 2019-03-14 2024-01-09 Snap Inc. Graph-based prediction for contact suggestion in a location sharing system
US11852554B1 (en) 2019-03-21 2023-12-26 Snap Inc. Barometer calibration in a location sharing system
US11249614B2 (en) 2019-03-28 2022-02-15 Snap Inc. Generating personalized map interface with enhanced icons
US10810782B1 (en) 2019-04-01 2020-10-20 Snap Inc. Semantic texture mapping system
CN110062157B (zh) * 2019-04-04 2021-09-17 北京字节跳动网络技术有限公司 渲染图像的方法、装置、电子设备和计算机可读存储介质
US10582453B1 (en) 2019-05-30 2020-03-03 Snap Inc. Wearable device location systems architecture
US10560898B1 (en) 2019-05-30 2020-02-11 Snap Inc. Wearable device location systems
US10893385B1 (en) 2019-06-07 2021-01-12 Snap Inc. Detection of a physical collision between two client devices in a location sharing system
US11307747B2 (en) 2019-07-11 2022-04-19 Snap Inc. Edge gesture interface with smart interactions
US11821742B2 (en) 2019-09-26 2023-11-21 Snap Inc. Travel based notifications
US11218838B2 (en) 2019-10-31 2022-01-04 Snap Inc. Focused map-based context information surfacing
US10818002B1 (en) 2019-11-26 2020-10-27 Capital One Services, Llc Systems and methods for processing images based on criteria
US11128715B1 (en) 2019-12-30 2021-09-21 Snap Inc. Physical friend proximity in chat
CN111050050A (zh) * 2019-12-30 2020-04-21 维沃移动通信有限公司 滤镜调节方法及电子设备
US11429618B2 (en) 2019-12-30 2022-08-30 Snap Inc. Surfacing augmented reality objects
US10880496B1 (en) 2019-12-30 2020-12-29 Snap Inc. Including video feed in message thread
US11343323B2 (en) 2019-12-31 2022-05-24 Snap Inc. Augmented reality objects registry
US11169658B2 (en) 2019-12-31 2021-11-09 Snap Inc. Combined map icon with action indicator
US11228551B1 (en) 2020-02-12 2022-01-18 Snap Inc. Multiple gateway message exchange
US11516167B2 (en) 2020-03-05 2022-11-29 Snap Inc. Storing data based on device location
US11619501B2 (en) 2020-03-11 2023-04-04 Snap Inc. Avatar based on trip
US10956743B1 (en) 2020-03-27 2021-03-23 Snap Inc. Shared augmented reality system
US11430091B2 (en) 2020-03-27 2022-08-30 Snap Inc. Location mapping for large scale augmented-reality
US11290851B2 (en) 2020-06-15 2022-03-29 Snap Inc. Location sharing using offline and online objects
US11483267B2 (en) 2020-06-15 2022-10-25 Snap Inc. Location sharing using different rate-limited links
US11314776B2 (en) 2020-06-15 2022-04-26 Snap Inc. Location sharing using friend list versions
US11503432B2 (en) 2020-06-15 2022-11-15 Snap Inc. Scalable real-time location sharing framework
US11308327B2 (en) 2020-06-29 2022-04-19 Snap Inc. Providing travel-based augmented reality content with a captured image
US11349797B2 (en) 2020-08-31 2022-05-31 Snap Inc. Co-location connection service
US11606756B2 (en) 2021-03-29 2023-03-14 Snap Inc. Scheduling requests for location data
US11645324B2 (en) 2021-03-31 2023-05-09 Snap Inc. Location-based timeline media content system
US11829834B2 (en) 2021-10-29 2023-11-28 Snap Inc. Extended QR code

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101751207A (zh) * 2008-12-12 2010-06-23 康佳集团股份有限公司 一种触摸屏移动终端图片显示控制方法、装置及移动终端
US20110109581A1 (en) * 2009-05-19 2011-05-12 Hiroyuki Ozawa Digital image processing device and associated methodology of performing touch-based image scaling
CN102082864A (zh) * 2010-11-22 2011-06-01 徐灏 Camare360手机摄影平台及瞬间处理方法
US20120176401A1 (en) * 2011-01-11 2012-07-12 Apple Inc. Gesture Mapping for Image Filter Input Parameters

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2003293771A1 (en) * 2002-12-10 2004-06-30 Sony Ericsson Mobile Communications Ab Creating effects for images
KR101749529B1 (ko) * 2010-11-25 2017-06-21 엘지전자 주식회사 휴대 단말기 및 그 동작 제어방법
DE202012012645U1 (de) * 2012-03-01 2013-07-11 Research In Motion Ltd. Ziehpunkt zum Anwenden von Bildfiltern in einem Bildeditor

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101751207A (zh) * 2008-12-12 2010-06-23 康佳集团股份有限公司 一种触摸屏移动终端图片显示控制方法、装置及移动终端
US20110109581A1 (en) * 2009-05-19 2011-05-12 Hiroyuki Ozawa Digital image processing device and associated methodology of performing touch-based image scaling
CN102082864A (zh) * 2010-11-22 2011-06-01 徐灏 Camare360手机摄影平台及瞬间处理方法
US20120176401A1 (en) * 2011-01-11 2012-07-12 Apple Inc. Gesture Mapping for Image Filter Input Parameters

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015169213A1 (zh) * 2014-05-07 2015-11-12 厦门美图之家科技有限公司 一种视频文件的滤镜处理方法

Also Published As

Publication number Publication date
US9674425B2 (en) 2017-06-06
CN103777852A (zh) 2014-05-07
US20150222814A1 (en) 2015-08-06
CN103777852B (zh) 2018-10-02

Similar Documents

Publication Publication Date Title
WO2014059886A1 (zh) 获取图像的方法和装置
US10788967B2 (en) Gesture mapping for image filter input parameters
TWI609317B (zh) 智慧型白板互動
US9405463B2 (en) Device and method for gesturally changing object attributes
US20120174029A1 (en) Dynamically magnifying logical segments of a view
US9128605B2 (en) Thumbnail-image selection of applications
JP4870212B2 (ja) 画像レイアウト制約生成
US10838607B2 (en) Managing objects in panorama display to navigate spreadsheet
US20130139102A1 (en) Systems and Methods for Organizing and Displaying Hierarchical Data Structures in Computing Devices
US20130027425A1 (en) Superimposed annotation output
US20160266878A1 (en) Automatic wireframing using images
WO2015130270A1 (en) Photo and document integration
TWI541748B (zh) 股票多模式詳細資訊裝置與方法
WO2018198703A1 (ja) 表示装置
WO2014019207A1 (zh) Widget处理方法、装置及移动终端
US20220155948A1 (en) Offset touch screen editing
JP2018073125A5 (ja) サーバ装置、情報処理方法、及びプログラム
US20160132478A1 (en) Method of displaying memo and device therefor
US20190088237A1 (en) System and Method of Generating Signals from Images
EP2911115B1 (en) Electronic device and method for color extraction
JP6448500B2 (ja) 画像処理装置、画像処理方法
CN111510646A (zh) 拼接屏的视频显示方法、显示装置、计算机设备和介质
US20160328215A1 (en) Pattern recognition of software program code in an integrated software development environment
US20130205201A1 (en) Touch Control Presentation System and the Method thereof
WO2016033934A1 (zh) 一种投影方法、投影装置及计算机存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13846638

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205N DATED 24/06/2015)

122 Ep: pct application non-entry in european phase

Ref document number: 13846638

Country of ref document: EP

Kind code of ref document: A1