WO2015132622A1 - Single mode & automatically visual media capture and excuting associate functions(s) - Google Patents

Single mode & automatically visual media capture and excuting associate functions(s) Download PDF

Info

Publication number
WO2015132622A1
WO2015132622A1 PCT/IB2014/059369 IB2014059369W WO2015132622A1 WO 2015132622 A1 WO2015132622 A1 WO 2015132622A1 IB 2014059369 W IB2014059369 W IB 2014059369W WO 2015132622 A1 WO2015132622 A1 WO 2015132622A1
Authority
WO
Grant status
Application
Patent type
Prior art keywords
video
electronic device
user
visual media
input
Prior art date
Application number
PCT/IB2014/059369
Other languages
French (fr)
Inventor
Yogesh Chunilal Rathod
Original Assignee
Onesnaps Technology Private Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, TV cameras, video cameras, camcorders, webcams, camera modules for embedding in other devices, e.g. mobile phones, computers or vehicles
    • H04N5/225Television cameras ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, camcorders, webcams, camera modules specially adapted for being embedded in other devices, e.g. mobile phones, computers or vehicles
    • H04N5/232Devices for controlling television cameras, e.g. remote control ; Control of cameras comprising an electronic image sensor
    • H04N5/23216Control of parameters, e.g. field or angle of view of camera via graphical user interface, e.g. touchscreen
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/30Information retrieval; Database structures therefor ; File system structures therefor
    • G06F17/30244Information retrieval; Database structures therefor ; File system structures therefor in image databases
    • G06F17/3028Information retrieval; Database structures therefor ; File system structures therefor in image databases data organisation and access thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/30Information retrieval; Database structures therefor ; File system structures therefor
    • G06F17/3074Audio data retrieval
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/002Specific input/output arrangements not covered by G06F3/02 - G06F3/16, e.g. facsimile, microfilm
    • G06F3/005Input arrangements through a video camera
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0346Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04817Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object or an image, setting a parameter value or selecting a range
    • G06F3/04842Selection of a displayed object
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object or an image, setting a parameter value or selecting a range
    • G06F3/04845Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object or an image, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object or an image, setting a parameter value or selecting a range
    • G06F3/04847Interaction techniques to control parameter settings, e.g. interaction with sliders, dials
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for entering handwritten data, e.g. gestures, text
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network-specific arrangements or communication protocols supporting networked applications
    • H04L67/06Network-specific arrangements or communication protocols supporting networked applications adapted for file transfer, e.g. file transfer protocol [FTP]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, TV cameras, video cameras, camcorders, webcams, camera modules for embedding in other devices, e.g. mobile phones, computers or vehicles
    • H04N5/225Television cameras ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, camcorders, webcams, camera modules specially adapted for being embedded in other devices, e.g. mobile phones, computers or vehicles
    • H04N5/232Devices for controlling television cameras, e.g. remote control ; Control of cameras comprising an electronic image sensor
    • H04N5/23245Operation mode switching of cameras, e.g. between still/video, sport/normal or high/low resolution mode
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, TV cameras, video cameras, camcorders, webcams, camera modules for embedding in other devices, e.g. mobile phones, computers or vehicles
    • H04N5/225Television cameras ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, camcorders, webcams, camera modules specially adapted for being embedded in other devices, e.g. mobile phones, computers or vehicles
    • H04N5/232Devices for controlling television cameras, e.g. remote control ; Control of cameras comprising an electronic image sensor
    • H04N5/23293Electronic viewfinders
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback

Abstract

An electronic device includes digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a touch controller to identify haptic contact engagement, haptic contact persistence and haptic contact release on the display; present the one or more input labels or icons on display; and based on haptic contact engagement and haptic contact release (single tap or double tap) on said particular label, enabling a visual media capture controller to capture or record the visual media as a photograph or video or voice or start recording of a video and triggering or invoking or presenting or executing of one or more label associated functions. In another embodiment an electronic device includes digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a sensor(s) including gyroscope, accelerometer, proximity sensor, compass, audio & gesture sensors to identify device/image orientation, proximity & stability on the display; and an auto visual media capture controller to alternately record the visual media as a photograph or a video based upon change in orientation and/or change proximity and/or stability of device or image within particular period of time or based upon evolution of the time period between the change in orientation and/or change proximity and/or stability of device or image.

Description

TITLE

Single mode & automatically visual media capture and executing associate function(s) FIELD OF INVENTION

The present invention relates generally to enabling user to single or double touch or tap preferred input icon or label(s) from set of default, pre-defined, auto presented or auto generated input icons or labels presented on display for capturing video or photo and execute associated functions or instructions for conducting one or more tasks, actions, processes, events, transactions, and workflow related to captured photo or video. In another embodiment present invention also enables auto mode capturing of photo or video, without any touch or tap(s) on display or hands free mode.

BACKGROUND OF THE INVENTION

At present users are using various photos or videos sharing applications for sharing photos and videos to connected or related users. After capturing photo or video user is required to specify the receivers of captured photos and videos, sources or destinations for posting or sharing of captured photos or videos, user actions including like, save, add to album, bookmark, invite, ask other user' s to reply, provide like or comments, types or categories of photos or videos, apply selected filters, sent to particular users, sent via particular methods of communication e.g. email, mms, web services and one or more types of communication interfaces, ask photo associate one or more types of information to one or more sources, start chat for captured & shared photos & videos. So it will consume lot of time of user to clicking or tapping or touching for conducting said tasks, to-dos, functions and actions with captured photos or videos. Present invention enabling user to touch or tap preferred input icon from set of input icons presented on display for capturing video or photo and execute associate functions or instructions for conducting one or more tasks, actions, events, transactions, workflow and saves users lot of time.

At present various photo or video capturing or recording applications provided with touch enabled button(s) to capture or record video or photo or voice. But user is engaged in selecting photo or video button, touch or tap on button each time capturing or recording of photo or starting or pausing or stopping of video. Present invention provides hands free environment for capturing photos or videos and enables user to automatically capture photo and/or video in single mode or session describe in detail in Figure 14.

Therefore, it is with respect to these considerations and others that the present invention has been made.

OBJECT OF THE INVENTION

The principal object of the present invention is to enabling user to use single mode capturing photos or videos as well as automatically conduct, execute, trigger associate or user preferred one or more tasks, activities, functions, actions, to-dos with said captured or recorded media including videos, photos, images, voice audio & sound by touching or clicking or tapping on preferred pre-defined input icon or label from presented one or more input icons or labels on display. The principal object of the present invention is to enable user to quick capture, record, select, share, post, sent, present media including videos, audio, voice, sound, images, photos, text and/or conduct one or more associate tasks, actions, events,, transactions, applications, workflow and participate with other users and make feel user always connected, logged-in, always on, and always engaged.

The main object of the present invention is to enable user to automatically capture or record, store, preview, save & share either photo or video in single mode.

DETAIL DESCRIPTION OF THE PREFERRED EMBODIMENT

The present invention now will be described more fully hereinafter with reference to the accompanying drawings, which form a part hereof, and which show, by way of illustration, specific exemplary embodiments by which the invention may be practiced. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. Among other things, the present invention may be embodied as methods or devices. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. The following detailed description is, therefore, not to be taken in a limiting sense.

Throughout the specification and claims, the following terms take the meanings explicitly associated herein, unless the context clearly dictates otherwise. The phrase "in one embodiment" as used herein does not necessarily refer to the same embodiment, though it may. Furthermore, the phrase "in another embodiment" as used herein does not necessarily refer to a different embodiment, although it may. Thus, as described below, various embodiments of the invention may be readily combined, without departing from the scope or spirit of the invention.

In addition, as used herein, the term "or" is an inclusive "or" operator, and is equivalent to the term "and/or," unless the context clearly dictates otherwise. The term "based on" is not exclusive and allows for being based on additional factors not described, unless the context clearly dictates otherwise. In addition, throughout the specification, the meaning of "a," "an," and "the" include plural references. The meaning of "in" includes "in" and "on."

Briefly stated, the present invention is directed towards enabling single mode & automatically capturing of visual media and executing associate function(s) In one embodiment present an electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a touch controller to identify haptic contact engagement, haptic contact persistence and haptic contact release on the display; present the one or more input labels or icons on display; and based on haptic contact engagement and haptic contact release (single tap or double tap) on said particular presented label, enabling a visual media capture controller to record the visual media as a photograph or video or voice or start recording of a video and triggering or invoking or presenting or executing one or more associate (auto identified, set, default, user selected) functions, events, transactions, processes, formatting, web services, interfaces, instructions, triggers, scripts, routines, object, modules, actions including sending (sharing, advertising, broadcasting, publishing & synchronizing) of recorded photograph or a video.

In another embodiment, label can presented with or alternatively present images or digital representation or dynamically generated strings including names of contacts/connections, group(s) of contacts/connections, web site or email address, user names, folder names, menu names, named entities, keywords, categories, taxonomies, ontology, key phrases, date & time, brand or product names, name or type of activities, actions, events, transactions, location & status, sources, destinations, nodes, identities.

In another embodiment, enabling user to select or operate visual media capture controller(s) based upon haptic contact engagement (single tap, double tap) or via digital or physical keyboard or via voice command(s).

In another embodiment, enabling user to search, match, select, make payment, purchase, download, register, upload, install, view details, un-install, sort, order, group, filter, rank, update, upgrade one or more functions from one or more developers, providers and sources.

In another embodiment, enabling user to set, make as default, apply, attach, associate, and select one or more functions with one or more or group or set of input icons or labels. In another embodiment, automatically select one or more functions with one or more input icons or labels.

In another embodiment, select one or more functions with one or more input icons or labels based on user defined or user programmed script(s) including execute, invoke, triggers, orchestrate functions, interfaces based on rules, commands, Boolean operators, programming languages.

In another embodiment, automatically and/or in background mode or silent mode execute, invoke, triggers, orchestrate function(s), interface(s) while capturing photo or video/recording of video or voice or run-time of capturing of video/recording of voice. In another embodiment, enabling user to further selecting or updating already selected or associated one or more functions while capturing of photo or video and/or recording of video or voice.

In another embodiment, automatically and/or in background mode or silent mode execute, invoke, triggers, orchestrate function(s), interface(s) while capturing video/recording of voice or run-time of capturing of video/recording of voice and present contextual contents including location information, information about associate or recognized or detected or sensed nodes, named entities, people, brands, products, items, objects, details associate with one or more or each or run-time or real-time capturing of video related photo frames.

In another embodiment, instructing user to input, select structured or un- structured details including form(s), interface(s) before executing of function(s) associate with input icon(s) or label(s).

In another embodiment, presenting controls, user actions, interface(s) before executing of function(s) associate with input icon(s) or label(s). In another embodiment, the visual media capture controller presents a single mode input labels or icons or images on the display to receive the haptic contact engagement, haptic contact persistence and haptic content release or single tap or double tap or other types of touches or multi touches for capturing or recording photo or video or voice as well as executing, invoking or triggering of associate one or more functions.

In another embodiment, the visual media capture controller selectively stores the photograph in a photograph library.

In another embodiment, the visual media capture controller selectively stores the video in a video library.

In another embodiment, end recording of started video and execute input icon(s) or label(s) associate function(s) after further receiving the haptic contact engagement by visual media capture controller. In another embodiment, end recording of started video and execute input icon(s) or label(s) associate function(s) after set particular duration including set number of seconds or minutes.

In another embodiment, the recognized touch input is identified as a short touch, a long touch, a multi-position touch, a touch & drag, tap, single tap, double tap, press & hold while, long press, two finger tap, pinch open and one or more types of touch or multi-touch gestures.

In another embodiment, the general image capture is performed when the identified type of touch input is the single tap or short touch; and the continuous image capture is performed when the identified type of touch input is the double tap.

In another embodiment, the visual media capture controller includes instructions executed by a processor. In another embodiment, further comprising a wireless signal processor, a power control circuit, one or more types of sensors including Accelerometer, Gyroscope, Gesture Sensor, Proximity Sensor, Direction sensor & Audio sensor and a global positioning processor connected to the processor. In another embodiment, an electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a touch controller to identify haptic contact engagement, haptic contact persistence and haptic contact release on the display; present the labels or icons or images or digital representation of contacts/connections and group(s) of contacts/connections on display as a input icons or labels for recording or capturing photograph or starting recording or capturing of a video, wherein a visual media capture controller to record the visual media as a photograph or start recording of a video and triggering of sending, sharing, advertising, broadcasting, publishing & synchronizing of recorded photograph or a video. In another embodiment, end recording of started video and send a recorded video after to selected label or icon or image or digital representation associate contact(s)/connection(s)/group(s) of contacts/connections further touching single visual media capture controller. In another embodiment, end recording of started video and send a recorded video to selected label or icon or image or digital representation associate contact(s)/connection(s)/group(s) of contacts/connections after set particular duration including set number of seconds or minutes. In another embodiment, the visual media capture controller presents a single mode input labels or icons or images of user's contacts/group(s) of contacts on the display to receive the haptic contact engagement, haptic contact persistence and haptic content release and triggering of sending or sharing of recorded photograph or video to label or icon or image or digital representation associate contact(s)/connection(s)/group(s) of contacts/connections.

In another embodiment, the visual media capture controller selectively stores the photograph in a photograph library.

In another embodiment, the visual media capture controller invokes a photograph preview mode and present media including photograph or video associate chat session interface.

In another embodiment, the chat interface enables sender user and receiving users to post messages & voice messages, view posted messages & listen voice messages, auto matched contextual contents based on posted messages, voice messages and participant users' data including profile & user data, view associate statistics including send to users, received users, viewed users, pending to view users, participated users in current media including photo/ video related chat session, likes by chat session participated users.

In another embodiment, the visual media capture controller invokes a photograph preview mode and present indications of receiving of other users' shared media including photos & videos via number icon.

In another embodiment, enable to switch among presented number icons for participate in shared media including photos & video associate chat session(s) for sending and receiving messages, voice messages, sound, edited media including photo or video, likes, location information, presence information, status, emoticons.

In another embodiment, number icon including sender name or identity or image, profile link, availability status, presence information.

In another embodiment, remove number icon after set number of time period. In another embodiment, remove number icon after expiring of chat session based on triggering of no single character inputted or selected or send by all participated users within set number of time period.

In another embodiment, add or update number icon after receiving shared media from other users or further capturing & sharing or sending of shared media to other users.

In another embodiment, based on settings the visual media capture controller invokes a visual media capture mode.

In another embodiment, the visual media capture controller selectively stores the video in a video library. In another embodiment, based on settings the visual media capture controller invokes a video preview mode.

In another embodiment, based on settings the visual media capture controller invokes a visual media capture mode.

In another embodiment, enable shared media receiving users to view shared media and associate chat conversations or messages or voice messages after expiration of chat session of other users.

In another embodiment, force remove received shared media of other users after set number of time period.

In another embodiment, enable user to scrolling presented number icons.

In another embodiment, enable user to paging presented contacts and group(s) of contacts on display.

In another embodiment, notify receiving user(s) about shared media including photos, videos, text, chat messages, comments, likes, voice messages, sounds, emoticons, locations, sending & receiving date & time, expiration or removal duration or time period of shared media and information about sender(s) including profile link, photo or image, name & identity, availability status, user posted status, presence information and information about shared media associate chat session related participant users including profile link, photo or image, name & identity, availability status, user posted status, presence information, In another embodiment, enabling user to opening application from notification and presenting each active shared media of each sending user with chat interface to each receiving user(s). In another embodiment, the visual media capture controller stores the photograph upon haptic contact engagement.

In another embodiment, the visual media capture controller includes instructions executed by a processor.

In another embodiment, further comprising a wireless signal processor, a power control circuit and a global positioning processor connected to the processor.

In another embodiment, present labels or icons or images or digital representation of contact(s)/connection(s)/group(s) of contacts/connections in switchable show and hide mode.

In another embodiment, switchable shows and hide mode based upon set time period.

In another embodiment, enabling user to search, match, select, remove or hide, add or attach or show, sort, rank, order or arrange, filter, view associate profile, details & statistics, group, page- wise group contacts or connections from display.

In another embodiment, enabling user to set particular period of time for one or more contacts or group(s) of contacts for removing shared photo or video from server and/or receiver's or receivers' device or application or database.

In another embodiment, dynamically change colors of label based on background color of photo or video. In another embodiment, an electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a touch controller to identify haptic contact engagement, haptic contact persistence and haptic contact release on the display; a presenting one or more types of input icon or label on display; enabling user to select one or more types of input icon or label based upon haptic contact engagement including single touch, double touch, one-tap, double-tap & one or more types of touches or multi touches gestures on respective icon or label of particular type of input icon or label; enabling user to select preferred input icon or label from one or more or set of input icons or labels related to particular type of input icon or label for recording the visual or voice media as a photograph or a video or a voice as well as executing or triggering or invoking associate one or more functions or commands or instructions or routines. In another embodiment, enabling user to select one or more types of input icons or labels for using or accessing one or more types of functions.

In another embodiment, enabling user to select one or more input icons or labels of one or more types of input icons or labels based upon haptic contact engagement on preferred input icon or label.

In another embodiment, one or more types of input icons or labels including send captured photo or video or voice to particular user or contact or group of contacts and user is enable to select said "Send To" type of input icon or label, wherein presenting said "Send To" type of input icon or label to capture photo or video and sent to particular user or contact or group within single touch or haptic contact engagement.

In another embodiment, executing or triggering or invoking associate one or more functions or commands or instructions or routines during or at the end of capturing or recording of visual or voice media.

In another embodiment, enabling user to search, match, select, make payment, purchase, download, register, upload, install, view details, un-install, sort, order, group, filter, rank, update, upgrade one or more functions from one or more developers, providers and sources.

In another embodiment, enabling user to set, make as default, apply, attach, associate, select one or more functions with one or more or group or set of input icons or labels.

In another embodiment, automatically select one or more functions with one or more input icons or labels.

In another embodiment, select one or more functions with one or more input icons or labels based on user defined or user programmed script(s) including execute, invoke, triggers, orchestrate functions, interfaces based on rules, commands, Boolean operators, programming languages. In another embodiment, automatically and/or in background mode or silent mode execute, invoke, triggers, orchestrate function(s), interface(s) while capturing photo or video/recording of video or voice or run-time of capturing of video/recording of voice. In another embodiment, enabling user to further selecting or updating already selected or associate one or more functions while capturing of photo or video and/or recording of video or voice.

In another embodiment, automatically and/or in background mode or silent mode execute, invoke, triggers, orchestrate function(s), interface(s) while capturing video/recording of voice or run-time of capturing of video/recording of voice and present contextual contents including location information, information about associate or recognized or detected or sensed nodes, named entities, people, brands, products, items, objects, details associate with one or more or each or run- time or real-time capturing of video related photo frames.

In another embodiment, instruct user to input, select structured or un-structured details including form(s), interface(s) before executing of function(s) associate with input icon or label(s).

In another embodiment, present controls, user actions, interface(s) before executing of function(s) associate with input icon or label(s).

In another embodiment, the visual media capture controller presents a single mode input labels or icons or images on the display to receive the haptic contact engagement, haptic contact persistence and haptic content release for capturing or recording photo or video or voice as well as executing, invoking or triggering of associate one or more functions.

In another embodiment, the visual media capture controller selectively stores the photograph in a photograph library. In another embodiment, the visual media capture controller selectively stores the video in a video library.

In another embodiment, end recording of started video and execute input icon(s) or label(s) associate function(s) after further receiving the haptic contact engagement by visual media capture controller. In another embodiment, end recording of started video and execute input icon(s) or label(s) associate function(s) after set particular duration including set number of seconds or minutes.

In another embodiment, the visual media capture controller includes instructions executed by a processor.

In another embodiment, further comprising a wireless signal processor, a power control circuit and a global positioning processor connected to the processor. In another embodiment, an electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a touch controller to identify haptic contact engagement, haptic contact persistence and haptic contact release on the display; and enabling to create or generate or auto generate one or more input icons or labels or one or more types of input icons or labels based on attaching, associating, selecting, setting, applying, customizing, defining, updating & programming one or more selected or auto matched or customize function(s) or group(s) of functions or types of functions, features, user actions, objects, applications, instructions, scripts, forms, controls, widgets, one or more types of contents & database or data, interfaces, web services with one or more input icons or labels or one or more types of input icons or labels .

In another embodiment, enabling to create or generate or auto generate input icons or labels based on user selected one or more function(s) or customized functions including selection of parameters, one or more contacts, group of contacts, web sites, sources, destinations, emails & Instant Messenger addresses & associate user identity, password or authentication information, one or more types of contents including templates, tags, suggested contents, keywords, questions, answers, captions, title, categories, images, photos, voice, video, audio, text, links, database or data, structured data, files or documents, emoticons, user profile, user data & commands, interfaces, applications, objects, services, widgets, controls, functions, features, keywords, categories, types, key phrases, name of folders, nodes, named entities and any combination thereof.

In another embodiment, enabling user to search, match, select, make payment, purchase, download, register, upload, install, view details, un-install, sort, order, group, filter, rank, update, upgrade one or more functions from one or more developers, providers and sources. In another embodiment, select one or more functions with one or more input icons or labels based on user defined or user programmed script(s) including execute, invoke, triggers, orchestrate functions, objects, modules, instructions, triggers, applications, interfaces, parameters based on rules, commands, Boolean operators, one or more types or programming language syntax, scripts.

In another embodiment, automatically and/or in background mode or silent mode execute, invoke, triggers, orchestrate function(s), interface(s) while capturing photo or video/recording of video or voice or run-time of capturing of video/recording of voice.

In another embodiment, instructing user to input, select structured or un- structured details including form(s), interface(s) before executing of function(s) associate with input icon(s) or label(s). In another embodiment, presenting controls, user actions, interface(s) before executing of function(s) associate with input icon(s) or label(s).

In another embodiment, the visual media capture controller presents a single mode input labels or icons or images on the display to receive the haptic contact engagement including one touch, two or more or multi touch and/or haptic contact persistence and/or haptic content release for capturing or recording photo or video or voice as well as executing, invoking or triggering of associate one or more functions.

In another embodiment, an electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a touch controller to identify haptic contact engagement, haptic contact persistence and haptic contact release on the display; enabling to create or generate or auto generate one or more input icons or labels or one or more types of input icons or labels based on attaching, associating, selecting, setting, applying, customizing, defining, updating & programming one or more selected or auto matched or customize function(s) or group(s) of functions or types of functions, features, user actions, objects, applications, instructions, scripts, forms, controls, widgets, one or more types of contents & database or data, interfaces, web services with one or more input icons or labels or one or more types of input icons or labels; a presenting one or more input icons or labels or one or more types of input icons or labels ; and enabling to select or select via one-touch, two touch or multi-touch or based upon haptic contact engagement one or more visual media capture controller(s). In another embodiment, enabling user to select one or more types of input icons or labels and related one or more input icons or labels via one -touch, two touch or multi-touch or based upon haptic contact engagement for recording media/photo/video/voice/content as well as executing, invoking, running, presenting, connecting, downloading, installing, opening, triggering one or more functions, objects, applications, web pages, widgets, web services, interfaces, wizards, programming scripts, instructions or source code, one or more types of contents including text, images, photos, audio, voice, video, files or documents, database, links, emoticons, location information, status & presence information, user profile, user data, user connections or contacts.

In another embodiment, dynamically present one or more input icons or labels or one or more types of input icons or labels based on updates in date, time, date & time ranges, updates, status, presence information, location, holidays or vacations, planned activates, actions, events, status, updates, & transactions, check-in place, user & connected users' profile & user data.

The many features and advantages of the invention are apparent from the detailed specification and, thus, it is intended by the appended claims to cover all such features and advantages of the invention that fall within the true spirit and scope of the invention. Further, since numerous modifications and changes will readily occur to those skilled in the art, it is not desired to limit the invention to the exact construction and operation illustrated and described, and accordingly all suitable modifications and equivalents may be resorted to, falling within the scope of the invention.

BRIEF DESCRIPTION OF THE DRAWINGS

Non-limiting and non-exhaustive embodiments of the present invention are described with reference to the following drawings. In the drawings, like reference numerals refer to like parts throughout the various figures unless otherwise specified.

For a better understanding of the present invention, reference will be made to the following Detailed Description, which is to be read in association with the accompanying drawings, wherein:

FIG. 1 illustrates components of an electronic device implementing single mode visual media capture and executes associate functions in accordance with the invention;

FIG. 2 illustrates processing operations associated with an embodiment of the invention.

FIG. 3 illustrates the exemplary graphical user interface for implementing single mode visual media capture and executes associate functions in accordance with the invention;

FIG. 4 illustrates the exemplary graphical user interface for presenting chat interface and chat session associate with particular shared or received media or contents enabling user to participate with current chat session and communicate with other chat participants;

FIG. 5 illustrates the exemplary graphical user interface for defining, creating, configuring, updating, adding, removing, customizing personalizing & generating one or more group(s) of or one or more label(s) of one or more group(s) for capturing photo or video and executing associate functions;

FIG. 6 illustrates the exemplary graphical user interface for selecting one or more typers of functions;

FIG. 7 illustrates the exemplary graphical user interface for describing one of the example or embodiment of present invention;

FIG. 8 illustrates the exemplary graphical user interface for describing another example or embodiment of present invention; FIG. 9 illustrates the exemplary graphical user interface for enabling sender and/or receiver to communicate with each other on shared media or contents; FIG. 10 illustrates the exemplary graphical user interface for describing another example or embodiment of present invention;

FIG. 11 illustrates the exemplary graphical user interface for describing another example or embodiment of present invention;

FIG. 12 illustrates the exemplary graphical user interface for describing another example or embodiment of present invention;

FIG. 13 illustrates the exemplary graphical user interface for describing another example or embodiment of present invention;

FIG. 14 illustrates components of an electronic device implementing auto & single mode visual media capture and executes associate functions in accordance with the invention; DETAILED DESCRIPTION OF THE DRAWINGS

FIG. 1 illustrates an electronic device 100 implementing operations of the invention. In one embodiment, the electronic device 100 is a smart phone with a processor 130 in communication with a memory 136. The processor 130 may be a central processing unit and/or a graphics processing unit. The memory 136 is a combination of flash memory and random access memory. The memory 136 stores a visual media capture controller(s) 140 and/or auto visual media capture controller(s) 138 to implement operations of the invention. The user touch or tap single mode input icon or label from set of presented input icons or labels 140 includes executable instructions to record a photograph or a video and execute associate one or more functions or instructions based upon the processing of haptic signals, as discussed below.

The user is presented with auto generated, auto presented, user created & selected one or more or group or set of single mode input icons or labels on the display 110. The visual media controller(s) 140/145 and/or auto visual media capture controller(s) 138 interacts with a photograph library controller 150, which includes executable instructions to store, organize and present photos 155. The visual media controller 140/145 and/or auto visual media capture controller(s) 138 also interacts with a video library controller 160, which includes executable instructions to store, organize and present videos 165. The visual media controller(s) 140/145 and/or auto visual media capture controller(s) 138 interacts with a functions library controller 167, which include executable instructions to store, updates, organize and execute function(s) 168.

The processor 130 is also coupled to image sensors 105. In another embodiment processor 130 is also coupled to other sensors including Accelerometer, Gyroscope, Gesture Sensor, Proximity Sensor, Direction sensor & Audio sensor 190. The image sensors 105 / 190 may be known digital image sensors. The image sensors capture visual media, which is presented on display 110. That is, in a visual media mode controlled by the visual media capture controller 140/145 and/or auto visual media capture controller(s) 138, the image sensors 105 capture visual media and present the visual media on the display 110 so that a user can observe the captured visual media.

A touch controller 115 is connected to the display 110 and the processor 130. The touch controller 115 is responsive to haptic signals applied to the display 110. In one embodiment, the visual media capture controller 140 and/or auto visual media capture controller(s) 138 presents an auto generated, user defined or created, user selected & auto presented one or more or group or set of input icons or labels input icon(s) on the display 110. That is, the visual media capture controller 140 and/or auto visual media capture controller(s) 138 includes executable instructions executed by the processor to present an auto generated, user defined or created, user selected & auto presented one or more or group or set of input icons or labels input icon(s) on the display 110.

The visual media capture controller 140 communicates with the processor 130 regarding haptic signals applied to the display 110, which are recorded by the touch controller 115. In one configuration, the visual media capture controller 140 processes haptic signals app ed to the user touched or tapped ser clicked or user selected input icon, as detailed in connection with the discussion of FIG. 2, and determines to execute associate functions or instructions along with capturing or recording of a photograph or a video and, as discussed below.

Auto visual media capture controller(s) 138 communicates with the processor 130 regarding auto identified or detected or sensed one or more types of senses derived from one or more sensors including Accelerometer, Gyroscope, Gesture Sensor, Proximity Sensor, Direction sensor & Audio sensor 1 0 applied to the display 110, which may also recorded by the touch controller 115. The electronic device 100 may also include other components commonly associated with a smart phone, such as a wireless signal processor 120, a power control circuit 125 and a global positioning system processor 135. While many of components of FIG. 1 are known in the art, new functionality is achieved through the visual media capture controller(s) 140/145. FIG. 2 illustrates processing operations associated with the visual media capture controller 140. Initially, a visual media capture mode is invoked 201. For example, a user may access an application presented on display 110 to invoke a visual media capture mode. FIG. 3 illustrates the exterior of electronic device 100. The figure also illustrates the display 110. The electronic device 100 is in a visual media capture mode and presents visual media 300. The display 1 10 also includes an input icon(s) or label(s) 302. In one embodiment, the amount of time that a user presses the preferred input icon(s) or label(s) 302 determines whether a photograph will be recorded or a starting of a video and execute input icon or label associate function(s). For example, if a user initially intends to take a photograph, then the icon(s) or label(s) 302 is engaged with a haptic signal. If the user decides that the visual media should instead be a video, the user continues to engage the icon(s) or label(s) 302. If the engagement persists for a specified period of time (e.g., 3 seconds), then the output of the visual media capture is determined to be video and starting of recording of a video. The video mode may be indicated on the display 110 with an icon 304. If the user decides to stop the video, the user further engages the icon(s) or label(s) 302 or an icon 304. When video capturing or recording is stops that triggers execution of input icon or label associate function(s).

In one embodiment user is presented with Input Icon(s) or Label(s) 207/302 by visual media capture controller 140 on display 110. In another one embodiment user is presented with categories 350 Input Icon(s) or Label(s) 207/302 by visual media capture controller 140 on display 110.

In another embodiment user can select categories 350 of Input Icon(s) or Label(s) 207/302 by visual media capture controller 140 on display 110. In another embodiment user can access default Input Icon(s) or Label(s) 207/302 by visual media capture controller 140 on display 110.

In one embodiment user is presented with auto generated Input Icon(s) or Label(s) 207/302 by visual media capture controller 140 on display 110 based on user's one or more activities, actions, senses, events, transactions, location(s), check-in place(s0, status, updates, presence information, user profile, user data, user connections and any combination thereof.

In another embodiment user can create, search, match, browse, select, add, remove, attach, detach, show, hide, order, sort, rank, filter, page wise group, group, lists, categories Input Icon(s) or Label(s).

In another embodiment system can automatically or semi-automatically can define, generate, create, search, match, select, add, remove, attach, detach, show, hide, order, sort, rank, filter, page wise group, group, & categories Input Icon(s) or Label(s).

Input Icon(s) or Label(s) may comprises button(s), image(s), photo(s), one or more types of digital representation(s), text including names, named entities, identities, nodes, keywords, key phrases, categories, taxonomies, contacts, connections, commands.

FIG. 2, haptic contact engagement is identified 210. For example, the haptic contact engagement may be at icon 302 on display 110. The touch controller 115 generates haptic contact engagement signals for processing by the visual media capture controller 140 in conjunction with the processor 130. Alternately, the haptic contact may be on any other icon or label on the display 110.

Identify haptic contact engagement area associate label(s), icon(s), button(s), image(s), digital representation(s) and associate applied, set, programmed, attached functions, privacy settings, function(s) settings, user provided or selected data, preferences, attributes, properties, associate data, system data & metadata 210.

Based on haptic contact engagement timer is started 215 and timer is stop 225 when haptic contact release 220. If threshold 230 not exceeded or user one-taps or single-touch or short-touch 231 then photo is captured 285 and stored 290 and Input Icon(s) or Label(s) associate functions are automatically and/or semi-automatically executed 295. If threshold 230 exceeded or user one-taps or single-touch or short-touch 232 then video capturing is started 235 and timer is started 235. If User applies or set maximum duration for length of video (e.g. set particular period of time - e.g. number of seconds) 241 and if threshold exceeded then stop timer and stop video 245 and store video 250 and execute user touched or selected or clicked single mode input icon or label associate function(s) (e.g. capture video and sent or share said recorded video to one or more or group of contacts) 275. If User not applies or set maximum duration for length of video and if threshold not exceeded then keep recording or capturing of video continue up-to haptic contact engagement 260 & haptic contact release 263 and stop video & stop timer 265 and execute user touched or selected or clicked single mode input icon or label associate function(s) (e.g. capture video and sent or share said recorded video to one or more or group of contacts) 275. The video or photo is recorded by the processor 130 operating in conjunction with the memory 136. The timer is executed by the processor 130 under the control of the visual media capture controller 140. Started video 235 continues to record and the timer 235 continues to ran up-to further haptic contact engagement response on the display. If threshold not exceeded and user wants to stop video then user engage in haptic contact engagement or touch on icon 304. Haptic contact release is subsequently identified 263. The timer is then stopped, as is the recording of video 265. If threshold exceeded (e.g. set time of period e.g. 10 seconds) then video automatically stopped and timer is stopped 245 and video is stored 250 and Input Icon(s) or Label(s) associate functions are automatically and/or semi-automatically executed 275.

The elapsed time recorded by the timer is then evaluated by the visual media capture controller 140 against a specified threshold (e.g. set time of period e.g. 10 seconds). If the threshold is exceeded (240— Yes), then video is stored 212. In particular, a video is sent to the video library controller 165 for handling. In one embodiment, the visual media capture controller 140 includes executable instructions to prompt the video library controller to enter a video preview mode. Consequently, a user can conveniently review a recently recorded video. In an alternate embodiment, the visual media capture mode is responsive to a first haptic contact signal (e.g., one tap) 231 to record a photograph and a second haptic contact signal (e.g., two taps) 232 to record a video. In this case, there is not persistent haptic contact, but different visual media modes are easily entered. Indeed, the visual media capture controller 140 may be configured to interpret two taps within a specified period as an invocation of the video mode. This allows a user to smoothly transition from intent to take a picture to a desire to record video. In another embodiment based on single touch or short touch, photo mode is initiated 231 and based on long touch or double touch or touch more than particular period of time then video mode initiated 232.

For example, if a user initially intends to take a photograph, then the icon(s) or label(s) 302 is engaged with a haptic signal. If the user decides that the visual media should instead be a video, the user continues to engage the icon(s) or label(s) 302 or double-touch the icon(s) or label(s) 302. If the engagement persists for a specified period of time and haptic contact release happened (e.g., 3 seconds), then the output of the visual media capture is determined to be video and video capturing or recording is started 235. The video mode may be indicated on the display 110 with an icon 304. If the user decides to stop the video, the user further engages the icon(s) or label(s) 302 or an icon 304. In another embodiment dynamically or automatically or semi-automatically search, match, select, activate, generate, create, sort, order, rank, filter, add, attach, detach, remove, invoke, update and present one or more or set or categories or group of input icon and/or label based on user senses and user device associate sensors related data, wherein user and/or user device(s) associate sensor(s) comprises plurality types of sensors including accelerometer, gyroscopes, gesture sensor, proximity sensor, direction sensor (magnetic compass), audio sensor

(microphone) - voice command, image sensor, localization (GPS) sensors, biometric or fingerprint sensors, touch screen sensors, and many other available sensors.

In another embodiment dynamically or automatically or semi-automatically search, match, select, sort, order, rank, filter, install, uninstall, purchase, make payment, subscribe, add, add or update with existing functions, attach, detach, remove, invoke, activate, execute, upgrade, update and present functions associate with one or more or set or categories or group of input icon and/or label based on user senses and user device associate sensors related data, wherein user and/or user device(s) associate sensor(s) comprises plurality types of sensors including accelerometer, gyroscopes, gesture sensor, proximity sensor, direction sensor (magnetic compass), audio sensor (microphone) - voice command, image sensor, localization (GPS) sensors, biometric or fingerprint sensors, touch screen sensors, and many other available sensors. In another embodiment user can use single-tap or double-tap to show/hide input all or partial icons or labels 302 from display 110.

In another embodiment notify user about availability of input icons or labels associate users, contacts or sources or connections via change in color of input icons or labels. For example turn color of input icons or labels into green when associate user or contact is available or online.

In another embodiment notify user about numbers of shared contents received, sent, pending to view, view by sender, view by receiver from icons or labels associate users via adding various numbers to input icons or labels. For example "3-Amita" input label 302 shows that connected user Amita sent to user 3 shared contents or chat messages including videos or photos or voice or text messages.

In another embodiment user can start chat or chat session 365 for captured or posted or shared or presented photos or videos discus in details in Figure 4.

In another embodiment automatically start chats or chat session 365 for captured or posted or shared or presented photos or videos discus in details in Figure 4. In another embodiment user is enable to paging next or previous or scroll up or down for viewing and selecting input icons or labels.

In another embodiment user is enable to manually search, view & select one or more contacts or group(s) of contacts from all contact 353 for sharing captured or selected media including videos or photos or voice or text messages and any combination thereof.

In another embodiment user is enable to set or apply or update period of time (e.g. number of seconds - 5 seconds) 351 with one or more input icons or labels 302 for auto deleting or auto removing or auto disappearing shared or presented media including videos or photos or voice or text messages and any combination thereof to other users. In another embodiment user can dynamically update or change period of time (e.g. number of seconds - 5 seconds) 351 for auto deleting or auto removing or auto disappearing shared or presented media including videos or photos or voice or text messages and any combination after sharing of media but before viewing by receiver(s). In another embodiment user is enable to set or apply or update period of time (e.g. number of seconds - 5 seconds) 351 via voice or voice command or voice numeric command at the time of capturing of particular media and/or after sending or sharing of particular media to other users for auto deleting or auto removing or auto disappearing shared or presented media including videos or photos or voice or text messages and any combination thereof.

In another embodiment present shared media or contents in transient mode.

In another embodiment can select one or more presented input icons or labels via one-tap with slide on selected or preferred input icon(s) or label(s).

In another embodiment only sender can forward or re-send or save locally sender user's own video, audio, photo, image, voice, contents, chat session associate messages and shared resources. In one embodiment auto presenting contextual contents with captured and or shared and/or presented media including video, photo, voice & text.

In one embodiment remove shared or presented media or contents after disconnecting or leaving of all participated members with chat session on particular shared media or contents

In another embodiment user can view all shared contents and associate status including send by user, delivered to user, received by user, pending to view, view or read, deleted or removed, snapshot taken, forwarded, saved 360. In another embodiment user can apply one or more types of settings 355 for sending, sharing live broadcasting and receiving of one or more types of media or contents from one or more sources and/or contacts or connections, apply default privacy settings for sending or sharing or posting or publishing or live streaming captured or recorded media including photos, videos, voice & contents to/at/on one or more contacts and/or group(s) of contacts add/or sources and/or destinations, apply privacy settings to set particular duration or period of time (e.g. set number of seconds) for auto removing or auto disappearing of shared contents, apply settings whether chat interface is present after capturing of video or photo or user can manually select chat interface via icon 365 (Fig. 4). In another embodiment user can dynamically change position, order, arrange of or move one or more input icons or labels on display via touch, slide & drag and drop anywhere on display 110.

In another embodiment user can dynamically select functions from list of functions and can add, attach, use, apply, set with one or more input icons or labels. For example input icon or label 302 associated with function "Sent capture video or photo to user Amita" and if user wants to add more functions with said input icon or label then user can select function(s) via icon 354 and add remove or update selected function(s) and associate parameters, settings, preferences, user inputs or selection, data & metadata with one or more input icons or labels presented on display 110.

In another embodiment Fig. 4 show that user is presented with chat interface 445 and starting of chat session 445 for sharing users including sender and receiver users on shared contents or media 425 including video, audio, images, text, voice, photos and any combination thereof. User can view 425 and chat 445 with other users to whom user sent or share captured or recorded or selected media or contents 425. In another embodiment User can select or switch multiple chat sessions 423 for chatting with one or more users on one or more presented send or received shared contents of one or more users. In another embodiment user can select various options 440/460 including send text messages 450/465, share emoticons 466, send voice message(s) 467, share location infuriation 468, add said captured or shared media to user's one or more selected stories 469 (apply privacy settings that who can view user's one or more stories and set period of time or duration after that user's daily story expired (e.g. 24 hours), user can set particular duration or period of time (e.g. number of seconds) for auto removing shared contents and/or chat conversation from receiver(s) device or application or server(s) or database(s) or storage medium 471, save captured or record or previewed media locally 472, add to album 473, provide rank 474, report 475, like 476, dislike 477 to other selected or particular user's 423 shared media or contents 425, edit shared media or contents 478, forward user's own or other user's allowed contents to one or more elected users or group(s) of users 480. User can view one or more shared contents or media of one or more users via previous and next icons 430/435 along with switching of users or stories via icons 423. User can again invoke visual media capture controller via icon 410. In another embodiment change or update color of number icon(s) 423 on triggering of one or more types of status, updates, updates in location or check-in place information, updates in availability status, chat participants availability or participation of one or more users with one or more chat sessions, receiving of new messages or voice messages, receiving ratings, like or dislike or comment with shared contents or media, receiving of new captured or recorded shared media by other users, expiration of period of time associate with received shared media or contents from other users.

In another embodiment user can capture or record media including photo, video & voice and chatting for captured or recorded shared contents with receivers or participants of chat within same screen. When user selects captured and/or shared mode via icon 410 then hide the chat interface and user is presented with visual media capture controller with other options as discuss in Figure 3 and when user touch or click on icon 365 then hide visual media capture controller & associate options and show chat interface and other associate options to user as discuss in Figure 4.

In another embodiment user can play, pause, stop video, audio & voice 460. In another embodiment user can view shared as well as received shared contents in slide show mode and can play, pause, stop media including video, audio & voice 460.

For example, in another embodiment Figure 3 shows auto generated and user defined or customized labels presented to user at user's device display 110 based on user's smart phone or mobile contacts with auto associated or user defined associated function(s) e.g. "Sent Function" with each generated , created or updated & presented label name associate contact, so when user touch or single or double tap on label or label active area then photo or video is captured or recorded & stored and said captured photo or video sent or share to said label associate contact person or group(s) of contacts associate with label.

Figure 5 illustrates graphical user interface, enabling user to add, update customize, generate and create one or more input icons or labels and or one or more group(s) input icons or labels. User can input 505 group name or user can select group(s) name(s) from existing list group(s) 507 for updating said selected group(s) or user can use selected group(s) 507 as tag or template for newly created group(s) 505. User can add or update icon 509 for newly created 505 or selected group 507. User can add description 512 & metadata 515 with newly created group 505 or update description of selected group 507. User can provide rank & order to one or more group(s) 517. User can download or update pre-created or shared or selected or contextual (based on user profile & user data) groups from server 520. User can selects default one or more functions including one or more main functions 522, additional functions 523 for all created, selected, and generated input icons or labels or user can select and update or attach or apply one or more main functions 502 and/or additional functions 503 for particular specific or individual created or selected input icon(s) or label(s) 501.

In another embodiment user is provided with functions, objects, interfaces, forms, controls, web services, applications management including search, match, browse, view, select, download, subscribe, purchase (make payment), install, un-install, update, upgrade, customize, program, test, attach, detach, sort, group, filter, rank, provide ratings & comments/reviews for functions, objects, interfaces, forms, controls, web services, applications 525. User can select each function associate parameters including one or more default selected contacts and/or group(s) of contact(s) 534 for created group 505 or user can select one or more contacts/group(s) of contacts 514 for particular label 501 of created group 505 or selected group 507, user can selects sources or destinations 530/514 for particular created group 505 or selected group 507 or for particular label 501 of particular created group 505 or selected group 507 including web sites names, email addresses, instant messenger credentials, cloud storage address & like and provide associate login, authentication & authorization information, user identity, password, metadata and other settings. Users can mange contacts 534 and/or sources 530 or destinations including add new 535, update selected 537, remove selected 539, update from server(s) & source(s) 540, share 542, search, find, invite & imports 544, search who can add user into their contacts 534, order & rank 545, filter 547, sort 550, group 552 user contacts, connections, group(s) of contacts and sources & destinations including web sites, web pages, servers, networks, devices, databases, applications, services, cloud storage, email addresses, instant messenger credentials, users of social networks, folder name(s), directories, groups, interfaces, social networks, search engines, peer to peer networks. User can add new 565, update existing 657, remove 569, suggest 570, updates from server(s) & source(s) 573, share 575, order & rank 578, filter 580, sort 585, group 587 & select pre-created categories or types 557 of input icons or labels and associate input icons or labels 560. In another embodiment user can set presentation settings for dynamically presenting said selected one or more types input icons or labels of and associate input icons or labels on display 110 including dynamically present based on particular date & time, ranges of date & time, location, check-in place, activities, actions, interactions, events, transactions, user profile & user data and connected users of user' s profile & user data and any combination thereof. User can add new 501 via button 565 or update existing selected 560 via button 567 input icon(s) or label(s) and customize or define including provide name or label 501, icon, add user mediated actions including one or more forms, types of information, structured fields, forms, interface(s), application 504, user can add or update one or more selected function(s) 502, additional functions 503 and apply, set, define, customize one or more types of function(s) associate parameters including provide sources & destinations including contacts, web sites, web pages, communication channels, servers, networks, application, services, databases, devices for sharing, sending, posting, submitting, presenting, publishing, synchronizing captured, shared & recorded media or contents including video, audio, image, photo & voice 514. User can save input icon or label 506. User can add input icon or label 501/560 specific contents 591 via various option 594 indulging select, add, edit & update texts, voice or audio or sound, photo or image, video, emoticons, photo or video effects or filters, location information, selected or searched contents, files or documents and metadata 592. User can also add suggested, auto match, contextual, preference or settings specific contents from various sources 595. User can Add, update, apply, configure, customize one or more domain or subject specific forms (e.g. deals, purchase details etc.), interfaces, controls, wizards & parameters and apply privacy settings & presentation settings including present said one or more attachments or forms, interfaces, controls & wizards to user after capturing of photo or video and before executing of functions 593, so user can input or provide contextual details or take one or more user contextual actions. User can add new, update or save existing, remove, share, apply privacy settings, presentation settings & preferences, auto generate one or more created or selected group(s), us can schedule to trigger or presenting of said group including associate input icons or labels and user can apply settings for dynamically presenting said group(s) based on particular date & time, ranges of date & time, location, check-in place, activities, actions, interactions, events, transactions, user profile & user data and connected users of user's profile & user data and any combination thereof 598.

In another embodiment user can apply settings including auto add date/time, location, check -in place, profile data or profile link, system data& metadata with one or more input icons or labels.

In another embodiment user can apply settings for ask user for input icon(s) or label(s) associate updating of settings, associate functions & configurations before executing of functions.

In another embodiment user can apply settings for present only input label or present icon with labels.

In another embodiment user can apply settings for enabling voice commands indulging present input labels based on voice commands e.g. Movie - R-mall (1) Viewing movie (2) shopping at <shop name at R-mall> (3) Food <Ice-cream-brand> and like. Figure 6 illustrates various types of exemplary main functions and additional functions. User can search, match, select, sort, order, rank, filter, install, uninstall, purchase, make payment, subscribe, add, add or update with existing functions, attach, detach, remove, invoke, activate, execute, upgrade, update and present functions associate with one or more or set or categories or group of input icon and/or label.

Figure 7 illustrates one of the examples of present invention, when user clicks on "Post or Publish To/Via" option 353 then user is presented with said group related set of input labels 745 on display 110 including "All Sites", "Facebook", "Twitter", "Flicker", "Instagram",

"Whatsapp", "Pinterest-Travel", "Pinterest-Food & Drink", "Blutooth", "List of Photo/Video Seller or e-commerce Sites", "Message", "Mail", "Slideshow". User is enable to single tap to capture photo or double tap to start capturing of video or press & hold to select one or more labels and can single tap on any select sign associate labels to capture photo or double tap on any select sign associate labels to start capturing of video - explain in detail in Figure 13. For example when user single tap on label "Pinterest-Travel" 702 then photo is captured and label associate "Post To <>" function and associate parameters of web site address, login information and category enables user to post said captured photo to "Travel" category of

Figure imgf000031_0001
Figure 8 illustrates one of the examples of present invention, For example user customize or creates or generates day to day activities group of labels 376 via interface describe in Figure 5, wherein each label under group had default settings including post capture photo to family & best friends, save at user's as well as receiver's device, each label associates with particular activity name (e.g. Figure 8 shows various types of activities for throughout the day or present only particular date & time range specific activities e.g. 7:00 AM to 10 AM) and associates with date & time range, so user come to know that what type or category of photo or video I have to capture and share with decided contacts including family & best friends. For example at 7:00 AM - 7:30 AM user notifies about or highlight label "Swimming" based on pre-settings and user or other user on behalf of user double tap on "Swimming" label for intention to capturing video and stops via icon 304 as describe detail in Figure 3. System stores capture video and executes said label "Swimming" associate functions including post said capture video, audio of video and voice of said user to pre-set contacts including family & best friends. As per pre-settings after capturing, storing & sharing user is presented with chat interface as describe in Figure 9. Figure 9 illustrates that user can view user's as well as other related users 423 shared contents or media related to same time range related activity specific shared contents, so user can compare activity of user with other users at same time i.e. what user's as well as other users were did or are doing or what's plan in near future. User can send and/or receive one or more messages, voice messages, emoticons, location information, like, rating, comments, and reviews 945 from/to one or more related users 423.

In another embodiment user is dynamically updated or presented with labels as per updates in date & time, time slot specific, location specific, holiday specific, daily known activities specific, past possible activities specific, pre selected daily generalize template of possible activities specific and based on user and connected or added users' profile and user data

Figure 10 illustrates one of the examples of present invention, where user can send request(s) 1005 to pre-set default users or system default or send to selected types of other users of network(s) including contextual users, one or more contacts or group(s) of contacts, followers, following users, similar interest groups, one or user devices of network(s) based on their location, presence information, check-in place(s), availability status, updates, profile data & user data for requesting request(s) specific live or near live sharing or broadcasting of contents or media including video, audio, voice, images, photos & text. User can receive contextual requests 1002 from other users of network(s) and present to user on display 110, so user can real time notify about and view new requests 1002 in chronological or any other preferred order. User can directly one tap or double tap on particular request 1002 and starts in-real time capturing or recording and broadcasting or sharing of request specific said captured media or contents to requestor user(s). In the absence of internet or network connectivity at sender's device, captured media saved to local storage of user's device and send later to requestor(s) when internet or network connectivity available. In the absence of internet or network connectivity of requestor user's device or availability of requestor, requestor receives & view request specific captured or recorded media or contents when internet or network connectivity available and/or user's is available. For example user is wondering at Baga Beach and based on that user receives or presented with and immediately notifies other users' requests 1002 related to "baga beach" e.g. "I want to sea view of baga beach", so user can immediately one-tap or double tap on request item 1002 and stats capturing or recording of video, audio, photos, user's voice comments related to "sea view of baga beach" and shared to requestor or one or more similar request(s) specific requestor(s). Figure 11 illustrates one of the examples of present invention, where user clicks or single touches or single taps on input label "Price" 1120 for capturing photo 1130 of products displayed at particular shop or mall with intent to ask price & details to other contextual users of network for said highlighted or user marked product via free form line 1150 inside said captured & previewed or vied photo 1130 and receives price & details of said highlighted products from particular seller of said highlighted product on display 110 of user's device.

Figure 12 illustrates one of the examples of present invention, where user clicks or single touches or single taps on input label "New Products" 1220 for capturing photo 1130 of new products displayed on user' s shop with intent to notify pre-set list of customers about said newly arrive products. User can sent or attache other contents with said captured photo or video including catalogue, information about products and like.

Figure 13 illustrates one of the examples of present invention, where user can selects one or more input labels from presented set of input labels 1302, 1305 & 1309 via long press or press and hold for a while or any other types of touches and release. When user release touch then selected labels highlighted with particular icon e.g. "right sign" icon with selected input labels 1302, 1305 & 1309 or labels highlighted in bold or particular color or any other manner and enabling user to single-tap or double-tap or via any other touch types on any selected labels and enable to capture video or photo and send or share or broadcast to label associate one or more functions associate one or more users, contacts, group(s), sources, destinations and like.

In an important embodiment Figure 14 illustrates hands free or auto capturing or recording of photo or video & voice.

FIG. 14 illustrates processing operations associated with the auto visual media capture controller 138. Initially, an auto visual media capture mode is invoked 1410. For example, a user may access an application presented on display 110 to invoke a auto visual media capture mode. FIG. 3 illustrates the exterior of electronic device 100. The figure also illustrates the display 110. The electronic device 100 is in an auto visual media capture mode and presents visual media 300. The display 110 also presents a auto mode input icon 333.

In one embodiment, the identification or determination or recognition or detection of orientation of device (e.g. vertical) in combination with proximity of device (e.g. some level far from eye or body) and/or stability level of device or display image (e.g. some level device and/or image stability for particular set period of time (e.g. number of seconds)) and the amount of time user holds device in particular orientation (e.g. vertical orientation) determines whether a photograph will be recorded or a video. For example, if a user initially intends to take a photograph, then the user have to invoke auto visual media capture mode and keep device in vertical position, so sensor can identify and provides orientation of device equal to vertical orientation and in combination of proximity sensor data that device is slight far from body or eye for capturing photo and/or detection of device stability based on compass, accelerometer and gyroscope sensors, system indicates initiating of capturing of photo. So in combination of said sensor data system identifies that user wants to initiate capturing of photo. Once photo capturing initiates and if the user decides that the visual media should instead be a video, the user continues to keep device in vertical position. If the orientation of device is vertical persists for a specified period of time (e.g., 3 seconds), then the output of the visual media capture is determined to be video. The video mode may be indicated on the display 110 with an icon 304. Thus, a single gesture allows the user to seamlessly transition from a photograph mode to a video mode and therefore control the media output during the recording process. This is accomplished without entering one mode or another prior to the capture sequence. Returning to FIG. 14, sensor data identified or detected or recognized or determined 1415. For example, the sensor(s) 190 generates signals data for processing by the auto visual media capture controller 138 in conjunction with the processor 130.

Video is recorded and a timer is started in response to required sensor data 1415. The video is recorded by the processor 130 operating in conjunction with the memory 136. Alternately, a still frame is taken from the video feed and is stored as a photograph in response to identification or recognition or determination or detection of device orientation as vertical 1416 in combination with proximity is equal to slight far from body or eye and/or device/image is some determined level stable and then video is recorded. The timer is executed by the processor 130 under the control of the auto visual media capture controller 138.

Video continues to record on the display 110 and the timer continues to run in response to persistent or recognition or determination or identification or detection of orientation of device equal to vertical in combination with proximity data and/or device stability related data from one or more types of sensors. Orientation of device is equal to horizontal is subsequently identified 1417. The timer is then stopped, as is the recording of video 1425. The elapsed time recorded by the timer is then evaluated by the auto visual media capture controller 138 against a specified threshold (e.g., 3 seconds). If the threshold is exceeded (1430— Yes), then video is stored 1435. In particular, a video is sent to the video library controller 160 for handling. In one embodiment, the auto visual media capture controller 138 includes executable instructions to prompt the video library controller to execute one or more default or auto identified or selected or user applied or custom functions (e.g. send to one or more selected contacts/group(s) or publish to one or more web sites and/or enter a video preview mode) 1437. If the threshold is not exceeded (1430— No), a frame of video is selected 1445 and is stored as a photograph 1450. As indicated above, an alternate approach is to capture a still frame from the camera video feed as a photograph upon identification or determination or recognition or detection of orientation of device (e.g. vertical) in combination with proximity of device (e.g. some level far from eye or body) and/or stability level of device or display image (e.g. some level device and/or image stability for particular set period of time (e.g. number of seconds)). Such a photograph is then passed to the photographic library controller 150 for storage. The visual media capture controller 138 may then execute one or more default or auto identified or selected or user applied or custom functions (e.g. send to one or more selected contacts/group(s) or publish to one or more web sites and/or enter a photo preview mode) 1455.

The foregoing embodiment relies upon identification or determination or recognition or detection of orientation of device (Vertical or Horizontal Orientation), proximity of device (Far from body or eye / Near to body) and device/image stability (sufficient level stable or not stable) for set No. of seconds. Based upon the expired time and continues device in vertical orientation or position, either a photograph or a video is preserved. Thus, a single recording mode allows one to seamlessly transition between photograph and video recording.

In another embodiment, a photograph is taken upon identific tion or determination or recognition or detection of orientation of device (e.g. vertical) in combination with proximity of device (e.g. some level far from eye or body) and/or stability level of device or display image (e.g. some level device and/or image stability for particular set period of time (e.g. number of seconds)) and a timer is started (but video is not recorded). If persistent identification or determination or recognition or detection of orientation of device (e.g. vertical) in combination with proximity of device (e.g. some level far from eye or body) and/or stability level of device or display image (e.g. some level device and/or image stability for particular set period of time (e.g. number of seconds)) exists, as measured by the timer, for a specified period of time, then video is recorded. In this case, the user may then access both the photograph and the video. Indeed, an option to choose both a photograph and video may be supplied in accordance with different embodiments of the invention.

In another embodiment, orientation of device (e.g. vertical or horizontal) and stability of device/image may identified via Compass and/or Gyroscope and/or Accelerometer and/or Gesture sensors and any combination thereof and proximity of device may identified via Proximity Sensor. Proximity Sensor detect how close your smart phone's screen is to your body is know in art. The Compass and/or Gyroscope and/or Accelerometer and/or Gesture Sensor(s) can be used to detect the orientation and direction of the phone is known in art.

In another embodiment, user can use voice commands to instructing device to capture photo, start video, audio or voice and stop or end recording of video, preview photo or video, save photo or video, add to story photo or video, record or nor record voice or audio with photo or video, share with or send to one or more contacts, group(s) of contacts, sources & destinations, set number of seconds to auto remove the said captured and shared contents from receiver users' devices. By using or Invoking Visual Media Capture Mode 1410 user can capture or record photo, video, audio, & voice completely hands free or without touching display of mobile device and fully concentrate on capturing photo or video.

In another embodiment, removing of last changed orientation (e.g. device orientation is horizontal by user for stopping video or photo) related frames from stored video 1435.

In another embodiment, user is enabling to apply settings for enabling or disabling photo and/or video preview mode. In another embodiment, based on settings provided by sender to remove shared photo / video from receiver's device after set no. of seconds (e.g. auto remove photo after 3 seconds or video up-to length of video), auto removing photo or video after receiver' s views shared or presented media including photo or video or voice or text after said set period of time. In another embodiment, user is enabling to manually or automatically select one or more functions for said auto captured or stored photo(s) and/or video(s) and/or voice.

In another embodiment, enabling user to set delay period of time (e.g. set number of seconds) for automatically send or share or publish photos or videos.

In another embodiment, enabling user to set delay period of time (e.g. set number of seconds) for automatically execution of functions and enabling user to stop said automatic execution of function(s) meanwhile and enabling to select one or more function(s) and associate parameters (e.g. change pre-selected contacts for sending captured photo or video to them) send applies to said automatically captured photo or video or voice.

In another embodiment, enabling user to pre-set or pre-select one or more functions for applying said selected functions and automatically executing said functions after auto capturing or recording of photo or video or voice. For example user can select function "Auto Sent To" and selects associate parameters including selects one or more contacts or group(s) of contacts e.g. Friends & Family group(s) of contacts for applying them with auto captured photo or video or voice. So when user captures photo or video or voice in auto mode then said function(s) automatically executed after capturing or recording of photo or video or voice i.e. sent said auto captured photo or video or voice to said pre-selected contacts i.e. Friends & Family group(s) of contacts.

In another embodiment, Auto Visual Media Capture Mode 1410 can enabled or disabled / activate or deactivate via shaking of smart phone or device based on one or more types of sensors including Compass and/or Gyroscope and/or Accelerometer and/or Gesture sensors .

In another embodiment, dedicated device can auto open display of camera and associate auto Invoke Visual Media Capture Mode or auto visual media capture controller application 1410. In another embodiment, an electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a sensor(s) including gyroscope, accelerometer, proximity sensor, compass, audio & gesture sensors to identify device/image orientation, proximity & stability on the display; and an auto visual media capture controller to alternately record the visual media as a photograph or a video based upon change in orientation and/or change proximity and/or stability of device or image within particular period of time or based upon evolution of the time period between the change in orientation and/or change proximity and/or stability of device or image.

In another embodiment, the visual media capture controller receive identified, monitored, tracked, recorded, recognized, detected, determined & sensed device and/or image orientation, stability & proximity data from sensors including gyroscope, accelerometer, proximity sensor, compass, audio & gesture sensors.

In another embodiment, the auto visual media capture controller selectively stores the photograph in a photograph library.

In another embodiment, the auto visual media capture controller invokes a photograph preview mode. In another embodiment, the auto visual media capture controller selectively stores the video in a video library.

In another embodiment, the auto visual media capture controller invokes a video preview mode. In another embodiment, the auto visual media capture controller selects a frame or image of the video to form or extract the photograph.

In another embodiment, the auto visual media capture controller stores the photograph upon change in orientation of device/image.

In another embodiment, the auto visual media capture controller includes instructions executed by a processor.

In another embodiment, further comprising a wireless signal processor, a power control circuit and a global positioning processor connected to the processor.

An embodiment of the present invention relates to a computer storage product with a computer readable storage medium having computer code thereon for performing various computer- implemented operations. The media and computer code may be those specially designed and constructed for the purposes of the present invention, or they may be of the kind well known and available to those having skill in the computer software arts. Examples of computer-readable media include, but are not limited to: magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROMs, DVDs and holographic devices; magneto- optical media; and hardware devices that are specially configured to store and execute program code, such as application-specific integrated circuits ("ASICs"), programmable logic devices ("PLDs") and ROM and RAM devices. Examples of computer code include machine code, such as produced by a compiler, and files containing higher-level code that are executed by a computer using an interpreter. For example, an embodiment of the invention may be implemented using JAVA.RTM., C++, or other object-oriented programming language and development tools. Another embodiment of the invention may be implemented in hardwired circuitry in place of, or in combination with, machine-executable software instructions.

The foregoing description, for purposes of explanation, used specific nomenclature to provide a thorough understanding of the invention. However, it will be apparent to one skilled in the art that specific details are not required in order to practice the invention. Thus, the foregoing descriptions of specific embodiments of the invention are presented for purposes of illustration and description. They are not intended to be exhaustive or to limit the invention to the precise forms disclosed; obviously, many modifications and variations are possible in view of the above teachings. The embodiments were chosen and described in order to best explain the principles of the invention and its practical applications, they thereby enable others skilled in the art to best utilize the invention and various embodiments with various modifications as are suited to the particular use contemplated. It is intended that the following claims and their equivalents define the scope of the invention. In an embodiment a program is written as a series of human understandable computer instructions that can be read by a compiler and linker, and translated into machine code so that a computer can understand and run it. A program is a list of instructions written in a programming language that is used to control the behavior of a machine, often a computer (in this case it is known as a computer program). A programming language's surface form is known as its syntax. Most programming languages are purely textual; they use sequences of text including words, numbers, and punctuation, much like written natural languages. On the other hand, there are some programming languages which are more graphical in nature, using visual relationships between symbols to specify a program. In computer science, the syntax of a computer language is the set of rules that defines the combinations of symbols that are considered to be a correctly structured document or fragment in that language. This applies both to programming languages, where the document represents source code, and markup languages, where the document represents data. The syntax of a language defines its surface form. Text-based computer languages are based on sequences of characters, while visual programming languages are based on the spatial layout and connections between symbols (which may be textual or graphical or flowchart(s)). Documents that are syntactically invalid are said to have a syntax error. Syntax - the form - is contrasted with semantics - the meaning. In processing computer languages, semantic processing generally comes after syntactic processing, but in some cases semantic processing is necessary for complete syntactic analysis, and these are done together or concurrently. In a compiler, the syntactic analysis comprises the frontend, while semantic analysis comprises the backend (and middle end, if this phase is distinguished). There are millions of possible combinations, sequences, ordering, permutations & formations of inputs, interpretations, and outputs or outcomes of set of instructions of standardized or specialized or generalized or structured or functional or object oriented programming language(s). The present invention has been described in particular detail with respect to a limited number of embodiments. Those of skill in the art will appreciate that the invention may additionally be practiced in other embodiments. First, the particular naming of the components, capitalization of terms, the attributes, data structures, or any other programming or structural aspect is not mandatory or significant, and the mechanisms that implement the invention or its features may have different names, formats, or protocols. Furthermore, the system may be implemented via a combination of hardware and software, as described, or entirely in hardware elements. Also, the particular division of functionality between the various system components described herein is merely exemplary, and not mandatory; functions performed by a single system component may instead be performed by multiple components, and functions performed by multiple components may instead performed by a single component. Additionally, although the foregoing embodiments have been described in the context of a social network website, it will apparent to one of ordinary skill in the art that the invention may be used with any social network service, even if it is not provided through a website. Any system that provides social networking functionality can be used in accordance with the present invention even if it relies, for example, on e-mail, instant messaging or any other form of peer-to-peer communications, or any other technique for communicating between users. Systems used to provide social networking functionality include a distributed computing system, client-side code modules or plug-ins, client-server architecture, a peer-to peer communication system or other systems. The invention is thus not limited to any particular type of communication system, network, protocol, format or application. The foregoing description of the embodiments of the invention has been presented for the purpose of illustration; it is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Persons skilled in the relevant art can appreciate that many modifications and variations are possible in light of the above disclosure.

Some portions of this description describe the embodiments of the invention in terms of algorithms and symbolic representations of operations on information. These algorithmic descriptions and representations are commonly used by those skilled in the data processing arts to convey the substance of their work effectively to others skilled in the art. These operations, while described functionally, computationally, or logically, are understood to be implemented by computer programs or equivalent electrical circuits, microcode, or the like. Furthermore, it has also proven convenient at times, to refer to these arrangements of operations as modules, without loss of generality. The described operations and their associated modules may be embodied in software, firmware, hardware, or any combinations thereof.

Any of the steps, operations, or processes described herein may be performed or implemented with one or more hardware or software modules, alone or in combination with other devices. In one embodiment, a software module is implemented with a computer program product comprising a computer-readable medium containing computer program code, which can be executed by a computer processor for performing any or all of the steps, operations, or processes described.

Embodiments of the invention may also relate to an apparatus for performing the operations herein. This apparatus may be specially constructed for the required purposes, and/or it may comprise a general-purpose computing device selectively activated or reconfigured by a computer program stored in the computer. Such a computer program may be stored in a tangible computer readable storage medium or any type of media suitable for storing electronic instructions, and coupled to a computer system bus. Furthermore, any computing systems referred to in the specification may include a single processor or may be architectures employing multiple processor designs for increased computing capability.

Embodiments of the invention may also relate to a computer data signal embodied in a carrier wave, where the computer data signal includes any embodiment of a computer program product or other data combination described herein. The computer data signal is a product that is presented in a tangible medium or carrier wave and modulated or otherwise encoded in the carrier wave, which is tangible, and transmitted according to any suitable transmission method.

Finally, the language used in the specification has been principally selected for readability and instructional purposes, and it may not have been selected to delineate or circumscribe the inventive subject matter. It is therefore intended that the scope of the invention be limited not by this detailed description, but rather by any claims that issue on an application based here on. Accordingly, the disclosure of the embodiments of the invention is intended to be illustrative, but not limiting, of the scope of the invention, which is set forth in the following claims.

Claims

An electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a touch controller to identify haptic contact engagement, haptic contact persistence and haptic contact release on the display; present the one or more input labels or icons on display; and based on haptic contact engagement and haptic contact release (single tap or double tap) on said particular presented label, enabling a visual media capture controller to record the visual media as a photograph or video or voice or start recording of a video and triggering or invoking or presenting or executing one or more associate (auto identified, set, default, user selected) functions, events, transactions, processes, formatting, web services, interfaces, instructions, triggers, scripts, routines, object, modules, actions including sending (sharing, advertising, broadcasting, publishing & synchronizing) of recorded photograph or a video.
The electronic device of claim 1 wherein enabling user to select or operate visual media capture controller(s) based upon haptic contact engagement (single tap, double tap) or via digital or physical keyboard or via voice command(s).
The electronic device of claim 1 wherein enabling user to search, match, select, make payment, purchase, download, register, upload, install, view details, un-install, sort, order, group, filter, rank, update, upgrade one or more functions from one or more developers, providers and sources.
The electronic device of claim 1 wherein enabling user to set, make as default, apply, attach, associate & select one or more functions with one or more or group or set of input icons or labels.
The electronic device of claim 1 wherein automatically select one or more functions with one or more input icons or labels.
The electronic device of claim 1 wherein select one or more functions with one or more input icons or labels based on user defined or user programmed script(s) including execute, invoke, triggers, orchestrate functions, interfaces based on rules, commands, Boolean operators, programming languages.
7. The electronic device of claim 1 wherein automatically and/or in background mode or silent mode execute, invoke, triggers, orchestrate function(s), interface(s) while capturing photo or video/recording of video or voice or run-time of capturing of video/recording of voice.
8. The electronic device of claim 1 wherein enabling user to further selecting or updating already selected or associated one or more functions while capturing of photo or video and/or recording of video or voice.
9. The electronic device of claim 1 wherein automatically and/or in background mode or silent mode execute, invoke, triggers, orchestrate function(s), interface(s) while capturing video/recording of voice or run-time of capturing of video/recording of voice and present contextual contents including location information, information about associate or recognized or detected or sensed nodes, named entities, people, brands, products, items, objects, details associate with one or more or each or run-time or real-time capturing of video related photo frames.
10. The electronic device of claim 1 wherein instructing user to input, select structured or unstructured details including form(s), interface(s) before executing of function(s) associate with input icon(s) or label(s).
11. The electronic device of claim 1 wherein presenting controls, user actions, interface(s) before executing of function(s) associate with input icon(s) or label(s).
12. The electronic device of claim 1 wherein the visual media capture controller presents a single mode input labels or icons or images on the display to receive the haptic contact engagement, haptic contact persistence and haptic content release for capturing or recording photo or video or voice as well as executing, invoking or triggering of associate one or more functions.
13. The electronic device of claim 1 wherein the visual media capture controller selectively stores the photograph in a photograph library.
14. The electronic device of claim 1 wherein the visual media capture controller selectively stores the video in a video library.
15. The electronic device of claim 1 wherein end recording of started video and execute input icon(s) or label(s) associate function(s) after further receiving the haptic contact engagement by visual media capture controller. 16. The electronic device of claim 1 wherein end recording of started video and execute input icon(s) or label(s) associate function(s) after set particular duration including set number of seconds or minutes.
17. The method of claim 1, wherein the recognized touch input is identified as a short touch, a long touch, a multi-position touch, a touch & drag, tap, single tap, double tap, press & hold while, long press, two finger tap, pinch open and one or more types of touch or multi-touch gestures.
18. The method of claim 1, wherein: the general image capture is performed when the identified type of touch input is the single tap or short touch; and the continuous image capture is performed when the identified type of touch input is the double tap.
19. The electronic device of claim 1 wherein the visual media capture controller includes instructions executed by a processor.
20. The electronic device of claim 9 further comprising a wireless signal processor, a power control circuit, one or more types of sensors including Accelerometer, Gyroscope, Gesture Sensor, Proximity Sensor, Direction sensor & Audio sensor and a global positioning processor connected to the processor.
21. An electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a touch controller to identify haptic contact engagement, haptic contact persistence and haptic contact release on the display; present the labels or icons or images or digital representation of contacts/connections and group(s) of contacts/connections on display as a input icons or labels for recording or capturing photograph or starting recording or capturing of a video, wherein a visual media capture controller to record the visual media as a photograph or start recording of a video and triggering of sending (sharing, advertising, broadcasting, publishing & synchronizing) of recorded photograph or a video.
22. The electronic device of claim 21 wherein end recording of started video and send a recorded video after to selected label or icon or image or digital representation associate contact(s)/connection(s)/group(s) of contacts/connections further touching single visual media capture controller.
23. The electronic device of claim 21 wherein end recording of started video and send a recorded video to selected label or icon or image or digital representation associate contact(s)/connection(s)/group(s) of contacts/connections after set particular duration including set number of seconds or minutes.
24. The electronic device of claim 21 wherein the visual media capture controller presents a single mode input labels or icons or images of user's contacts/group(s) of contacts on the display to receive the haptic contact engagement, haptic contact persistence and haptic content release and triggering of sending or sharing of recorded photograph or video to label or icon or image or digital representation associate contact(s)/connection(s)/group(s) of contacts/connections .
25. The electronic device of claim 21 wherein the visual media capture controller selectively stores the photograph in a photograph library.
26. The electronic device of claim 25 wherein based on settings the visual media capture controller invokes a photograph preview mode.
27. The electronic device of claim 26 wherein the visual media capture controller invokes a photograph preview mode and present media including photograph or video associate chat session interface.
28. The electronic device of claim 27 wherein the chat interface enables sender user and receiving users to post messages & voice messages, view posted messages & listen voice messages, auto matched contextual contents based on posted messages, voice messages and participant users' data including profile & user data, view associate statistics including send to users, received users, viewed users, pending to view users, participated users in current media including photo/ video related chat session, likes by chat session participated users.
29. The electronic device of claim 26 wherein the visual media capture controller invokes a photograph preview mode and present indications of receiving of other users' shared media including photos & videos via number icon.
30. The electronic device of claim 29 wherein enable to switch among presented number icons for participate in shared media including photos & video associate chat session(s) for sending and receiving messages, voice messages, sound, edited media including photo or video, likes, location information, presence information, status, emoticons.
31. The electronic device of claim 29 wherein number icon including sender name or identity or image, profile link, availability status, presence information.
32. The electronic device of claim 29 wherein remove number icon after set number of time period.
33. The electronic device of claim 29 wherein remove number icon after expiring of chat session based on triggering of no single character inputted or selected or send by all participated users within set number of time period.
34. The electronic device of claim 29 wherein add or update number icon after receiving shared media from other users or further capturing & sharing or sending of shared media to other users.
35. The electronic device of claim 25 wherein based on settings the visual media capture controller invokes a visual media capture mode.
36. The electronic device of claim 21 wherein the visual media capture controller selectively stores the video in a video library. 37. The electronic device of claim 36 wherein based on settings the visual media capture controller invokes a video preview mode.
38. The electronic device of claim 36 wherein based on settings the visual media capture controller invokes a visual media capture mode.
39. The electronic device of claim 21 wherein enable shared media receiving users to view shared media and associate chat conversations or messages or voice messages after expiration of chat session of other users.
40. The electronic device of claim 29 wherein force remove received shared media of other users after set number of time period. 41. The electronic device of claim 21 wherein enabling user to scrolling presented number icons.
42. The electronic device of claim 21 wherein enabling user to paging presented contacts and group(s) of contacts on display. 43. The electronic device of claim 21 wherein notify receiving user(s) about shared media including photos, videos, text, chat messages, comments, likes, voice messages, sounds, emoticons, locations, sending & receiving date & time, expiration or removal duration or time period of shared media and information about sender(s) including profile link, photo or image, name & identity, availability status, user posted status, presence information and information about shared media associate chat session related participant users including profile link, photo or image, name & identity, availability status, user posted status, presence information,
44. The electronic device of claim 43 wherein enabling user to opening application from notification and presenting each active shared media of each sending user with chat interface to each receiving user(s).
45. The electronic device of claim 21 wherein the visual media capture controller stores the photograph upon haptic contact engagement.
46. The electronic device of claim 21 wherein the visual media capture controller includes instructions executed by a processor.
47. The electronic device of claim 46 further comprising a wireless signal processor, a power control circuit and a global positioning processor connected to the processor.
48. The electronic device of claim 21 wherein presenting labels or icons or images or digital representation of contact(s)/connection(s)/group(s) of contacts/connections in switchable show and hide mode.
49. The electronic device of claim 21 wherein switchable shows and hide mode based upon set time period.
50. The electronic device of claim 21 wherein enabling user to search, match, select, remove or hide, add or attach or show, sort, rank, order or arrange, filter, view associate profile, details & statistics, group, page- wise group contacts or connections from display.
51. The electronic device of claim 21 wherein enabling user to set particular period of time for one or more contacts or group(s) of contacts for removing shared photo or video from server and/or receiver's or receivers' device or application or database. 52. The electronic device of claim 1 wherein dynamically change colors of label based on background color of photo or video.
53. A non-transient computer readable storage medium, comprising executable instructions to: process haptic contact signals from a label or icon or image or digital representation of contacts/connections presented on display; record a photograph and send a photograph to label or icon or image or digital representation associate contacts/connections based upon a first haptic contact signal; and start recording of a video based upon a second haptic contact signal, wherein the second haptic contact signal is a haptic contact release signal that occurs after a specified period of time after the first haptic contact signal.
54. The non-transient computer readable storage medium of claim 54 wherein end recording of started video and send a recorded video to selected label or icon or image or digital representation associate contact(s)/connection(s)/group(s) of contacts/connections based on further receiving and processing haptic contact signals from visual media capture controller.
55. The non-transient computer readable storage medium of claim 54 wherein end recording of started video and send a recorded video to selected label or icon or image or digital representation associate contact(s)/connection(s)/group(s) of contacts/connections after set particular duration including set number of seconds or minutes.
56. The non-transient computer readable storage medium of claim 54 wherein the first haptic contact signal is a haptic contact engagement signal.
57. The non-transient computer readable storage medium of claim 54 wherein the first haptic contact signal and the second haptic contact signal occur at single mode input icon.
58. The non-transient computer readable storage medium of claim 54 wherein the first haptic contact signal is a first specified gesture performed on the display.
59. The non-transient computer readable storage medium of claim 54 wherein the second haptic contact signal is a second specified gesture performed on the display.
60. The non-transient computer readable storage medium of claim 54 further comprising executable instructions to store the photograph in a photograph library. 61. The non-transient computer readable storage medium of claim 60 further comprising executable instructions to invoke a photograph preview mode.
62. The non-transient computer readable storage medium of claim 54 further comprising executable instructions to store the video in a video library.
63. The non-transient computer readable storage medium of claim 62 further comprising executable instructions to invoke a video preview mode.
64. An electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a touch controller to identify haptic contact engagement, haptic contact persistence and haptic contact release on the display; a presenting one or more types of input icon or label on display; enabling user to select one or more types of input icon or label based upon haptic contact engagement including single touch, double touch, one-tap, double-tap & one or more types of touches or multi touches gestures on respective icon or label of particular type of input icon or label; enabling user to select preferred input icon or label from one or more or set of input icons or labels related to particular type of input icon or label for recording the visual or voice media as a photograph or a video or a voice as well as executing or triggering or invoking associate one or more functions or commands or instructions or routines.
65. The electronic device of claim 64 wherein enabling user to select one or more types of input icons or labels for using or accessing one or more types of functions.
66. The electronic device of claim 64 wherein enabling user to select one or more input icons or labels of one or more types of input icons or labels based upon haptic contact engagement on preferred input icon or label.
67. The electronic device of claim 64 wherein one or more types of input icons or labels including send captured photo or video or voice to particular user or contact or group of contacts and user is enable to select said "Send To" type of input icon or label, wherein presenting said "Send To" type of input icon or label to capture photo or video and sent to particular user or contact or group within single touch or haptic contact engagement.
68. The electronic device of claim 64 wherein executing or triggering or invoking associate one or more functions or commands or instructions or routines during or at the end of capturing or recording of visual or voice media.
69. The electronic device of claim 64 wherein enabling user to search, match, select, make payment, purchase, download, register, upload, install, view details, un-install, sort, order, group, filter, rank, update, upgrade one or more functions from one or more developers, providers and sources.
70. The electronic device of claim 64 wherein enabling user to set, make as default, apply, attach, associate, select one or more functions with one or more or group or set of input icons or labels .
71. The electronic device of claim 64 wherein automatically select one or more functions with one or more input icons or labels.
72. The electronic device of claim 64 wherein select one or more functions with one or more input icons or labels based on user defined or user programmed script(s) including execute, invoke, triggers, orchestrate functions, interfaces based on rules, commands, Boolean operators, programming languages.
73. The electronic device of claim 64 wherein automatically and/or in background mode or silent mode execute, invoke, triggers, orchestrate function(s), interface(s) while capturing photo or video/recording of video or voice or run-time of capturing of video/recording of voice.
74. The electronic device of claim 64 wherein enabling user to further selecting or updating already selected or associate one or more functions while capturing of photo or video and/or recording of video or voice.
75. The electronic device of claim 64 wherein automatically and/or in background mode or silent mode execute, invoke, triggers, orchestrate function(s), interface(s) while capturing video/recording of voice or run-time of capturing of video/recording of voice and present contextual contents including location information, information about associate or recognized or detected or sensed nodes, named entities, people, brands, products, items, objects, details associate with one or more or each or run-time or real-time capturing of video related photo frames.
76. The electronic device of claim 64 wherein instruct user to input, select structured or unstructured details including form(s), interface(s) before executing of function(s) associate with input icon or label(s).
77. The electronic device of claim 64 wherein present controls, user actions, interface(s) before executing of function(s) associate with input icon or label(s).
78. The electronic device of claim 64 wherein the visual media capture controller presents a single mode input labels or icons or images on the display to receive the haptic contact engagement, haptic contact persistence and haptic content release for capturing or recording photo or video or voice as well as executing, invoking or triggering of associate one or more functions.
79. The electronic device of claim 64 wherein the visual media capture controller selectively stores the photograph in a photograph library.
80. The electronic device of claim 64 wherein the visual media capture controller selectively stores the video in a video library.
81. The electronic device of claim 64 wherein end recording of started video and execute input icon(s) or label(s) associate function(s) after further receiving the haptic contact engagement by visual media capture controller.
82. The electronic device of claim 64 wherein end recording of started video and execute input icon(s) or label(s) associate function(s) after set particular duration including set number of seconds or minutes.
83. The electronic device of claim 64 wherein the visual media capture controller includes instructions executed by a processor.
84. The electronic device of claim 83 further comprising a wireless signal processor, a power control circuit and a global positioning processor connected to the processor. 85. An electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a touch controller to identify haptic contact engagement, haptic contact persistence and haptic contact release on the display; and enabling to create or generate or auto generate one or more input icons or labels or one or more types of input icons or labels based on attaching, associating, selecting, setting, applying, customizing, defining, updating & programming one or more selected or auto matched or customize function(s) or group(s) of functions or types of functions, features, user actions, objects, applications, instructions, scripts, forms, controls, widgets, one or more types of contents & database or data, interfaces, web services with one or more input icons or labels or one or more types of input icons or labels .
86. The electronic device of claim 85 wherein enabling to create or generate or auto generate input icons or labels based on user selected one or more function(s) or customized functions including selection of parameters, one or more contacts, group of contacts, web sites, sources, destinations, emails & Instant Messenger addresses & associate user identity, password or authentication information, one or more types of contents including templates, tags, suggested contents, keywords, questions, answers, captions, title, categories, images, photos, voice, video, audio, text, links, database or data, structured data, files or documents, emoticons, user profile, user data & commands, interfaces, applications, objects, services, widgets, controls, functions, features, keywords, categories, types, key phrases, name of folders, nodes, named entities and any combination thereof.
87. The electronic device of claim 85 wherein enabling user to search, match, select, make payment, purchase, download, register, upload, install, view details, un-install, sort, order, group, filter, rank, update, upgrade one or more functions from one or more developers, providers and sources.
88. The electronic device of claim 85 wherein select one or more functions with one or more input icons or labels based on user defined or user programmed script(s) including execute, invoke, triggers, orchestrate functions, objects, modules, instructions, triggers, applications, interfaces, parameters based on rules, commands, Boolean operators, one or more types or programming language syntax, scripts.
89. The electronic device of claim 85 wherein automatically and/or in background mode or silent mode execute, invoke, triggers, orchestrate function(s), interface(s) while capturing photo or video/recording of video or voice or run-time of capturing of video/recording of voice.
90. The electronic device of claim 85 wherein instructing user to input, select structured or unstructured details including form(s), interface(s) before executing of function(s) associate with input icon(s) or label(s). 91. The electronic device of claim 85 wherein presenting controls, user actions, interface(s) before executing of function(s) associate with input icon(s) or label(s).
92. The electronic device of claim 85 wherein the visual media capture controller presents a single mode input labels or icons or images on the display to receive the haptic contact engagement including one touch, two or more or multi touch and/or haptic contact persistence and/or haptic content release for capturing or recording photo or video or voice as well as executing, invoking or triggering of associate one or more functions.
93. An electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a touch controller to identify haptic contact engagement, haptic contact persistence and haptic contact release on the display; enabling to create or generate or auto generate one or more input icons or labels or one or more types of input icons or labels based on attaching, associating, selecting, setting, applying, customizing, defining, updating & programming one or more selected or auto matched or customize function(s) or group(s) of functions or types of functions, features, user actions, objects, applications, instructions, scripts, forms, controls, widgets, one or more types of contents & database or data, interfaces, web services with one or more input icons or labels or one or more types of input icons or labels; a presenting one or more input icons or labels or one or more types of input icons or labels ; and enabling to select or select via one- touch, two touch or multi-touch or based upon haptic contact engagement one or more visual media capture controller(s).
94. The electronic device of claim 93 wherein enabling user to select one or more types of input icons or labels and related one or more input icons or labels via one-touch, two touch or multi-touch or based upon haptic contact engagement for recording media/photo/video/voice/content as well as executing, invoking, running, presenting, connecting, downloading, installing, opening, triggering one or more functions, objects, applications, web pages, widgets, web services, interfaces, wizards, programming scripts, instructions or source code, one or more types of contents including text, images, photos, audio, voice, video, files or documents, database, links, emoticons, location information, status & presence information, user profile, user data, user connections or contacts.
95. The electronic device of claim 93 wherein dynamically present one or more input icons or labels or one or more types of input icons or labels based on updates in date, time, date & time ranges, updates, status, presence information, location, holidays or vacations, planned activates, actions, events, status, updates, & transactions, check-in place, user & connected users' profile & user data
96. An electronic device, comprising: digital image sensors to capture visual media; a display to present the visual media from the digital image sensors; a sensor(s) including gyroscope, accelerometer, proximity sensor, compass, audio & gesture sensors to identify device/image orientation, proximity & stability on the display; and an auto visual media capture controller to alternately record the visual media as a photograph or a video based upon change in orientation and/or change proximity and/or stability of device or image within particular period of time or based upon evolution of the time period between the change in orientation and/or change proximity and/or stability of device or image.
97. The electronic device of claim 96 wherein the visual media capture controller receive identified, monitored, tracked, recorded, recognized, detected, determined & sensed device and/or image orientation, stability & proximity data from sensors including gyroscope, accelerometer, proximity sensor, compass, audio & gesture sensors.
98. The electronic device of claim 96 wherein the auto visual media capture controller selectively stores the photograph in a photograph library. 99. The electronic device of claim 98 wherein the auto visual media capture controller invokes a photograph preview mode.
100. The electronic device of claim 96 wherein the auto visual media capture controller selectively stores the video in a video library.
101. The electronic device of claim 100 wherein the auto visual media capture controller invokes a video preview mode.
102. The electronic device of claim 96 wherein the auto visual media capture controller selects a frame or image of the video to form or extract the photograph.
103. The electronic device of claim 96 wherein the auto visual media capture controller stores the photograph upon change in orientation of device/image. 104. The electronic device of claim 96 wherein the auto visual media capture controller includes instructions executed by a processor.
105. The electronic device of claim 104 further comprising a wireless signal processor, a power control circuit and a global positioning processor connected to the processor.
PCT/IB2014/059369 2014-03-02 2014-03-02 Single mode & automatically visual media capture and excuting associate functions(s) WO2015132622A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/IB2014/059369 WO2015132622A1 (en) 2014-03-02 2014-03-02 Single mode & automatically visual media capture and excuting associate functions(s)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
PCT/IB2014/059369 WO2015132622A1 (en) 2014-03-02 2014-03-02 Single mode & automatically visual media capture and excuting associate functions(s)
US14995137 US20160132231A1 (en) 2014-03-02 2016-01-13 Communications devices and methods for single-mode and automatic media capture
US15079836 US20160202901A1 (en) 2014-03-02 2016-03-24 Communications devices and methods for single-mode and automatic media capture
US15426008 US20170150037A1 (en) 2014-03-02 2017-02-06 Communications devices and methods for single-mode and automatic media capture

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14995137 Continuation US20160132231A1 (en) 2014-03-02 2016-01-13 Communications devices and methods for single-mode and automatic media capture

Publications (1)

Publication Number Publication Date
WO2015132622A1 true true WO2015132622A1 (en) 2015-09-11

Family

ID=50842275

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2014/059369 WO2015132622A1 (en) 2014-03-02 2014-03-02 Single mode & automatically visual media capture and excuting associate functions(s)

Country Status (2)

Country Link
US (3) US20160132231A1 (en)
WO (1) WO2015132622A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3159773A3 (en) * 2015-10-22 2017-07-26 LG Electronics Inc. Mobile device and method of controlling therefor

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8655404B1 (en) * 2009-11-11 2014-02-18 Gurminder Singh Systems and methods for distributing content with a portable electronic device
US10127273B2 (en) 2014-04-15 2018-11-13 Splunk Inc. Distributed processing of network data using remote capture agents
US20160057087A1 (en) * 2014-08-21 2016-02-25 Facebook, Inc. Processing media messages based on the capabilities of the receiving device
US20180025451A1 (en) * 2014-10-15 2018-01-25 UScope Technologies, Inc. Systems and Methods for Improved Property Inspection Management
CN105592261A (en) * 2014-11-04 2016-05-18 深圳富泰宏精密工业有限公司 Auxiliary shooting system and method
US20160164809A1 (en) * 2014-12-05 2016-06-09 Facebook, Inc. Identifying and selecting contacts to include in a face tray of a messaging application
US20160165128A1 (en) * 2014-12-05 2016-06-09 Facebook, Inc. Capturing and sending images and videos based on a single user interaction with a user interface element
US10001911B2 (en) * 2015-04-10 2018-06-19 International Business Machines Corporation Establishing a communication link between plural participants based on preferences
US9774693B2 (en) * 2015-04-29 2017-09-26 Facebook, Inc. Methods and systems for viewing user feedback
US20160373388A1 (en) * 2015-06-19 2016-12-22 Voxer Ip Llc Messaging application for recording and inserting a video message into a chat

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8428453B1 (en) * 2012-08-08 2013-04-23 Snapchat, Inc. Single mode visual media capture

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060252442A1 (en) * 2005-05-04 2006-11-09 Nokia Corporation Method for establishing a PoC connection in a terminal device with a touch-screen display, an application used in the method and a terminal device
KR101722616B1 (en) * 2009-12-24 2017-04-19 삼성전자주식회사 Method and apparatus for operating application of a touch device having touch-based input interface
US9237305B2 (en) * 2010-10-18 2016-01-12 Apple Inc. Overlay for a video conferencing application
US20120290968A1 (en) * 2011-05-13 2012-11-15 Christopher Michael Cecora Web Link Quick Send Add-On For Internet Browsers
US20130041948A1 (en) * 2011-08-12 2013-02-14 Erick Tseng Zero-Click Photo Upload
US8887096B2 (en) * 2011-10-27 2014-11-11 Disney Enterprises, Inc. Friends lists with dynamic ordering and dynamic avatar appearance
US9179021B2 (en) * 2012-04-25 2015-11-03 Microsoft Technology Licensing, Llc Proximity and connection based photo sharing
US20160050289A1 (en) * 2012-06-26 2016-02-18 Google Inc. Automatic sharing of digital content
US20150206349A1 (en) * 2012-08-22 2015-07-23 Goldrun Corporation Augmented reality virtual content platform apparatuses, methods and systems
US20140055633A1 (en) * 2012-08-27 2014-02-27 Richard E. MARLIN Device and method for photo and video capture
US8775972B2 (en) * 2012-11-08 2014-07-08 Snapchat, Inc. Apparatus and method for single action control of social network profile access
US20140282140A1 (en) * 2013-03-13 2014-09-18 Outright, Inc. Method for mobile content management
US20140281994A1 (en) * 2013-03-15 2014-09-18 Xiaomi Inc. Interactive method, terminal device and system for communicating multimedia information
US20150205492A1 (en) * 2014-01-20 2015-07-23 John B. Nobil Navigating audio content and selecting portions thereof using circular dial on user interface

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8428453B1 (en) * 2012-08-08 2013-04-23 Snapchat, Inc. Single mode visual media capture

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
APPLE: "iPhone 3G user manual", 11 July 2008 (2008-07-11), pages 1 - 154, XP007907619, Retrieved from the Internet <URL:http://manuals.info.apple.com/en_US/iPhone_User_Guide.pdf> *
None
SAMSUNG: "Samsung GALAXY Note 10.1 - Android Tablet - User Manual", SAMSUNG'S USER MANUALS, 21 August 2012 (2012-08-21), pages 1 - 249, XP055117806, Retrieved from the Internet <URL:http://downloadcenter.samsung.com/content/UM/201208/20120821042319330/GT-N8013_English_User_Manual_LGB_F3.pdf> [retrieved on 20140513] *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3159773A3 (en) * 2015-10-22 2017-07-26 LG Electronics Inc. Mobile device and method of controlling therefor
US10013227B2 (en) 2015-10-22 2018-07-03 Lg Electronics Inc. Mobile device and method of controlling therefor

Also Published As

Publication number Publication date Type
US20170150037A1 (en) 2017-05-25 application
US20160132231A1 (en) 2016-05-12 application
US20160202901A1 (en) 2016-07-14 application

Similar Documents

Publication Publication Date Title
US7954058B2 (en) Sharing of content and hop distance over a social network
US20080005664A1 (en) Method and Apparatus for Creating, Configuring and Displaying a Widget to Display User-Generated Highlights
US20110125847A1 (en) Collaboration networks based on user interactions with media archives
US20080222687A1 (en) Device, system, and method of electronic communication utilizing audiovisual clips
US20110202864A1 (en) Apparatus and methods of receiving and acting on user-entered information
US20060085515A1 (en) Advanced text analysis and supplemental content processing in an instant messaging environment
US20110041076A1 (en) Platform for delivery of heavy content to a user
US20100299391A1 (en) Configuring channels for sharing media
US20080016164A1 (en) Method and Apparatus for Automatically Embedding and emailing User-Generated Highlights
US20110021250A1 (en) Aggregated, interactive communication timeline
US20110167357A1 (en) Scenario-Based Content Organization and Retrieval
US20110131299A1 (en) Networked multimedia environment allowing asynchronous issue tracking and collaboration using mobile devices
US20110246502A1 (en) Creating and propagating annotated information
US20120044153A1 (en) Method and apparatus for browsing content files
US8599801B2 (en) Collecting implicit information for determining context of event actions
US20140067702A1 (en) Method and System for Dynamically Publishing, Sharing, Communication and Subscribing
US20110167078A1 (en) User Interfaces for Content Categorization and Retrieval
US20140136990A1 (en) System for and method of embedding rich media into text messages
US8650210B1 (en) Identifying non-search actions based on a search query
US20120110464A1 (en) Content sharing interface for sharing content in social networks
US20160162172A1 (en) Presenting plurality types of interfaces and functions for conducting various activities
US20110087739A1 (en) Routing User Data Entries to Applications
WO2013089674A1 (en) Real-time mapping and navigation of multiple media types through a metadata-based infrastructure
US8996625B1 (en) Aggregate display of messages
US20140181197A1 (en) Tagging Posts Within A Media Stream

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14727231

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase in:

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14727231

Country of ref document: EP

Kind code of ref document: A1