US20190109810A1 - Social-topical adaptive networking (stan) system allowing for group based contextual transaction offers and acceptances and hot topic watchdogging - Google Patents

Social-topical adaptive networking (stan) system allowing for group based contextual transaction offers and acceptances and hot topic watchdogging Download PDF

Info

Publication number
US20190109810A1
US20190109810A1 US16/196,542 US201816196542A US2019109810A1 US 20190109810 A1 US20190109810 A1 US 20190109810A1 US 201816196542 A US201816196542 A US 201816196542A US 2019109810 A1 US2019109810 A1 US 2019109810A1
Authority
US
United States
Prior art keywords
user
topic
space
system
nodes
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US16/196,542
Inventor
Jeffrey Alan Rapaport
Seymour Rapaport
Kenneth Allen Smith
James Beattie
Gideon Gimlan
Original Assignee
Jeffrey Alan Rapaport
Seymour Rapaport
Kenneth Allen Smith
James Beattie
Gideon Gimlan
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US201161485409P priority Critical
Priority to US201161551338P priority
Priority to US13/367,642 priority patent/US8676937B2/en
Priority to US14/192,119 priority patent/US10142276B2/en
Application filed by Jeffrey Alan Rapaport, Seymour Rapaport, Kenneth Allen Smith, James Beattie, Gideon Gimlan filed Critical Jeffrey Alan Rapaport
Priority to US16/196,542 priority patent/US20190109810A1/en
Publication of US20190109810A1 publication Critical patent/US20190109810A1/en
Application status is Pending legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00Arrangements for user-to-user messaging in packet-switching networks, e.g. e-mail or instant messages
    • H04L51/32Messaging within social networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06QDATA PROCESSING SYSTEMS OR METHODS, SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation, e.g. computer aided management of electronic mail or groupware; Time management, e.g. calendars, reminders, meetings or time accounting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/16Arrangements for providing special services to substations
    • H04L12/18Arrangements for providing special services to substations for broadcast or conference, e.g. multicast
    • H04L12/1813Arrangements for providing special services to substations for broadcast or conference, e.g. multicast for computer conferences, e.g. chat rooms
    • H04L12/1818Conference organisation arrangements, e.g. handling schedules, setting up parameters needed by nodes to attend a conference, booking network resources, notifying involved parties
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network-specific arrangements or communication protocols supporting networked applications
    • H04L67/30Network-specific arrangements or communication protocols supporting networked applications involving profiles
    • H04L67/306User profiles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/835Generation of protective data, e.g. certificates
    • H04N21/8358Generation of protective data, e.g. certificates involving watermark
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network-specific arrangements or communication protocols supporting networked applications
    • H04L67/02Network-specific arrangements or communication protocols supporting networked applications involving the use of web-based technology, e.g. hyper text transfer protocol [HTTP]

Abstract

Disclosed is a Social-Topical Adaptive Networking (STAN) system that can inform users of cross-correlations between currently focused-upon topic or other nodes in a corresponding topic or other data-objects organizing space maintained by the system and various social entities monitored by the system. More specifically, one of the cross-correlations may be as between the top N now-hottest topics being focused-upon by a first social entity and the amounts of focus ‘heat’ that other social entities (e.g., friends and family) are casting on the same topics (or other subregions of other cognitive attention receiving spaces) in a relevant time period.

Description

    1. FIELD OF DISCLOSURE
  • The present disclosure of invention relates generally to online networking systems and uses thereof.
  • The disclosure relates more specifically to Social-Topical/contextual Adaptive Networking (STAN) systems that, among other things, empower co-compatible users to on-the-fly join into corresponding online chat or other forum participation sessions based on user context and/or on likely topics currently being focused-upon by the respective users. Such STAN systems can additionally provide transaction offerings to groups of people based on system determined contexts of the users, on system determined topics of most likely current focus and/or based on other usages of the STAN system by the respective users. Yet more specifically, one system disclosed herein maintains logically interconnected and continuously updated representations of communal cognitions spaces (e.g., topic space, keyword space, URL space, context space, content space and so on) where points, nodes or subregions of such spaces link to one another and/or to cross-related online chat or other forum participation opportunities and/or to cross-related informational resources. By automatically determining where in at least one of these spaces a given user's attention is currently being focused, the system can automatically provide the given user with currently relevant links to the interrelated chat or other forum participation opportunities and/or to the interrelated other informational resources. In one embodiment, such currently relevant links are served up as continuing flows of more up to date invitations that empower the user to immediately link up with the link targets.
  • 2A. CROSS REFERENCE TO AND INCORPORATION OF CO-OWNED NON PROVISIONAL APPLICATIONS
  • The following copending U.S. patent applications are owned by the owner of the present application, and their disclosures are incorporated herein by reference in their entireties as originally filed:
  • (A) Ser. No. 12/369,274 filed Feb. 11, 2009 by Jeffrey A. Rapaport et al. and which is originally entitled, ‘Social Network Driven Indexing System for Instantly Clustering People with Concurrent Focus on Same Topic into On Topic Chat Rooms and/or for Generating On-topic Search Results Tailored to User Preferences Regarding Topic', where said application was early published as US 2010-0205541 A1; and
  • (B) Ser. No. 12/854,082 filed Aug. 10, 2010 by Seymour A. Rapaport et al. and which is originally entitled, Social-Topical Adaptive Networking (STAN) System Allowing for Cooperative Inter-coupling with External Social Networking Systems and Other Content Sources.
  • 2B. CROSS REFERENCE TO AND INCORPORATION OF CO-OWNED PROVISIONAL APPLICATIONS
  • The following copending U.S. provisional patent applications are owned by the owner of the present application, and their disclosures are incorporated herein by reference in their entireties as originally filed:
  • (A) Ser. No. 61/485,409 filed May 12, 2011 by Jeffrey A. Rapaport, et al. [atty docket: RAPA17334-1V US] and entitled Social-Topical Adaptive Networking (STAN) System Allowing for Group Based Contextual Transaction Offers and Acceptances and Hot Topic Watchdogging;
  • and
  • (B) Ser. No. 61/551,338 filed Oct. 25, 2011 [atty docket: RAPA17334-2V US] and entitled Social-Topical Adaptive Networking (STAN) System Allowing for Group Based Contextual Transaction Offers and Acceptances and Hot Topic Watchdogging.
  • 2C. CROSS REFERENCE TO OTHER PATENTS/PUBLICATIONS
  • The disclosures of the following U.S. patents or Published U.S. patent applications are incorporated herein by reference:
  • (A) U.S. Pub. 20090195392 published Aug. 6, 2009 to Zalewski; Gary and entitled: Laugh Detector and System and Method for Tracking an Emotional Response to a Media Presentation;
  • (B) U.S. Pub. 2005/0289582 published Dec. 29, 2005 to Tavares, Clifford; et al. and entitled: System and method for capturing and using biometrics to review a product, service, creative work or thing;
  • (C) U.S. Pub. 2003/0139654 published Jul. 24, 2003 to Kim, Kyung-Hwan; et al. and entitled: System and method for recognizing user's emotional state using short-time monitoring of physiological signals; and
  • (D) U.S. Pub. 20030055654 published Mar. 20, 2003 to Oudeyer, Pierre Yves and entitled: Emotion recognition method and device.
  • Preliminary Introduction to Disclosed Subject Matter
  • Imagine a set of virtual elevator doors opening up on your N-th generation smart cellphone (a.k.a. smartphone) or tablet computer screen (where N≤3 here) and imagine an on-screen energetic bouncing ball hopping into the elevator, dragging you along visually with it into the insides of a dimly lighted virtual elevator. Imagine the ball bouncing back and forth between the elevator walls while blinking sets of virtual light emitters embedded in the ball illuminate different areas within the virtual elevator. You keep your eyes trained on the attention grabbing ball. What will it do next?
  • Suddenly the ball jumps to the elevator control panel and presses the button for floor number 86. A sign lights up next to the button. It glowingly says “Superbowl™ Sunday Party Today”. You already had a subconscious notion that this is where this virtual elevator ride was going to next take you. Surprisingly, another, softer lit sign on the control panel momentarily flashes the message: “Reminder: Help Grandma Tomorrow”. Then it fades. You are glad for the gentle reminder. You had momentarily forgotten that you promised to help Grandma with some chores tomorrow. In today's world of mental overload and overwhelming information deluges (and required cognition staminas for handling those deluges) it is hard to remember where to cast one's limited energies (of the cognitive kind) and when and how intensely to cast them on competing points of potential focus. It is impossible to focus one's attentions everywhere and at everything. The human mind has a problem in that, unlike the eye's relatively small and well understood blind spot (the eye's optic disc), the mind's conscious blind spots are vast and almost everywhere except in the very few areas one currently concentrates one's attentions on. Hopefully, the bouncing virtual ball will remember to remind you yet again, and at an appropriate closer time tomorrow that it is “Help Grandma Day”. (It will.) You make a mental note to not stay at today's party very late because you need to reserve some of your limited energies for tomorrow's chores.
  • Soon the doors of your virtual elevator open up and you find yourself looking at a refreshed display screen (the screen of your real life (ReL) intelligent personal digital assistant (a.k.a. PDA, smartphone or tablet computer). Now it has a center display area populated with websites related to today's Superbowl™ football game (the American game of football, not British “football”, a.k.a. soccer). On the left side of your screen is a list of friends whom you often like to talk to (literally or by way of electronic messaging) about sports related matters. Sometimes you forget one or two of them. But your computer system seems not to forget and thankfully lists all the vital ones for this hour's planned activities. Next to their names are a strange set of revolving pyramids with red lit bars disposed along the slanted side areas of those pyramids. At the top of your screen there is a virtual serving tray supporting a set of so-called, invitation-serving plates. Each serving plate appears to serve up a stack of pancake-like or donut-like objects, where the served stacks or combinations of pancake or donut-like objects each invites you to join a recently initiated, or soon-to-be-started, online chat and where the user-to-user exchanges of these chats are (or will be) primarily directed to your current topic of attention; which today at this hour happens to be on the day's Superbowl™ Sunday football game. Rather than you going out hunting for such chats, they appear to have miraculously hunted for, and found you instead. On the bottom of your screen is another virtual serving tray that is serving up a set of transaction offers related to buying Superbowl™ associated paraphernalia. One of the promotional offerings is for T-shirts with your favorite team's name on them and proclaiming them the champions of this year's climactic but-not-yet-played-out game. You think to yourself, “I'm ready to buy that, and I'm fairly certain my team will win”.
  • As you muse over this screenful of information that was automatically served up to you by your wirelessly networked computer device (e.g., smartphone) and as you muse over what today's date is, as well as considering the real life surroundings where you are located and the context of that location, you realize in the back of your mind that the virtual bouncing ball and its virtual elevator friend had guessed correctly about you, about where you are or where you were heading, your surrounding physical context, your surrounding social context, what you are thinking about at the moment (your mental context), your current emotional mood (happy and ready to engage with sports-minded friends of similar dispositions to yours) and what automatically presented invitations or promotional offerings you will now be ready to now welcome. Indeed, today is Superbowl™ Sunday and at the moment you are about to sit down (in real life) on the couch in your friend's house (Ken's house) getting ready to watch the big game on Ken's big-screen TV along with a few other like minded colleagues. The thing of it is that today you not only have the topic of the “Superbowl™ Sunday football game” as a central focal point or central attention receiving area in your mind, but you also have the unfolding dynamics of a real life social event (meeting with friends at Ken's house) as an equally important region of focus in your mind. If you had instead been sitting at home alone and watching the game on your small kitchen TV, the surrounding social dynamics probably would not have been such a big part of your current thought patterns. However, the combination of the surrounding physical cues and social context inferences plus the main topic of focus in your mind places you in Ken's house, in front of his big screen, high definition TV and happily trading quips with similarly situated friends sitting next to you.
  • You surmise that the smart virtual ball inside your smartphone (or inside another mobile data processing device) and whatever external system it wirelessly connects with must have been empowered to use a GPS and/or other sensor embedded in the smart cellphone (or tablet or other mobile device) as well as to use your online digitized calendar to make best-estimate guesses at where you are (or soon will be), which other people are near you (or soon will be with you), what symmetric or asymmetric social relations probably exist between you and the nearby other people, what you are probably now doing, how you mentally perceive your current context, and what online content you might now find to be of greatest and most welcomed interest to you due to your currently adopted contexts and current points of focus (where, ultimately in this scenario; you are the one deciding what your currently adopted contexts are: e.g., Am I at work or at play? and which if any of the offerings automatically presented to you by your mobile data processing device you will now accept).
  • Perhaps your mobile data processing device was empowered, you further surmise; to pick up on sounds surrounding you (e.g., sounds from the turned-on TV set) or images surrounding you (e.g., sampled video from the TV set as well as automatically recognized faces of friends who happen to be there in real life (ReL)) and it was empowered to report these context-indicating signals to a remote and more powerful data processing system by way of networking? Perhaps that is how the limited computing power associated with your relatively small and low powered smartphone determined your most likely current physical and mental contexts? The question intrigues you for only a flash of a moment and then you are interrupted in your thoughts by Ken offering you a bowl full of potato chips.
  • With thoughts about how the computer systems might work quickly fading into the back of your subconscious, you thank Ken and then you start paying conscious attention to one of the automatically presented websites now found within a first focused-upon area of your smartphone screen. It is reporting on the health condition of your favorite football player, Joe-the-Throw Nebraska (best quarterback, in your humble opinion; since Joe Montana (a.k.a. “Golden Joe”, “Comeback Joe”) hung up his football cleats). Meanwhile in your real life background, the Hi-Def TV is already blaring with the pre-game announcements and Ken has started blasting some party music from the kitchen area while he opens up more bags of pretzels and potato chips. As you return focus to the web content presented by your PDA-style (Personal Digital Assistant type) smartphone, a small on-screen advertisement icon pops up next to the side of the athlete's health-condition reporting frame. You hover a pointer over it and the advertisement icon automatically expands to say: “Pizza: Big Local Discount, Only while it lasts, First 10 Households, Press here for more”. This promotional offering you realize is not at all annoying to you. Actually it is welcomed. You were starting to feel a wee bit hungry just before the ad popped up. Maybe it was the sound and smell of the bags of potato chips being opened in the kitchen or maybe it was the party music. You hadn't eaten pizza in a while and the thought of it starts your mouth salivating. So you pop the small teaser advertisement open to see even more.
  • The further enlarged promotional informs you that at least 50 households in your current, local neighborhood are having similar Superbowl™ Sunday parties and that a reputable pizza store nearby is ready to deliver two large sized pizza pies to each accepting household at a heavily discounted price, where the offered deal requires at least 10 households in the same, small radius neighborhood to accept the deal within the next 30 minutes; otherwise the deal lapses. Additional pies and other items are available at different discount rates, first not as good of a deal as the opening teaser rate, but then getting better and better again as you order larger and larger volumes (or more expensive ones) of those items. (In an alternate version of this hypothetical story, the deal minimum is not based on number of households but rather on number of pizzas ordered, or number of people who send their email addresses to the promoter or on some other basis that may be beneficial to the product vendor for reasons known to him. Also, in an alternate version, special bonus prizes are promised if you convince the next door neighbor to join in on your group order so that two adjacent houses are simultaneously ordering from the same pizza store.)
  • This promotional offering not only sounds like a great deal for you, but as you think on it some more, you realize it is also a win-win deal for the local pizza pie vendor. The pizza store owner can greatly reduce his delivery overhead costs by delivering in one delivery run, a large volume of same-time ordered pizzas to a same one local neighborhood (especially if there are a few large-sized social gatherings i.e., parties, in the one small-radiused neighborhood) and all the pizzas should be relatively fresh if the 10 or more closely-located households all order in the allotted 30 minutes (which could instead be 20 minutes, 40 minutes or some other number). Additionally, the pizza store can time a mass-production run of the pizzas, and a common storage of the volume-ordered hot pizzas (and of other co-ordered items) so they will all arrive fresh and hot (or at least lukewarm) in the next hour to all the accepting customers in the one small neighborhood. Everyone ends up pleased with this deal; customers and promoter. Additionally, if the pizza store owner can capture new customers at the party because they are impressed with the speed and quality of the delivery and the taste and freshness of the food, that is one additional bonus for the promotion offering vendor (e.g., the local pizza store).
  • You ask around the room and discover that a number of other people at the party (in Ken's house, including Ken) are also very much in the mood for some hot fresh pizza. One of them has his tablet computer running and he just got the same promotional invitation from the same vendor and, as a matter of fact, he was about to ask you if you wanted to join with him in signing up for the deal. He too indicates he hasn't had pizza in a week and therefore he is “game” for it. Now Jim chimes in and says he wants spicy chicken wings to go along with his pizza. Another friend (Jeff) tells you not to forget the garlic bread. Sye, another friend, says we need more drinks, it's important to hydrate (he is always health conscious). As you hit the virtual acceptance button within your on-screen offer, you begin to wonder; how did the pizza store, or more correctly your smartphone's computer and whatever it is remotely connected to; know this would happen just now—that all these people would welcome this particular promotional offering? You start filling in the order details on your screen while keeping an eye on an on-screen deal-acceptance counter. The deal counter indicates how many nearby neighbors have also signed up for the neighborhood group discount (and/or other promotional offering) before the offer deadline lapses. Next to the sign-up count there is a countdown timer decrementing from 30 minutes towards zero. Soon the required minimum number of acceptances is reached, well before the countdown timer reaches zero. How did all this come to be? Details will follow below.
  • After you place the pizza order, a not-unwelcomed further suggestion icon or box pops open on your screen. It says: “This is the kind of party that your friends A) Henry and B) Charlie would like to be at, but they are not present. Would you like to send a personalized invitation to one or more of them? Please select: 0) No, 1) Initiate Instant Chat, 2) Text message to their cellphones or tablets using pre-drafted invitation template, 3) Dial their cellphone or other device now for personal voice invite, 4) Email, 5) more . . . ”. The automatically generated suggestion further says, “Please select one of the following, on-topic messaging templates and select the persons (A,B,C, etc.) to apply it to.” The first listed topic reads: “SuperBowl Party, Come ASAP”. You think to yourself, yes this is indeed a party where Charlie is sorely missed. How did my computer realize this when it had slipped my mind? I'm going to press the number 2) “Text message” option right now. In response to the press, a pre-drafted invitation template addressed to Charlie automatically pops open. It says: “Charlie, We are over at Ken's house having a Superbowl™ Sunday Party. We sorely miss you. Please join ASAP. P.S. Do you want pizza?” Further details for empowering this kind of feature will follow below.
  • Your eyes flick back to the on-screen news story concerning the health of your favorite sports celebrity (Joe-the-Throw Nebraska—a hypothetical name). A new frame has now appeared next to it: “Will Joe Throw Today?”. You start reading avidly. In the background, the doorbell rings. Someone says, “Pizza is here!” The new frame on your screen says “Best Chat Comments re Joe's Health”. From experience you know that this is a compilation of contributions collected from numerous chat rooms, blog comments, etc.; a sort of community collection of best and voted most-worthy-to-see comments so far regarding the topic of Joe-the-Throw Nebraska, his health status and today's American football game. You know from past experience that these “community board” type of comments have been voted on, and have been ranked as the best liked and/or currently ‘hottest’ and they are all directed to substantially the same topic you are currently centering your attention on, namely, the health condition of your favorite sports celebrity's (e.g., “Is Joe well enough to play full throttle today?”) and how it will impact today's game. The best comments have percolated to the top of the list (a.k.a., community board). You have given up trying to figure out how your smartphone (and whatever computer system it is wirelessly hooked up to) can do this too. Details for empowering this kind of feature will also follow below.
  • Definitions
  • As used herein, terms such as “cloud”, “server”, “software”, “software agent”, “BOT”, “virtual BOT”, “virtual agent”, “virtual ball”, “virtual elevator” and the like do not mean nonphysical abstractions but instead always entail a physically real and tangibly implemented aspect unless otherwise explicitly stated to the contrary at that spot.
  • Claims appended hereto which use such terms (e.g., “cloud”, “server”, “software”, etc.) do not preclude others from thinking about, speaking about or similarly non-usefully using abstract ideas, or laws of nature or naturally occurring phenomenon. Instead, such “virtual” or non-virtual entities as described herein are always accompanied by changes of physical state of real physical, tangible and non-transitory objects. For example, when it is in an active (e.g., an executing) mode, a “software” module or entity, be it a “virtual agent”, a spyware program or the alike is understood to be a physical ongoing process (at the time it is executed) which is being carried out in one or more real, tangible and specific physical machines (e.g., data processing machines) where the machine(s) entropically consume(s) electrical power and/or other forms of real energy per unit time as a consequence of said physical ongoing process being carried out there within. Parts or wholes of software implementations may be substituted for by substantially similar in functionality hardware or firmware including for example implementation of functions by way of field programmable gate arrays (FPGA's) or other such programmable logic devices (PLD's). When it is in a static (e.g., non-executing) mode, an instantiated “software” entity or module, or “virtual agent” or the alike is understood (unless explicitly stated otherwise herein) to be embodied as a substantially unique and functionally operative and nontransitory pattern of transformed physical matter preserved in a more-than-elusively-transitory manner in one or more physical memory devices so that it can functionally and cooperatively interact with a commandable or instructable machine as opposed to being merely descriptive and totally nonfunctional matter. The one or more physical memory devices mentioned herein can include, but are not limited to, PLD's and/or memory devices which utilize electrostatic effects to represent stored data, memory devices which utilize magnetic effects to represent stored data, memory devices which utilize magnetic and/or other phase change effects to represent stored data, memory devices which utilize optical and/or other phase change effects to represent stored data, and so on.
  • As used herein, the terms, “signaling”, “transmitting”, “informing” “indicating”, “logical linking”, and the like do not mean nonphysical and abstract events but rather physical and not elusively transitory events where the former physical events are ones whose existence can be verified by modern scientific techniques. Claims appended hereto that use the aforementioned terms, “signaling”, “transmitting”, “informing”, “indicating”, “logical linking”, and the like or their equivalents do not preclude others from thinking about, speaking about or similarly using in a non-useful way abstract ideas, laws of nature or naturally occurring phenomenon.
  • As used herein, the terms, “empower”, “empowerment” and the like refer to a physically transformative process that provides a present or near-term ability to a data producing/processing device or the like to be recognized by and/or to communicate with a functionally more powerful data processing system (e.g., an on network or in cloud server) where the provided abilities include at least one of: transmitting status reporting signals to, and receiving responsive information-containing signals from the more powerful data processing system where the more powerful system will recognize at least some of the reporting signals and will responsively change stored state-representing signals for a corresponding one or more system-recognized personas and/or for a corresponding one or more system-recognized and in-field data producing and/or data processing devices and where at least some of the responsive information-containing signals, if provided at all, will be based on the stored state-representing signals. The term, “empowerment” may include a process of registering a person or persona (real or virtual) or a process of logging in a registered entity for the purpose of having the functionally more powerful data processing system recognize that registered entity and respond to reporting signals associated with that recognized entity. The term, “empowerment” may include a process of registering a data processing and/or data-producing and/or information inputting and/or outputting device or a process of logging in a registered such device for the purpose of having the functionally more powerful data processing system recognize that registered device and respond to reporting signals associated with that recognized device and/or supply information-containing and/or instruction-containing signals to that recognized device.
  • Background and Further Introduction to Related Technology
  • The above identified and herein incorporated by reference U.S. patent application Ser. No. 12/369,274 (filed Feb. 11, 2009) and Ser. No. 12/854,082 (filed Aug. 10, 2010) disclose certain types of Social-Topical Adaptive Networking (STAN) Systems (hereafter, also referred to respectively as “Sierra#1” or “STAN_1” and “Sierra#2” or “STAN_2”) which empower and enable physically isolated online users of a network to automatically join with one another (electronically or otherwise) so as to form a topic-specific and/or otherwise based information-exchanging group (e.g., a ‘TCONE’—as such is described in the STAN_2 application). A primary feature of the STAN systems is that they provide and maintain one or more so-called, topic space defining objects (e.g., topic-to-topic associating database records) which are represented by physical signals stored in machine memory and which topic space defining objects can define (and thus model) topic nodes and logical interconnections (cross-associations) between, and/or spatial clusterings of those nodes and/or can provide logical links to forums associated with topics modeled by the respective nodes and/or to persons or other social entities associated with topics of the nodes and/or to on-topic other material associated with topics of the nodes. The topic space defining objects (e.g., database records, also referred to herein as potentially-attention-receiving modeled points, nodes or subregions of a Cognitive Attention Receiving Space (CARS), which space in this case is topic space) can be used by the STAN systems to automatically provide, for example, invitations to plural persons or to other social entities to join in on-topic online chats or other Notes Exchange sessions (forum sessions) when those social entities are deemed to be currently focusing-upon (e.g., casting their respective attention giving energies on) such topics or clusters of such topics and/or when those social entities are deemed to be co-compatible for interacting at least online with one another. (In one embodiment, co-compatibilities are established by automatically verifying reputations and/or attributes of persons seeking to enter a STAN-sponsored chat room or other such Notes Exchange session, e.g., a Topic Center “Owned” Notes Exchange session or “TCONE”.) Additionally, the topic space defining objects (e.g., database records) are used by the STAN systems to automatically provide suggestions to users regarding on-topic other content and/or regarding further social entities whom they may wish to connect with for topic-related activities and/or socially co-compatible activities.
  • During operation of the STAN systems, a variety of different kinds of informational signals may be collected by a STAN system in regard to the current states of its users; including but not limited to, the user's geographic location, the user's transactional disposition (e.g., at work? at a party? at home? etc.); the user's recent online activities; the user's recent biometric states; the user's habitual trends, behavioral routines, the user's biological states (e.g., hungry tired, muscles fatigued from workout) and so on. The purpose of this collected information is to facilitate automated joinder of like-minded and co-compatible persons for their mutual benefit. More specifically, a STAN-system-facilitated joinder may occur between users at times when they are in the mood to do so (to join in a so-called Notes Exchange session) and when they have roughly concurrent focus on same or similar detectable content and/or when they apparently have approximately concurrent interest in a same or similar particular topic or topics and/or when they have current personality co-compatibility for instantly chatting with, or for otherwise exchanging information with one another or otherwise transacting with one another.
  • In terms of a more concrete example of the above concepts, the imaginative and hypothetical introduction that was provided above revolved around a group of hypothetical people who all seemed to be currently thinking about a same popular event (the day's Superbowl™ football game) and many of whom seemed to be concurrently interested in then obtaining event-relevant refreshments (e.g., pizza) and/or other event-relevant paraphernalia (e.g., T-shirts). The group-based discount offer sought to join them, along with others, in an online manner for a mutually beneficial commercial transaction (e.g., volume purchase and localized delivery of a discounted item that is normally sold in smaller quantities to individual and geographically dispersed customers one at a time). The unsolicited and thus “pushed” solicitation was not one that generally annoyed the recipients as would conventionally pushed unsolicited and undesired advertisements. It's almost as if the users pulled the solicitation in to them by means of their subconscious will power rather than having the solicitations rudely pushed onto them by an insistent high pressure salesperson. The underlying mechanisms that can automatically achieve this will be detailed below. At this introductory phase of the present disclosure it is worthwhile merely to note that some wants and desires can arise at the subconscious level and these can be inferred to a reasonable degree of confidence by carefully reading a person's facial expressions (e.g., micro-expressions) and/or other body gestures, by monitoring the persons' computer usage activities, by tracking the person's recent habitual or routine activities, and so on, without giving away that such is going on and without inappropriately intruding on reasonable expectations of privacy by the person. Proper reading of each individual's body-language expressions may require access to a Personal Emotion Expression Profile (PEEP) that has been pre-developed for that individual and for certain contexts in which the person may find themselves. Example structures for such PEEP records are disclosed in at least one of the here incorporated U.S. Ser. No. 12/369,274 and Ser. No. 12/854,082. Appropriate PEEP records for each individual may be activated based on automated determination of time, place and other context revealing hints or clues (e.g., the individual's digitized calendar or recent email records which show a plan, for example, to attend a certain friend's “Superbowl™ Sunday Party” at a pre-arranged time and place, for example 1:00 PM at Ken's house). Of course, user permission for accessing and using such information should be obtained by the system beforehand, and the users should be able to rescind the permissions whenever they want to do so, whether manually or by automated command (e.g., IF Location=Charlie's Tavern THEN Disable All STAN monitoring”). In one embodiment, user permission automatically fades over time for all or for one or more prespecified regions of topic space and needs to be reestablished by contacting the user and either obtaining affirmative consent or permission from the user or at least notifying the user and reminding the user of the option to rescind. In one embodiment, certain prespecified regions of topic space are tagged by system operators and/or the respective users as being of a sensitive nature and special double permissions are required before information regarding user direct or indirect ‘touchings’ into these sensitive regions of topic space is automatically shared with one or more prespecified other social entities (e.g., most trusted friends and family).
  • Before delving deeper into such aspects, a rough explanation of the term “STAN system” as used herein is provided. The term arises from the nature of the respective network systems, namely, STAN_1 as disclosed in here-incorporated U.S. Ser. No. 12/369,274 and STAN_2 as disclosed in here-incorporated U.S. Ser. No. 12/854,082. Generically they are referred to herein as Social-Topical ‘Adaptive’ Networking (STAN) systems or STAN systems for short. One of the things that such STAN systems can generally do is to maintain in machine memory one or more virtual spaces (data-objects organizing spaces) populated by interrelated data objects stored therein such as interrelated topic nodes (or ‘topic centers’ as they are referred to in the Ser. No. 12/854,082 application) where the nodes may be hierarchically interconnected (via logical graphing) to one another and/or logically linked to topic-related forums (e.g., online chat rooms) and/or to topic-related other content. Such system-maintained and logically interconnected and continuously updated representations of topic nodes and associated forums (e.g., online chat rooms) may be viewed as social and dynamically changing communal cognition spaces. (The definition of such communal cognition spaces is expanded on herein as will be seen below.) In accordance with one aspect of the present disclosure, if there are not enough online users tethered to one topic node so as to adequately fill a social mix recipe of a given chat or other forum participation session, users from hierarchically and/or spatially nearby other topic nodes those of substantially similar topic may be automatically recruited to fill the void. In other words, one chat room can simultaneously service plural ones of topic nodes. (The concept of social mix recipe will be explained later below.) The STAN_1 and STAN_2 systems (as well as the STAN_3 of the present disclosure) can cross match current users with respective topic nodes that are determined by machine means as representing topics likely to be currently focused-upon ones in the respective users' minds. The STAN systems can also cross match current users with other current users (e.g., co-compatible other users) so as to create logical linkages between users where the created linkages are at least one if not both of being topically relevant and socially acceptable for such users of the STAN system. Incidentally, hierarchical graphing of topic-to-topic associations (T2T) is not a necessary or only way that STAN systems can graph T2T associations via a physical database or otherwise. Topic-to-topic associations (T2T) may alternatively or additionally be defined by non-hierarchical graphs (ones that do not have clear parent to child relationships as between nodes) and/or by spatial and distance based positionings within a specified virtual positioning space.
  • The “adaptive” aspect of the “STAN” acronym correlates in one sense to the “plasticity” (neuroplasticity) of the individual human mind and correlates in a second sense to a similar “plasticity” of the collective or societal mind. Because both individualized people and groups thereof; and their respective areas of focused attention tend to change with time, location, new events and variation of physical and/or social context (as examples), the STAN systems are structured to adaptively change (e.g., update) their definitions regarding what parts of a system-maintained, Cognitive Attention Receiving Space (referred to herein also as a “CARS”) are currently cross-associated with what other parts of the same CARS and/or with what specific parts of other CARS. The adaptive changes can also modify what the different parts currently represent (e.g., what is the current definition of a topic of a respective topic node when the CARS is defined as being the topic space). The adaptive changes can also vary the assigned intensity of attention giving energies for respective users when the users are determined by the machine means to be focused-upon specific subareas within, for example, a topics-defining map (e.g., hierarchical and/or spatial). The adaptive changes can also determine how and/or at what rate the cross-associated parts (e.g., topic nodes) and their respective interlinkings and their respective definitions change with changing times and changing external conditions. In other words, the STAN systems are structured to adaptively change the topics-defining maps themselves (a.k.a. topic spaces, which topic maps/spaces have corresponding, physically represented, topic nodes or the like defined by data signals recorded in databases or other appropriate memory means of the STAN_system and which topic nodes or groups thereof can be pointed to with logical pointer mechanisms). Such adaptive change of perspective regarding virtual positions or graphed interlinks in topic space and/or reworking of the topic space and of topic space content (and/or of alike subregions of other Cognitive Attention Receiving Spaces) helps the STAN systems to keep in tune with variable external conditions and with their variable user populations as the latter migrate to new topics (e.g., fad of the day) and/or to new personal dispositions (e.g., higher levels of expertise, different moods, etc.).
  • One of the adaptive mechanisms that can be relied upon by the STAN system is the generation and collection of implicit vote or CVi signals (where CVi may stand for Current (and implied or explicit) Vote-Indicating record). CVi's are vote-representing signals which are typically automatically collected from user surrounding machines and used to infer subconscious positive or negative votes cast by users as they go about their normal machine usage activities or normal life activities, where those activities are open to being monitored (due to rescindable permissions given by the user for such monitoring) by surrounding information gathering equipment. User PEEP files may be used in combination with collected CFi and CVi signals to automatically determine most probable, user-implied votes regarding focused-upon material even if those votes are only at the subconscious level. Stated otherwise, users can implicitly urge the STAN system topic space and pointers thereto to change (or pointers/links within the topic space to change) in response to subconscious votes that the users cast where the subconscious votes are inferred from telemetry gathered about user facial grimaces, body language, vocal grunts, breathing patterns, eye movements, and the like. (Note: The above notion of a current cross-association between different parts of a same CARS (e.g., topic space or some other Cognitive Attention Receiving Space) is also referred to herein as an IntrA-Space cross-associating link or “InS-CAX” for short. The above notion of a current cross-association between points, nodes or subregions of different CARS's is also referred to herein as an IntEr-Space cross-associating link or “IoS-CAX” for short, where the “o” in the “IoS-CAX” acronym signifies that the link crosses to outside of the respective space. See for example, IoS-CAX 370.6 of FIG. 3E and IoS-CAX 390.6 of the same figure where these will be further described later below.)
  • Although not specifically given as an example in the earlier filed and here incorporated U.S. Ser. No. 12/854,082 (STAN_2), one example of a changing and “neuro-plastic” cognition landscape might revolve around a keyword such as “surfing”. In the decade of the 1960's, the word “surfing” may most likely have conjured up in the minds of most individuals and groups, the notion of waves breaking on a Hawaiian or Californian beach and young men taking to the waves with their “surf boards” so they can ride or “surf” those waves. By contrast, after the decade of the 1990's, the word “surfing” may more likely have conjured up in the minds of most up-to-date individuals (and groups of the same), the notion of people using personal computers and using the Internet and searching through it (surfing the net) to find websites of interest. Moreover, in the decade of the 1960's there was essentially no popular attention giving activities directed to the notion of “surfing” meaning the idea of journeying through webs of data by means of personally controlled computers. By contrast, beginning with the decade of the 1990's (and the explosive growth of the World Wide Web), it became exponentially more and more popular to focus one's attention giving energies on the notion of “surfing” as it applies to riding through the growing mounds of information found on the World Wide Web or elsewhere within the Internet and/or within other network systems. Indeed, another word that changed in meaning in a plastic cognition way is the word sounded out as “Google”. In the decade of the 1960's such a sounded out word (more correctly spelled as “Googol”) was understood to mean the number 10 raised to the 100th power. Thinking about sorting through a Googol-ful of computerized data meant looking for a needle in a haystack. The likelihood of finding the sought item was close to nil. Ironically, with the advent of the internet searching engine known as Google™, the probability of finding a website whose content matches with user-picked keywords increased dramatically and the popularly assumed meaning for the corresponding sound bite (“Googol” or “Google”) changed, and the topics cross-correlated to that sound bite also changed; quite significantly.
  • The sounded-out words, “surfing and “Google” are but two of many examples of the “plasticity” attribute of the individual human mind and of the “plasticity” attribute of the collective or societal mind. Change has and continues to come to many other words, and to their most likely meanings and to their most likely associations to other words (and/or other cognitions). The changes can come not only due to passage of time, be it over a period of years; or sometimes over a matter of days or hours, but also due to unanticipated events (e.g., the term “911”—pronounced as nine eleven—took on sudden and new meaning on Sep. 11, 2001). Other examples of words or phrases that have plastically changed over time include, being “online”, opening a “window”, being infected by a “virus”, looking at your “cellular”, going “phishing”, worrying about “climate change”, “occupying” a street such as one named Wall St., and so on. Indeed, not only do meanings and connotations of same-sounding words change over time, but new words and new ideas associated with them are constantly being added. The notion of having an adaptive and user-changeable topic space was included even in the here-incorporated STAN_1 disclosure (U.S. Ser. No. 12/369,274).
  • In addition to disclosing an adaptively changing topics space/map (topic-to-topic (T2T) associations space), the here also-incorporated U.S. Ser. No. 12/854,082 (STAN_2) discloses the notion of a user-to-user (U2U) associations space as well as a user-to-topic (U2T) cross associations space. Here, an extension of the user-to-user (U2U) associations space will be disclosed where that extension will be referred to as Social/Persona Entities Interrelation Spaces (SPEIS'es for short). A single such space is a SPEIS. However, there often are many such spaces due to the typical presence of multiple social networking (SN) platforms like FaceBook™, LinkedIn™, MySpace™, Quora™, etc. and the many different kinds of user-to-user associations which can be formed by activities carried out on these various platforms in addition to user activities carried out on a STAN platform. The concept of different “personas” for each one real world person was explained in the here incorporated U.S. Ser. No. 12/854,082 (STAN_2). In this disclosure however, Social/Persona Entities (SPE's) may include not only the one or different personas of a real world, single flesh and blood person, but also personas of hybrid real/virtual persons (e.g., a Second Life™ avatar driven by a committee of real persons) and personas of collectives such as a group of real persons and/or a group of hybrid real/virtual persons and/or purely virtual persons (e.g., those driven entirely by an executing computer program). In one embodiment, each STAN user can define his or her own custom groups or the user can use system-provided templates (e.g., My Immediate Family). The Group social entity may be used to keep a collective tab on what a relevant group of social entities are doing (e.g., What topic or other thing are they collectively and recently focusing-upon?).
  • When it comes to automated formation of social groups, one of the extensions or improvements disclosed herein involves formation of a group of online real persons who are to be considered for receiving a group discount offer (e.g., reduced price pizza) or another such transaction/promotional offering. More specifically, the present disclosure provides for a machine-implemented method that can use the automatically gathered CFi and/or CVi signals (current focus indicator and current voting indicator signals respectively) of a STAN system advantageously to automatically infer therefrom what unsolicited solicitations (e.g., group offers and the like) would likely be welcome at a given moment by a targeted group of potential offerees (real or even possibly virtual if the offer is to their virtual life counterparts, e.g., their SecondLife™ avatars) and which solicitations would less likely be welcomed and thus should not be now pushed onto the targeted personas, because of the danger of creating ill-will or degrading previously developed goodwill. Another feature of the present disclosure is to automatically sort potential offerees according to likelihood of welcoming and accepting different ones of possible solicitations and pushing the M most likely-to-be-now-welcomed solicitations to a corresponding top N ones of the potential offerees who are currently likely to accept (where here M and N are corresponding predetermined numbers). Outcomes can change according to changing moods/ideas of socially-interactive user populations as well as those of individual users (e.g., user mood or other current user persona state). A potential offeree who is automatically determined to be less likely to welcome a first of simultaneously brewing group offers may nonetheless be determined to more likely to now welcome a second of the brewing group offers. Thus brewing offers are competitively and automatically sorted by machine means so that each is transmitted (pushed) to a respective offerees population that is populated by persons deemed most likely to then accept that offer and offerees are not inundated with too many or unwelcomed offers. More details follow below.
  • Another novel use disclosed herein of the Group entity is that of tracking group migrations and migration trends through topic space and/or through other cognition cross-associating spaces (e.g., keyword space, context space, etc.). If a predefined group of influential personas (e.g., Tipping Point Persons) is automatically tracked as having traveled along a sequence of paths or a time parallel set of paths through topic space (by virtue of making direct or indirect ‘touchings’ in topic space), then predictions can be automatically made about the paths that their followers (e.g., twitter fans) will soon follow and/or of what the influential group will next likely do as a group. This can be useful for formulating promotional offerings to the influential group and/or their followers. Also, the leaders may be solicited by vendors for endorsing vendor provided goods and/or services. Detection of sequential paths and/or time parallel paths through topic space is not limited to predefined influential groups. It can also apply to individual STAN users. The tracking need not look at (or only at) the topic nodes they directly or indirectly ‘touched’ in topic space. It can include a tracking of the sequential and/or time parallel patterns of CFi's and/or CVi's (e.g., keywords, meta-tags, hybrid combinations of different kinds of CFi's (e.g., keywords and context-reporting CFi's), etc.) produced by the tracked individual STAN users. Such trackings can be useful for automatically formulating promotional offerings to the corresponding individuals. In one embodiment, so-called, hybrid spaces are created and represented by data stored in machine memory where the hybrid spaces can include but are not limited to, a hybrid topic-and-context space, a hybrid keyword-and-context space, a hybrid URL-and-context space, whereby system users whose recently collected CFi's indicate a combination of current context and current other focused-upon attribute (e.g., keyword) can be identified and serviced according to their current dispositions in the respective hybrid spaces and/or according to their current trajectories of journeying through the respective hybrid spaces.
  • It is to be understood that this background and further introduction section is intended to provide useful background for understanding the here disclosed inventive technology and as such, this technology background section may and probably does include ideas, concepts or recognitions that were not part of what was known or appreciated by others skilled in the pertinent arts prior to corresponding invention dates of invented subject matter disclosed herein. As such, this background of technology section is not to be construed as any admission whatsoever regarding what is or is not prior art. A clearer picture of the inventive technology will unfold below.
  • SUMMARY
  • In accordance with one aspect of the present disclosure, likely to-be-welcomed group-based offers or other offers are automatically presented to STAN system users based on information gathered from their STAN (Social-Topical Adaptive Networking) system usage activities. The gathered information may include current mood or disposition as implied by a currently active PEEP (Personal Emotion Expression Profile) of the user as well as recently collected CFi signals (Current Focus indicator signals), recently collected CVi signals (Current Voting (implicit or explicit indicator signals) and recently collected context-indicating signals (e.g., XP signals) uploaded for the user and recent topic space (TS) usage patterns or hybrid space (HS) usage patterns or attention giving energies being recently cast onto other Cognitive Attention Receiving Points, Nodes or SubRegions (CAR PNoS's) of other cognition cross-associating spaces (CARS) maintained by the system or trends therethrough as detected of the user and/or associated group and/or recent friendship space usage patterns or trends detected of the user (where latter is more correctly referred to here as recent SPEIS'es usage patterns or trends {usage of Social/Persona Entities Interrelation Spaces}). Current mood and/or disposition may be inferred from currently focused-upon nodes and/or subregions of other spaces besides just topic space (TS) as well as from detected hints or clues about the user's real life (ReL) surroundings (e.g., identifying music playing in the background or other sounds and/or odors emanating from the background, such as for example the sounds and/or smells of potato chip bags being popped open at the hypothetical “Superbowl™ Sunday Party” described above).
  • In accordance with another aspect of the present disclosure, various user interface techniques are provided for allowing a user to conveniently interface (even when using a small screen portable device; e.g., smartphone) with resources of the STAN system including by means of device tilt, body gesture, facial expressions, head tilt and/or wobble inputs and/or touch screen inputs as well as pupil pointing, pupil dilation changes (independent of light level change), eye widening, tongue display, lips/eyebrows/tongue contortions display, and so on, as such may be detected by tablet and/or palmtop and/or other data processing units proximate to STAN system users and communicating with telemetry gathering resources of a STAN system.
  • Although numerous examples given herein are directed to situations where the user of the STAN_system is carrying a small-sized mobile data processing device such as a tablet computer with a tappable touch screen, it is within the contemplation of the present disclosure to have a user enter an instrumented room or other such area (e.g., instrumented with audio visual display resources and other user interface resources) and with the user having essentially no noticeable device in hand, where the instrumented area automatically recognizes the user and his/her identity, automatically logs the user into his/her STAN_system account, automatically presents the user with one or more of the STAN_system generated presentations described herein (e.g., invitations to immediately join in on chat or other forum participation sessions related to a subportion of a Cognitive Attention Receiving Space, which subportion the user is deemed to be currently focusing-upon) and automatically responds to user voice and/or gesture commands and/or changes in user biometric states.
  • In accordance with yet another aspect of the present disclosure, a user-viewable screen area is organized to have user-relevant social entities (e.g., My Friends and Family) iconically represented in one subarea (e.g., hideable side tray area) of the screen and user-relevant topical and contextual material (e.g., My Top 5 Now Topics While Being Here) iconically represented in another subarea (e.g., hideable top tray area) of the screen, where an indication is provided to the user regarding which user-relevant social entities are currently focusing-upon which user-relevant topics (and/or other points, nodes or subregions in other Cognitive Attention Receiving Spaces). Thus the user can readily appreciate which of persons or other social entities relevant to him/her (e.g., My Friends and Family, My Followed Influencers) are likely to be currently interested in what topics that are same or similar (as measured by hierarchical and/or spatial distances in topic space) to those being current focused-upon by the user in the user's current context (e.g., at a bus stop, bored and waiting for the bus to arrive) or in topics that the user has not yet focused-upon. Alternatively, when the on-screen indications are provided to the user with regard to other points, nodes or subregions in other Cognitive Attention Receiving Spaces (e.g., keyword space, URL space, content space) the user can learn of user-relevant other social entities who are currently focusing-upon such user-relevant other spaces (including upon same or similar base symbols in a clustered symbols layer of the respective Cognitions-representing Space (CARS)).
  • Other aspects of the disclosure will become apparent from the below yet more detailed description.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The below detailed description section makes reference to the accompanying drawings, in which:
  • FIG. 1A is a block diagram of a portable tablet microcomputer which is structured for electromagnetic linking (e.g., electronically and/or optically linking, this including wirelessly linking) with a networking environment that includes a Social-Topical Adaptive Networking (STAN_3) system where, in accordance with the present disclosure, the STAN_3 system includes means for automatically creating individual or group transaction offerings based on usages of the STAN_3 system;
  • FIG. 1B shows in greater detail, a multi-dimensional and rotatable “current heats” indicating construct that may be used in a so-called, SPEIS radar display column of FIG. 1A where the illustrated heats indicating construct is indicative of intensity of current focus (or earlier timed focus) on certain topic nodes of the STAN_3 system by certain SPE's (Social/Persona Entities) who are context wise related to a top-of-column SPE (e.g., “Me”);
  • FIG. 1C shows in greater detail, another multi-dimensional and rotatable “heats” indicating construct that may be used in the radar display column of FIG. 1A where the illustrated heats indicating construct is indicative of intensity of discussion or other data exchanges as may be occurring between pairs of persons or groups of persons (SPE's) when using the STAN_3 system;
  • FIG. 1D shows in greater detail, another way of displaying current or previous heats as a function of time and of personas or groups involved and/or of topic nodes (or nodes/subregions of other spaces) involved;
  • FIG. 1E shows a machine-implemented method for determining what topics are currently the top N topics being focused-upon by each social entity;
  • FIG. 1F shows a machine-implemented system for computing heat attributes that are attributable to a respective first user (e.g., Me) and to a cross-correlation between a given topic space region and a preselected one or more second users (e.g., My Friends and Family) of the system;
  • FIG. 1G shows an automated community board posting system that includes a posts ranking and/or promoting sub-system in accordance with the disclosure;
  • FIG. 1H shows an automated process that may be used in conjunction with the automated community board posting and posts ranking/promoting system of FIG. 1G;
  • FIG. 1I shows a cell/smartphone or tablet computer having a mobile-compatible user interface for presenting 1-click chat-now and alike, on-topic joinder opportunities to users of the STAN_3 system;
  • FIG. 1J shows a smartphone and tablet computer compatible user interface method for presenting on-topic location based congregation opportunities to users of the STAN_3 system where the congregation opportunities may depend on availability of local resources (e.g., lecture halls, multimedia presentation resources, laboratory supplies, etc.);
  • FIG. 1K shows a smartphone and tablet computer compatible user interface method for presenting an M out of N, now commonly focused-upon topics and optional location based chat or other joinder opportunities to users of the STAN_3 system;
  • FIG. 1L shows a smartphone and tablet computer compatible user interface method that includes a topics digression mapping tool;
  • FIG. 1M shows a smartphone and tablet computer compatible user interface method that includes a social dynamics mapping tool;
  • FIG. 1N shows how the layout and content of each floor in a virtual multi-storied building can be re-organized as the user desires (e.g., for a “Help Grandma Today” day);
  • FIG. 2 is a perspective block diagram of a user environment that includes a portable palmtop microcomputer and/or intelligent cellphone (smartphone) or tablet computer which is structured for electromagnetic linking (e.g., electronically and/or optically linking) with a networking environment that includes a Social-Topical Adaptive Networking (STAN_3) system where, in accordance with one aspect of the present disclosure, the STAN_3 system includes means for automatically presenting through the mobile user interface, individual or group transaction offerings based on user context and on usages of the STAN_3 system;
  • FIGS. 3A-3B illustrate automated systems for passing user click or user tap or other user inputting streams and/or other energetic and contemporary focusing activities of a user through an intermediary server (e.g., webpage downloading server) to the STAN_3 system for thereby having the STAN_3 system return topic-related information for optional downloading to the user of the intermediary server;
  • FIG. 3C provides a flow chart of machine-implemented method that can be used in the system of FIG. 3A;
  • FIG. 3D provides a data flow schematic for explaining how individualized CFi's are automatically converted into normalized and/or categorized CFi's and thereafter mapped by the system to corresponding subregions or nodes within various data-organizing spaces (cognitions coding-for or symbolizing-of spaces) of the system (e.g., topic space, context space, etc.) so that topic-relevant and/or context sensitive results can be produced for or on behalf of a monitored user;
  • FIG. 3E provides a data structure schematic for explaining how cross links can be provided as between different data organizing spaces of the system, including for example, as between the recorded and adaptively updated topic space (Ts) of the system and a keywords organizing space, a URL's organizing space, a meta-tags organizing space and hybrid organizing spaces which cross organize data objects (e.g., nodes) of two or more different, data organizing spaces and wherein at least one data organizing space has an adaptively updateable, expressions, codings, or other symbols clustering layer;
  • FIGS. 3F-3I respectively show data structures of data object primitives useable for example in a music-nodes data organizing space, a sounds-nodes data organizing space, a voice nodes data organizing space, and a linguistics nodes data organizing space;
  • FIG. 3J shows data structures of data object primitives useable in a context nodes data organizing space;
  • FIG. 3K shows data structures usable in defining nodes being focused-upon and/or space subregions (e.g., TSR's) being focused-upon within a predetermined time duration by an identified social entity;
  • FIG. 3L shows an example of a data structure such as that of FIG. 3K logically linking to a hybrid operator node in a hybrid space formed by the intersection of a music space, a context space and a portion of topic space;
  • FIGS. 3M-3P respectively show data structures of data object primitives useable for example in an images nodes data organizing space, a body-parts/gestures nodes data organizing space, a biological states organizing space, and a chemical states organizing space;
  • FIG. 3Q shows an example of a data structure that may be used to define an operator node;
  • FIG. 3R illustrates in a perspective schematic format how child and co-sibling nodes (CSiN's) may be organized within a branch space owned by a parent node (such as a parent topic node of PaTN) and how personalized codings of different users in corresponding individualized contexts progress to become collective (communal) codings and collectively usable resources within, or linked to by, the CSiN's organized within the perspective-wise illustrated branch space;
  • FIG. 3S illustrates in a perspective schematic format how topic-less, catch-all nodes and/or topic-less, catch-all chat rooms (or other forum participation sessions) can respectively migrate to become topic-affiliated nodes placed in a branch space of a hierarchical topics tree and to become topic-affiliated chat rooms (or other forum participation sessions) that are strongly or weakly tethered to such topic-affiliated nodes;
  • FIG. 3Ta and FIG. 3Tb show an example of a data structure that may be used for representing a corresponding topic node in the system of FIGS. 3R-3S;
  • FIG. 3U shows an example of a data structure that may be used for implementing a generic CFi's collecting (clustering) node in the system of FIGS. 3R-3S;
  • FIG. 3V shows an example of a data structure that may be used for implementing a species of a CFi's collecting node specific to textual types of CFi's;
  • FIG. 3W shows an example of a data structure that may be used for implementing a textual expression primitive object;
  • FIG. 3X illustrates a system for locating equivalent and near-equivalent (same or similar) nodes within a corresponding data organizing space;
  • FIG. 3Y illustrates a system that automatically scans through a hybrid context-plus-other space (e.g., context-plus-keyword expressions space) in order to identify context appropriate topic nodes and/or subregions that score highest for correspondence with CFi's received under the assumed context;
  • FIG. 4A is a block diagram of a networked system that includes network interconnected mechanisms for maintaining one or more Social/Persona Entities Interrelation Spaces (SPEIS), for maintaining one or more kinds of topic spaces (TS's, including a hybrid context plus topic space) and for supplying group offers to users of a Social-Topical Adaptive Networking system (STAN3) that supports the SPEIS and TS's as well as other relationships (e.g., L2U/T/C, which here denotes location to user(s), topic node(s), content(s) and other such data entities);
  • FIG. 4B shows a combination of flow chart and popped up screen shots illustrating how user-to-user associations (U2U) from external platforms can be acquired by (imported into) the STAN_3 system;
  • FIG. 4C shows a combination of a data structure and examples of user-to-user associations (U2U) for explaining an embodiment of FIG. 4B in greater detail;
  • FIG. 4D is a perspective type of schematic view showing mappings between different kinds of spaces and also showing how different user-to-user associations (U2U) may be utilized by a STAN_3 server that determines, for example, “What topics are my friends now focusing on and what patterns of journeys have they recently taken through one or more spaces supported by the STAN_3 system?”;
  • FIG. 4E illustrates how spatial clusterings of points, nodes or subregions in a given Cognitive Attention Receiving Space (CARS) may be displayed and how significant ‘touchings’ by identified (e.g., demographically filtered) social entities in corresponding 2D or higher dimensioned maps of data organizing spaces (e.g., topic space) can also be identified and displayed;
  • FIG. 4F illustrates how geographic clusterings of on-topic chat or other forum participation sessions can be displayed and how availability of nearby promotional or other resources can also be displayed;
  • FIG. 5A illustrates a profiling data structure (PHA_FUEL) usable for determining habits, routines, and likes and dislikes of STAN users;
  • FIG. 5B illustrates another profiling data structure (PSDIP) usable for determining time and context dependent social dynamic traits of STAN users;
  • FIG. 5C is a block diagram of a social dynamics aware system that automatically populates chat or other forum participation opportunity spaces in an assembly line fashion with various types of social entities based on predetermined or variably adaptive social dynamic recipes; and
  • FIG. 6 is a flow chart indicating how an offering recipients-space may be populated by identities of persons who are likely to accept a corresponding offered transaction where the populating or depopulating of the offering recipients-space may be a function of usage by the targeted offerees of the STAN_3 system.
  • MORE DETAILED DESCRIPTION
  • Some of the detailed description found immediately below is substantially repetitive of detailed description of a ‘FIG. 1A’ found in the here-incorporated U.S. Ser. No. 12/854,082 application (STAN_2) and thus readers familiar with the details of the STAN_2 disclosure may elect to skim through to a part further below that begins to detail a tablet computer 100 illustrated by FIG. 1A of the present disclosure. FIG. 4A of the present disclosure corresponds to, but is not completely the same as the ‘FIG. 1A’ provided in the here-incorporated U.S. Ser. No. 12/854,082 application (STAN_2).
  • Referring to FIG. 4A of the present disclosure, shown is a block diagram of an electromagnetically inter-linked (e.g., electronically and/or optically linked, this optionally including wirelessly linked) networking environment 400 that includes a Social-Topical Adaptive Networking (STAN_3) sub-system 410 configured in accordance with the present disclosure. The encompassing environment 400 shown in FIG. 4A includes other sub-network systems (e.g., Non-STAN subnets 441, 442, etc., generally denoted herein as 44X). Although the electromagnetically inter-linked networking environment 400 will be often described as one using “the Internet” 401 for providing communications between, and data processing support for persons or other social entities and/or providing communications therebetween as well, and data processing support for, respective communication and data processing devices thereof, the networking environment 400 is not limited to just using “the Internet” and may include alternative or additional forms of communicative interlinkings. The Internet 401 is just one example of a panoply of communications-supporting and data processing supporting resources that may be used by the STAN_3 system 410. Other examples include, but are not limited to, telephone systems such as cellular telephony systems (e.g., 3G, 4G, etc.), including those wherein users or their devices can exchange text, images (including video, moving images or series of images) or other messages with one another as well as voice messages. More generically, the present disclosure contemplates various means by way of which individualized, physical codings by a first user that are representative of probable mental cognitions of that first user may be communicated directly or indirectly to one or more other users. (An example of an individualized, physical coding might be the text string, “The Golden Great” by way of which string, a given individual user might refer to American football player, Joseph “Joe” Montana, Jr. whereas others may refer to him as “Joe Cool” or “Golden Joe” or otherwise. The significance of individualized, physical codings versus collectively recognized codings will be explained later below. A text string is merely one of different ways in which coded symbols can be used to represent individualized mental cognitions of respective system users. Other examples include sign language, body language, music, and so on.) Yet other examples of communicative means by way of which user codings can be communicated include cable television systems, satellite dish systems, near field networking systems (optical and/or radio based), and so on; any of which can act as conduits and/or routers (e.g., uni-cast, multi-cast broadcast) for not only digitized or analog TV signals but also for various other digitized or analog signals, including those that convey codings representative of individualized and/or collectively recognized codings. Yet other examples of such communicative means include wide area wireless broadcast systems and local area wireless broadcast, uni-cast, and/or multi-cast systems. (Incidental note: In this disclosure, the terms STAN_3, STAN#3, STAN-3, STAN3, or the like are used interchangeably to represent the third generation Social-Topical Adaptive Networking (STAN) system. STAN_1, STAN_2 similarly represent the respective first and second generations.)
  • The resources of the schematically illustrated environment 400 may be used to define so-called, user-to-user association codings (U2U) including for example, so-called “friendship spaces” (which spaces are a subset of the broader concept of Social/Persona Entities Interrelation Spaces (SPEIS) as disclosed herein and as represented by data signals stored in a SPEIS database area 411 of the STAN_3 system portion 410 of FIG. 4A. Examples of friendship spaces may include a graphed representation (as digitally encoded) of real persons whom a first user (e.g., 431) friends and/or de-friends over a predetermined time period when that first user utilizes an available version of the FaceBook™ platform 441. See also, briefly; FIG. 4C. Another friendship space may be defined by a graphed representation (as digitally encoded) of real persons whom the user 431 friends and/or de-friends over a predetermined time period when that first user utilizes an available version of the MySpace™ platform 442. Other Social/Personal Interrelations may be defined by the first user 431 utilizing other available social networking (SN) systems such as LinkedIn™ 444, Twitter™ and so on. As those skilled in the art of computer-facilitated social networking (SN) will be aware, the well known FaceBook™ platform 441 and MySpace™ platform 442 are relatively pioneering implementations of social media approaches to exploiting user-to-user associations (U2U) for providing network users with socially meaningful experiences while using computer-facilitated and electronic communication facilitated resources. However there is much room for improvement over the pioneering implementations and numerous such improvements may be found at least in the present disclosure if not also in the earlier the disclosures of the here incorporated U.S. Ser. No. 12/369,274 (filed Feb. 11, 2009) and U.S. Ser. No. 12/854,082 (filed Aug. 10, 2010).
  • The present disclosure will show how various matrix-like cross-correlations between one or more SPEIS 411 (e.g., friendship relation spaces) and topic-to-topic associations (T2T, a.k.a. topic spaces) 413 and hybrid context associations (e.g., location to users to topic associations) 416 may be used to enhance online experiences of real person users (e.g., 431, 432) of the one or more of the sub-networks 410, 441, 442, . . . , 44X, etc. due to cross-correlating actions automatically taken by the STAN_3 sub-network system 410 of FIG. 4A.
  • Yet more detailed background descriptions on how Social-Topical Adaptive Networking (STAN) sub-systems may operate can be found in the above-cited and here incorporated U.S. application Ser. No. 12/369,274 and Ser. No. 12/854,082 and therefore as already mentioned, detailed repetitions of said incorporated-by-reference materials will not all be provided here. For sake of avoiding confusion between the drawings of Ser. No. 12/369,274 (STAN_1) and the figures of the present application, drawings of Ser. No. 12/369,274 will be identified by the prefix, “giF.” (which is “Fig.” written backwards) while figures of the present application will be identified by the normal figure prefix, “Fig.”. It is to be noted that, if there are conflicts as between any two or more of the two earlier filed and here incorporated applications and this application, the later filed disclosure controls as to conflicting teachings.
  • In brief, giF. 1A of the here incorporated '274 application shows how topics that are currently being focused-upon by (not to be confused with sub-portions of content being currently ‘focused upon’ by) individual online participants may be automatically determined based on detection of certain content sub-portions being currently and emotively ‘focused upon’ by the respective online participants and based upon pre-developed profiles of the respective users (e.g., registered and logged-in users of the STAN_1 system). (Incidentally, in the here disclosed STAN_3 system, the notion is included of determining what group offers a user is likely to currently welcome or not welcome based on a variety of factors including habit histories, trending histories, detected context and so on.)
  • Further in brief, giF. 1B of the incorporated '274 application shows a data structure of a first stored chat co-compatibility profile that can change with changes of user persona (e.g., change of mood); giF. 1C shows a data structure of a stored topic co-compatibility profile that can also change with change of user persona (e.g., change of mood, change of surroundings); and giF. 1E shows a data structure of a stored personal emotive expression profile of a given user, whereby biometrically detected facial or other biotic expressions of the profiled user may be used to deduce emotional involvement with on-screen content and thus degree of emotional involvement with focused upon content. One embodiment of the STAN_1 system disclosed in the here incorporated '274 application uses uploaded CFi (current focus indicator) packets to automatically determine what topic or topics are most likely ones that each user is currently thinking about based on the content that is being currently focused upon with above-threshold intensity. The determined topic is logically linked by operations of the STAN_1 system to topic nodes (herein also referred to as topic centers or TC's) within a hierarchical parent-child tree represented by data stored in the STAN_1 system.
  • Yet further and in brief, giF. 2A of the incorporated '274 application shows a possible data structure of a stored CFi record while giF. 2B shows a possible data structure of an implied vote-indicating record (CVi) which may be automatically extracted from biometric information obtained from the user. The giF. 3B diagram shows an exemplary screen display wherein so-called chat opportunity invitations (herein referred to as in-STAN-Vitations™) are provided to the user based on the STAN_1 system's understanding of what topics are currently of prime interest to the user. The giF. 3C diagram shows how one embodiment of the STAN_1 system (of the '274 application) can automatically determine what topic or domain of topics might most likely be of current interest for a given user and then responsively can recommend, based on likelihood rankings, content (e.g., chat rooms) which are most likely to be on-topic for that user and compatible with the user's current status (e.g., level of expertise in the topic).
  • Moreover, in the here incorporated '274 application, giF. 4A shows a structure of a cloud computing system (e.g., a chunky grained cloud) that may be used to implement a STAN_1 system on a geographic region by geographic region basis. Importantly, each data center of giF. 4A has an automated Domains/Topics Lookup Service (DLUX) executing therein which receives up- or in-loaded CFi data packets (Current Focus indicating records) from users and combines these with user histories uploaded form the user's local machine and/or user histories already stored in the cloud to automatically determine probable topics of current interest then on the user's mind. In one embodiment the DLUX points to so-called topic nodes of a hierarchical topics tree. An exemplary data structure for such a topics tree is provided in giF. 4B which shows details of a stored and adaptively updated topic mapping data structure used by one embodiment of the STAN_1 system. Also each data center of giF. 4A further has one or more automated Domain-specific Matching Services (DsMS's) executing therein which are selected by the DLUX to further process the up- or in-loaded CFi data packets and match alike users to one another or to matching chat rooms and then presents the latter as scored chat opportunities. Also each data center of giF. 4A further has one or more automated Chat Rooms management Services (CRS) executing therein for managing chat rooms or the like operating under auspices of the STAN_1 system. Also each data center of giF. 4A further has an automated Trending Data Store service that keeps track of progression of respective users over time in different topic sectors and makes trend projections based thereon.
  • The here incorporated '274 application is extensive and has many other drawings as well as descriptions that will not all be briefed upon here but are nonetheless incorporated herein by reference. (Note again that where there are conflicts as between any two or more of the earlier filed and here incorporated applications and this application, the later filed disclosure controls as to conflicting teachings.)
  • Referring again to FIG. 4A of the present disclosure, in the illustrated environment 400 which includes a more advanced, third generation or STAN_3 system 410, a first real and living user 431 (also USER-A, also “Stan”) is shown to have access to a first data processing device 431 a (also CPU-1, where “CPU” does not limit the device to a centralized or single data processing engine, but rather is shorthand for denoting any single or multi-processing digital or mixed signals device capable of providing the commensurate functionality). The first user 431 may routinely log into and utilize the illustrated STAN_3 Social-Topical Adaptive Networking system 410 by causing CPU-1 to send a corresponding user identification package 431 u 1 (e.g., user name and user password data signals and optionally, user fingerprint and/or other biometric identification data) to a log-in interface portion 418 of the STAN_3 system 410. In response to validation of such log-in, the STAN_3 system 410 automatically fetches various profiles of the logged-in user (431, “Stan”) from a database (DB, 419) thereof for the purpose of determining the user's currently probable topics of prime interest and current focus-upon, moods, chat co-compatibilities and so forth. As will be explained in conjunction with FIG. 3D, user profiling may start with fail-safe default profiles (301 d) and then switch to more context appropriate, current profiles (301 p). In one embodiment, a same user (e.g., 431 of FIG. 4A) may have plural personal log-in pages, for example, one that allows him to log in as “Stan” and another which allows that same real life person user to log-in under the alter ego identity (persona) of say, “Stewart” if that user is in the mood to assume the “Stewart” persona at the moment rather than the “Stan” persona. If a user (e.g., 431) logs-in via interface 418 with a second alter ego identity (e.g., “Stewart”) rather than with a first alter ego identity (e.g., “Stan”), the STAN_3 Social-Topical Adaptive Networking system 410 automatically activates corresponding personal profile records (e.g., CpCCp's, DsCCp's, PEEP's, PHAFUEL's, PSDIP, etc.; where the latter two will be explained below) of the second alter ego identity (e.g., “Stewart”) rather than those of the first alter ego identity (e.g., “Stan”). Topics of current interest that the machine system determines as being currently focused-upon by the logged-in persona may be identified as being logically associated with specific nodes (herein also referred to as TC's or topic centers) on a topics domain-parent/child tree structure such as the one schematically indicated at 415 within the drawn symbol that represents the STAN_3 system 410 in FIG. 4A. A corresponding stored data structure that represents the tree structure in the earlier STAN_1 system (not shown) is illustratively represented by drawing number giF. 4B. (A more advanced data structure for topic nodes will be described in conjunction with FIGS. 3Ta and FIG. 3Tb of the present disclosure.) The topics defining tree 415 as well as user profiles of registered STAN_3 users may be stored in various parts of the STAN_3 maintained database (DB) 419 which latter entity could be part of a cloud computing system and/or partly implemented in the user's local equipment and/or in remotely-instantiated data processing equipment (e.g., CPU-1, CPU-2, etc.). The database (DB) 419 may be a centralized one, or one that is semi-redundantly distributed over different service centers of a geographically distributed cloud computing system. In the distributed cloud computing environment, if one service center becomes nonoperational or overwhelmed with service requests, another somewhat redundant (partially overlapping in terms of resources) service center can function as a backup (where yet more details are provided in the here incorporated STAN_1 patent application). The STAN_1 cloud computing system is of chunky granularity rather than being homogeneous in that local resources (cloud data centers) are more dedicated to servicing local STAN user than to seamlessly backing up geographically distant centers should the latter become overwhelmed or temporarily nonoperational.
  • As used herein, the term, “local data processing equipment” includes data processing equipment that is remote from the user but is nonetheless controllable by a local means available to the user. More specifically, the user (e.g., 431) may have a so-called net-computer (e.g., 431 a) in his local possession and in the form for example of a tablet computer (see also 100 of FIG. 1A) or in the form for example of a palmtop smart cellphone/computer (see also 199 of FIG. 2) where that networked-computer is operatively coupled by wireless or other means to a virtual computer or to a virtual desktop space instantiated in one or more servers on a connected to network (e.g., the Internet 401). In such cases the user 431 may access, through operations of the relatively less-fully equipped net-computer (e.g., tablet 100 of FIG. 1A or palmtop 199 of FIG. 2, or more generally CPU-1 of FIG. 4A), the greater computing and data storing resources (hardware and/or software) available in the instantiated server(s) of the supporting cloud or other networked super-system (e.g., a system of data processing machines cooperatively interconnected by one or more networks to form a cooperative larger machine system). As a result, the user 431 is made to feel as if he has a much more resourceful computer locally in his possession (more resourceful in terms of hardware and/or software and/or functionality, any of which are physical manifestations as those terms are used herein) even though that might not be true of the physically possessed hardware and/or software. For example, the user's locally possessed net-computer (e.g., 431 a in FIG. 4A, 100 in FIG. 1A) may not have a hard disk or a key pad but rather a touch-detecting display screen and/or other user interface means appropriate for the nature of the locally possessed net-computer (e.g., 100 in FIG. 1A) and the local context in which it is used (e.g., while driving a car and thus based more on voice-based and/or gesture-based user-to-machine interface rather than on a graphical user interface). However the server (or cloud) instantiated virtual machine or other automated physical process that services that net-computer can project itself as having an extremely large hard disk or other memory means and a versatile keyboard-like interface that appears with context variable keys by way of the user's touch-responsive display and/or otherwise interactive screen. Occasionally the term “downloading” will be used herein under the assumption that the user's personally controlled computer (e.g., 431 a) is receiving the downloaded content. However, in the case of a net-book or the like local computer, the term “downloaded” is to be understood as including the more general notion of in- or cross-loaded, wherein a virtual computer on the network (or in a cloud computing system) is inloaded (or cross-loaded) with the content rather than having that content being “downloaded” from the network to an actual local and complete computer (e.g., tablet 100 of FIG. 1A) that is in direct possession of the user.
  • Of course, certain resources such as the illustrated GPS-2 peripheral part of CPU-2 (in FIG. 4A, or imbedded GPS 106 and gyroscopic (107) peripherals of FIG. 1A) may not always be capable of being operatively mimicked with an in-net or in-cloud virtual counterpart; in which case it is understood that the locally-required resource (e.g., GPS, gyroscope, IR beam source 109, barcode scanner, RFID tag reader, wireless interrogator of local-nodes (e.g., for indoor location and assets determination), user-proximate microphone(s), etc.) is a physically local resource. On the other hand, cell phone triangulation technology, RFID (radio frequency based wireless identification) technology, image recognition technology (e.g., recognizing a landmark) and/or other technologies may be used to mimic the effect of having a GPS unit although one might not be directly locally present. It is to be understood that GPS or other such local measuring, interrogating, detecting or telemetry collecting means need not be directly embedded in a portable data processing device that is hand carried or worn by the user. A portable/mobile device of the user may temporarily inherit such functionality from nearby other devices. More specifically, if the user's portable/mobile device does not have a temperature measuring sensor embedded therein for measuring ambient air temperature but the portable/mobile device is respectively located adjacent to, or between one; two or more other devices that do have air temperature measuring means, the user's portable/mobile device may temporarily adopt the measurements made by the nearby one; two or more other devices and extrapolate and/or add an estimated error indication to the adopted measurement reading based on distance from the nearby measurement equipment and/or based on other factors such as local wind velocity. The same concept substantially applies to obtaining GPS-like location information. If the user's portable/mobile device is interposed between two or more GPS-equipped, and relatively close by, other devices that it can communicate with and the user's portable/mobile device can estimate distances between itself and the other devices, then the user's portable/mobile device may automatically determine its current location based on the adopted location measurements of the nearby other devices and on an extrapolation or estimate of where the user's portable/mobile device is located relative to those other devices. Similarly, the user's portable/mobile device may temporarily co-opt other detection or measurement functionalities that neighboring devices have but it itself does not directly possess such as, but not limited to, sound detection and/or measurement capabilities, biometric data detection and/or measurement capabilities, image capture and/or processing capabilities, odor and/or other chemical detection, measurement and/or analysis capabilities and so on.
  • It is to be understood that the CPU-1 device (431 a) used by first user 431 when interacting with (e.g., being tracked, monitored in real time by) the STAN_3 system 410 is not limited to a desktop computer having for example a “central” processing unit (CPU), but rather that many varieties of data processing devices having appropriate minimal intelligence capability are contemplated as being usable, including laptop computers, palmtop PDA's (e.g., 199 of FIG. 2), tablet computers (e.g., 100 of FIG. 1a ), other forms of net-computers, including 3rd generation or higher smartphones (e.g., an iPhone™, and Android™ phone), wearable computers, and so on. The CPU-1 device (431 a) used by first user 431 may have any number of different user interface (UI) and environment detecting devices included therein such as, but not limited to, one or more integrally incorporated webcams (one of which may be robotically aimed to focus on what off screen view the user appears to be looking at, e.g. 210 of FIG. 2), one or more integrally incorporated ear-piece and/or head-piece subsystems (e.g., Bluetooth™) interfacing devices (e.g., 201 b of FIG. 2), an integrally incorporated GPS (Global Positioning System) location identifier and/or other automatic location identifying means, integrally incorporated accelerometers (e.g., 107 of FIG. 1) and/or other such MEMs devices (micro-electromechanical devices), various biometric sensors (e.g., vascular pulse, respiration rate, tongue protrusion, in-mouth tongue actuations, eye blink rate, eye focus angle, pupil dilation and change of dilation and rate of dilation (while taking into consideration ambient light strength and changes), body odor, breath chemistry—e.g., as may be collected and analyzed by combination microphone and exhalation sampler 201 c of FIG. 2) that are operatively coupleable to the user 431 and so on. As those skilled in the art will appreciate from the here incorporated STAN_1 and STAN_2 disclosures, automated location determining devices such as integrally incorporated GPS and/or audio pickups and/or odor pickups may be used to determine user surroundings (e.g., at work versus at home, alone or in noisy party, near odor emitting items or not) and to thus infer from this sensing of environment and user state within that environment, the more probable current user persona (e.g., mood, frame of mind, etc.). One or more (e.g., stereoscopic) first sensors (e.g., 106, 109 of FIG. 1A) may be provided in one embodiment for automatically determining what specific off-screen or on-screen object(s) the user is currently looking at; and if off-screen, a robotically aimmable further sensor (e.g., webcam 210) may be automatically trained onto the off-screen view (e.g., 198 in FIG. 2) in order to identify it, categorize it and optionally provide a virtually-augmented presentation of that off-screen specific object (198). In one embodiment, an automated image categorizing tool such as GoogleGoggles™ or IQ_Engine™ (e.g., www.iqengines.com) may be used to automatically categorize imagery or objects (including real world objects) that the user appears to be focusing upon. The categorization data of the automatically categorized image/objects may then be used as an additional “encoding” and hint presentations for assisting the STAN_3 system 410 in determining what topic or finite set (e.g., top 5) of topics the user (e.g., 431) currently most probably has in focus within his or her mind given the detected or presumable context of the user.
  • It is within the contemplation of the present disclosure that alternatively or in addition to having an imaging device near the user and using an automated image/object categorizing tool such as GoogleGoggles™, IQ_Engine™, etc., other encoding detecting devices and automated categorizing tools may be deployed such as, but not limited to, sound detecting, analyzing and categorizing tools; non-visible light band detecting, analyzing, recognizing and categorizing tools (e.g., IR band scanning and detecting tools); near field apparatus identifying communication tools, ambient chemistry and temperature detecting, analyzing and categorizing tools (e.g., What human olfactorable and/or unsmellable vapors, gases are in the air surrounding the user and at what changing concentration levels?); velocity and/or acceleration detecting, analyzing and categorizing tools (e.g., Is the user in a moving vehicle and if so, heading in what direction at what speed or acceleration?); gravitational orientation and/or motion detecting, analyzing and categorizing tools (e.g., Is the user titling, shaking or otherwise manipulating his palmtop device?); and virtually-surrounding or physically-surrounding other people detecting, analyzing and categorizing tools (e.g., Is the user in virtual and/or physical contact or proximity with other personas, and if so what are their current attributes?).
  • Each user (e.g., 431, 432) may project a respective one of different personas and assumed roles (e.g., “at work” versus “at play” persona, where the selected persona may then imply a selected context) based on the specific environment (including proximate presence of other people virtually or physically) that the user finds him or herself in. For example, there may be an at-the-office or at-work-site persona that is different from an at-home or an on-vacation persona and these may have respectively different habits, routines and/or personal expression preferences due to corresponding contexts. (See also briefly the context identifying signal 316 o of FIG. 3D which will detailed below. Most likely context may be identified in part based on user selected persona.) More specifically, one of the many selectable personas that the first user 431 may have is one that predominates in a specific real and/or virtual environment 431 e 2 (e.g., as geographically detected by integral GPS-2 device of CPU-2 and/or as socially detected by a connected/nearby others detector). When user 431 is in this environmental context (431 e 2), that first user 431 may choose to identify him or herself with (or have his CPU device automatically choose for him/her) a different user identification (UAID-2, also 431 u 2) than the one utilized (UAID-1, also 431 u 1) when typically interacting in real time with the STAN_3 system 410. A variety of automated tools may be used to detect, analyze and categorize user environment (e.g., place, time, calendar date, velocity, acceleration, surroundings—physically or virtually nearby objects and/or nearby people and their respectively assumed roles, etc.). These may include but are not limited to, webcams, IR Beam (IRB) face scanners, GPS locators, electronic time keeper, MEMs, chemical sniffers, etc.
  • When operating under this alternate persona (431 u 2), the first user 431 may choose (or pre-elect) to not be wholly or partially monitored in real time by the STAN_3 system (e.g., through its CFi, CVi or other such monitoring and reporting mechanisms) or to otherwise not be generally interacting with the STAN_3 system 410. Instead, the user 431 may elect to log into a different kind of social networking (SN) system or other content providing system (e.g., 441, . . . , 448, 460) and to fly, so-to-speak, STAN-free inside that external platform 441—etc. While so interacting in a free-of-STAN mode with the alternate social networking (SN) system (e.g., FaceBook™, MySpace™, LinkedIn™, YouTube™, GoogleWave™, ClearSpring™, etc.), the user may develop various types of user-to-user associations (U2U, see block 411) unique to that outside-of-STAN platform. More specifically, the user 431 may develop a historically changing record of newly-made “friends”/“frenemys” on the FaceBook™ platform 441 such as: recently de-friended persons, recently allowed-behind the private wall friends (because they are more trusted) and so on. The user 431 may develop a historically changing record of newly-made live-video chat buddies on the FaceBook™ platform 441. The user 431 may develop a historically changing record of newly-made 1st degree “contacts” on the LinkedIn™ platform 444, newly joined groups and so on. The user 431 may then wish to import some of these outside-of-STAN-formed user-to-user associations (U2U) to the STAN_3 system 410 for the purpose of keeping track of what topics in one or more topic spaces 413 (or other nodes in other spaces) the respective friends, non-friends, contacts, buddies etc. are currently focusing-upon in either a direct ‘touching’ manner or through indirect heat ‘touching’. Importation of user-to-user association (U2U) records into the STAN_3 system 410 may be done under joint import/export agreements as between various platform operators or via user transfer of records from an external platform (e.g., 441) to the STAN_3 system 410.
  • Referring next, and on a brief basis to FIG. 1A (more details are provided later below), shown here is a display screen 111 of a corresponding tablet computer 100 on whose touch-sensitive screen 111 there are displayed a variety of machine-instantiated virtual objects. Although the illustrated example has but one touch-sensitive display screen 111 on which all is displayed, it is within the contemplation of the present disclosure for the computer 100 (a.k.a. first data processing device usable by a corresponding first user) to be operatively coupleable by wireless and/or wired means to one or more auxiliary displays and/or auxiliary user-to-machine interface means (e.g., a large screen TV with built in gesture recognition and for which the computer 100 appears to act as a remote control). Additionally, while not shown in FIG. 1A, it will become clearer below that the illustrated computer 100 is operatively couplable to a point(s)-of-attention modeling system (e.g., in-cloud STAN server(s)) that has access to signals (e.g., CFi's) representing attention indicative activities of the first user (at what is the user focusing his/her attentions upon?). Moreover, it is to be understood that the visual information outputting function of display screen 111 is but one way of presenting (outputting) information to the user and that it is within the contemplation of the present disclosure to present (output) information to the user in additional or alternative ways including by way of sound (e.g., voice and/or tones and/or musical scores) and/or haptic means (e.g., variable Braille dots for the blind and/or vibrating or force producing devices that communicate with the user by means of different vibrations and/or differently directed force applications).
  • In the exemplary illustration, the displayed objects of screen 111 are clustered into major screen regions including a major left column region 101 (a.k.a. first axis), a topside and hideable tray region 102 (a second axis), a major right column region 103 (a third axis) and a bottomside and hideable tray region 104 (a fourth axis). The corners at which the column and row regions 101-104 meet also have noteworthy objects. The bottom right corner (first axes crossing—of axes 103 and 104) contains an elevator tool 113 which can be used to travel to different virtual floors of multi-storied virtual structure (e.g., building). Such a multi-storied virtual structure may be used to define a virtual space within which the user virtually travels to get to virtual rooms or virtual other areas having respective combinations of invitation presenting trays and/or such tools. (See also briefly, FIG. 1N.) The upper left corner (second axes crossing) of screen 111 contains an elevator floor indicating tool 113 a which indicates which virtual floor is currently being visited (e.g., the floor that automatically serves up in area 102 a set of opportunity serving plates labeled as the Me and My Friends and Family Top Topics Now serving plates). In one embodiment, the floor indicating tool 113 a may be used to change the currently displayed floor (for example to rapidly jump to the User-Customized Help Grandma floor of FIG. 1N). The bottom left corner (third axes crossing) contains a settings tool 114. The top right corner (fourth axes crossing—of axes 102 and 103) is reserved for a status indicating tool 112 that tells the user at least whether monitoring by the STAN_3 system is currently active or not, and if so, optionally what parts of his/her screen(s) and/or activities are being monitored (e.g., full screen and all activities versus just one data processing device, just one window or pane therein and/or just certain filter-defined activities). The center of the display screen 111 is reserved for centrally focused-upon content that the user will usually be focusing-upon (e.g., window 117, not to scale, and showing in subportions (e.g., 117 a) thereof content related to an eBook Discussion Group that the user belongs to). It is to be understood that the described axes (102-104) and axes crossings can be rearranged into different configurations.
  • Among the objects displayed in the left column area 101 are urgency valued or importance valued ones that collectively define a sorted list of social entities or groups thereof, such as “My Family” 101 b (valued in this example as second most important/relevant after the “Me” entity 101 a) and/or “My Friends” 101 c (valued in this example as third in terms of importance/urgency after “Me” and after “My Family”) where the represented social entities and their positionings along the list are pre-specified by the current user of the device 100 or accepted as such by the user after having been automatically recommended by the system.
  • The topmost social entity along the left-side vertical column 101 (the sorted list of now-important/relevant social entities) is specially denoted as the current King-of-the-Hill Social Entity (e.g., KoH=“Me” 101 a) while the person or group representing objects disposed below the current King-of-the-Hill (101 a) are understood to be subservient to or secondary relative to the KOH object 101 a in that certain categories of attributes painted-on or attached to those subservient objects (101 b, 101 c, etc.) are inherited from the KOH object 101 a and mirrored onto the subservient objects or attachments thereof. (The KOH object may alternatively be called the Pharaoh of the Pyramids for reasons soon to become apparent.) Each of the displayed first items (e.g., social entity representing items 101 a-101 d) may include one or both a correspondingly displayed label (e.g., “Me”) and a correspondingly displayed icon (e.g., up-facing disc). Alternatively or additionally, the presentation of the first items may come by way of voice presentation. Different ones of the presented first items may have unique musical tones and/or color tones associated with them, where in the case of the display being used, the corresponding musical tones and/or color tones are presented as the user hovers a cursor or the like over the item.
  • In terms of more specifics, and referring also to FIG. 1B, adjacent to the KOH object 101 a of the first vertical axis 101 of FIG. 1A there may be provided along a second vertical axis 101 r, a corresponding status reporting pyramid 101 ra belonging to the KOH object 101 a. Displayed on a first face of that status-reporting pyramid 101 ra are a set of painted histogram bars denoted as Heat of My Top 5 Now Topics (see 101 w′ of FIG. 1B). It is understood that each such histogram bar corresponds to a respective one of a Top 5 Now (being-now-focused-upon) Topics of the King-of-the-Hill Social Entity (e.g., KoH=“Me” 101 a) and it reports on a “heat” attribute (e.g., attentive energies) cast by the row's social entity with regard to that topic. The mere presence of the histogram bar indicates that attention is being cast by the row's social entity with regard to the bar's associated topic. The height of the bar (and/or another attribute thereof) indicates how much attention. The amount of attention can have numerous sub-attributes such as emotional attention, deep neo-cortical thinking attention, physical activity attention (i.e., keeping one's eyes trained on content directed to the specific topic) and so on.
  • From usage of the system, it becomes understood to users of the system that the associated topic of each such histogram bar on the attached status pyramid (e.g., 101 rb in FIG. 1A) of a subservient social entity (101 b, 101 c, etc.) corresponds in category mirroring fashion to a respective one of the Top 5 Now (being-focused-upon) Topics of the KOH. In other words, it is not necessarily a top-now-topic of the subservient social entity (e.g., 101 b), but rather it is a top-now topic of the King-of-the-Hill (KOH) Social Entity 101 a.
  • Therefore, if the social entity identified as “Me” by the top item of column 101 is King-of-the-Hill and the Top 5 Now Topics of “Me” are represented by bars on a face of the KOH's adjacent reporting pyramid 101 ra, the same Top 5 Now Topics of “Me” will be represented by (mirrored by) respective locations of bars on a corresponding face of subservient reporting pyramids (e.g., 101 rb). Accordingly, with one quick look, the user can see what Top 5 Now Topics of “Me” (if “Me” is the KOH) are also being focused-upon (if at all), and if so with what “heat” (emotional and/or otherwise) by associated other social entities (e.g., by “My Family” 101 b, by “My Friends” 101 c and so on).
  • The designation of who is currently the King-of-the-Hill Social Entity (e.g., KoH=“Me” 101 a) can be indicated by means other than or in addition to displaying the KOH entity object 101 a at the top of first vertical column 101. For example, KOH status may be indicated by displaying a virtual crown (not shown) on the entity representing object (e.g., 101 a) who is King and/or coloring or blinking the KOH entity representing object 101 a differently and so on. Placement at the top of the stack 101 is used here as a convenient way of explaining the KOH concept and also explaining the concept of a sorted array of social entities whose positional placement is based on the user's current valuation of them (e.g., who is now most important, who is most urgent to focus-upon, etc.). The user's data processing device 100 may include a ‘Help’ function (activated by right clicking to activate, or otherwise activating a context sensitive menu 111 a) that provides detailed explanation of the KOH function and the sorted array function (e.g., is it sorting its items 101 a-10 d based on urgency, based on importance or based on some other metrics?). Although for sake of an easiest to understand example, the “Me” disc 101 a is disposed in the KOH position, the representative disc of any other social entity (individual or group), say, “My Others” 101 d can instead be designated as the KOH item, placed on top, and then the Top 5 Now Topics of the group called “My Others” (101 d) will be mirrored onto the status reporting pyramids of the remaining social entity objects (including “Me”) of column 101. The relative sorting of the secondary social entities relative to the new KoH entity will be based on what the user of the system (not the KoH) thinks it should be. However, in one embodiment, the user may ask the system to sort the secondary social entities according to the way the KoH sorts those items on his computer.
  • Although FIG. 1A shows the left vertical column 101 (first vertical array) as providing a sorted array of disc objects 101 a-101 d representing corresponding social entities, where these are sorted according to different valuation criteria such as importance of relation or urgency of relation or priority (in terms for example of needing attention by the user), it is within the contemplation of the present disclosure to have the first vertical column 101 provide a sorted array of corresponding first items representing other things; for example things associated with one or more prespecified social entities; and more specifically, projects or other to-do items associated with one or more social entities. Yet more specifically, the chosen social entity might be “Me” and then the first vertical column 101 may provides a sorted array of first items (e.g., disc objects) representing work projects attributed to the “Me” entity (e.g., “My Project#1”, “My Project#2”, etc.—not shown) where the array is sorted according to urgency, priority, current financial risk projections or other valuations regarding relative importance and timing priorities. As another example, the sorted array of disc-like objects in the first vertical column 101 might respectively represent, in top down order of display, first the most urgent work project assigned to the “Me” entity, then the most urgent work project assigned to the “My Boss” entity, and then the most urgent work project associated with the “His Boss” entity. At the same time, the upper serving tray 102 (first horizontal axis) may serve up chat or other forum participation opportunities corresponding to keywords, URL's etc. associated with the respective projects, where any of the served up participation opportunities can be immediately seized upon by the user double clicking or otherwise opening up the opportunity-representing icon to thereby immediately display the underlying chat or other forum participation session.
  • According to yet another variation (not shown), the arrayed first items 101 a-101 d of the first vertical column 101 may respectively represent different versions of the “Me” entity; as such for example “Me When at Home” (a first context); “Me When at Work” (a second context); “Me While on the Road” (a third context); “Me While Logged in as Persona#1 on social networking Platform#2” (a fourth context) and so on.
  • In one embodiment, the sorted first array of disc objects 101 a-101 d and what they represent are automatically chosen or automatically offered to be chosen based on an automatically detected current context of the device user. For example, if the user of data processing device 100 is detected to be at his usual work place (and more specifically, in his usual work area and at his usual work station), then the sorted first array of disc objects 101 a-101 d might respectively represent work-related personas or work-related projects. In an alternate or same embodiment, the sorted array of disc objects 101 a-101 d and what they represent can be automatically chosen or automatically offered to be chosen based on the current Layer-Vator™ floor number (as indicated by tool 113 a). In an alternate or same embodiment, the sorted array of disc objects 101 a-101 d and what they represent can be automatically chosen or automatically offered to be chosen based on current time of day, day of week, date within year and/or current geographic location or compass heading of the user or his vehicle and/or scheduled events in the user's computerized calendar files.
  • Returning to the specific example of the items actually shown to be arrayed in first vertical column 101 of FIG. 1A and looking here at yet more specific examples of what such social entity objects (e.g., 101 a-101 d) might represent, the displayed circular disc denoted as the “My Friends”-representing object 101 c can represent a filtered subset of a current user's FaceBook™ friends, where identification records of those friends have been imported from the corresponding external platform (e.g., 441 of FIG. 4A) and then optionally further filtered according to a user-chosen filtering algorithm (e.g., just include all my trusted, behind the wall friends of the past week who haven't been de-friended by me in the past 2 weeks). Additionally, the “My Friends” representing object 101 c is not limited to picking friends from just one source (e.g., the FaceBook™ platform 441 whose counterpart is displayed as platform representing object 103 b at the far right side 103 of the screen 111). A user can slice and dice and mix individual personas or other social entities (standard groups or customized groups) from different sources; for example by setting “My Friends” equal to My Three Thursday Night Bowling Buddies plus my trusted, behind the wall FaceBook™ friends of the past week. An EDIT function provided by an on-screen menu 111 a includes tools (not shown) for allowing the user to select who or what social entity or entities will be members of each user-defined, social entity-representing or entities-representing object (e.g., discs 101 a-101 d). The “Me” representing object 101 a does not, for example, have to represent only the device user alone (although such representation is easier to comprehend) and it may be modified by the EDIT function so that, for example, “Me” represents a current online persona of the user's plus one or more identified significant others (SO's, e.g., a spouse) if so desired. Additional user preference tools (114) may be employed for changing how King-of-the-Hill (KOH) status is indicated (if at all) and whether such designation requires that the KOH representing object (e.g., the “Me” object 101 a) be placed at the top of the stack 101. In one embodiment, if none of the displayed social entity representing objects 101 a-101 d in the left vertical column 101 is designated as KOH, then topic mirroring is turned off and each status-reporting pyramid 101 ra-101 rd (or pyramids column 101 r) reports a “heat” status for the respective Top 5 Now Topics of that respective social entity. In other words, reporting pyramid 101 rd then reports the “heat” status for the Top 5 Now Topics of the social group entity identified as “My Others” and represented by object 101 d rather than showing “heat” cast by “My Others” on the Top 5 Now Topics of the KOH (the King-of the-Hill). The concept of “cast heat”, incidentally, will be explained in more detail below (see FIGS. 1E and 1F). For now, it may be thought of as indicating how intensely in terms of emotions or otherwise, the corresponding social entity or social group (e.g., “My Others” 101 d) is currently focusing-upon or paying attention to each of the identified topics even if the corresponding social entity is not consciously aware of his or her paying prime attention to the topic per se.
  • As may be appreciated, the current “heat” reporting function of the status reporting objects in column 101 r (they do not have to be pyramids) provides a convenient summarizing view, for example, for: (1) identifying relevant social-associates of the user (e.g., “Me” 101 a), (2) for indicating how those socially-associated entities 101 b-101 d are grouped and/or filtered and/or prioritized relative to one another (e.g., “My Friends” equals only all my trusted, behind the wall friends of the past week plus my three bowling buddies); (3) for tracking some of their current activities (if not blocked by privacy settings) in an adjacent column 101 r by indicating cross-correlation with the KOH's Top 5 Now Topics or by indicating “heat” cast by each on their own Top 5 Now Topics if there is no designated KOH.
  • Although in the illustrated example, the subsidiary adjacent column 101 r (social radars column) indicates what top-5 topics of the entity “Me” (101 a) are also being focused-upon in recent time periods (e.g., now and 15 minutes ago, see faces 101 t and 101 x of magnified pyramid 101 rb in FIG. 1A) and to what extent (amount of “heat”) by associated friends or family or other social entities (101 b-101 d), various other kinds of status reports may be provided at the user's discretion. For example, the user may wish to see what the top N topics were (where N does not have to be 5) last week, or last month of the respective social entities. By way of another example, the user may wish to see what top N URL's and/or keywords were ‘touched’ upon by his relevant social entities in the last 6, 12, 24, 48 or other number of hours. (“Keywords” are generally understood here to mean the small number of words used for submitting to a popular search engine tool for thereby homing in on and identifying content best described by such keywords. “Content”, on the other hand, may refer to a much broader class of presentable information where the mere presentation of such information does not mean that a user is focusing-upon all of it or even a small sub-portion of it. “Content” is not to be conflated with “Topic”. A presented collection of content could have many possible topics associated with it.)
  • Focused-upon “topics” or topic regions are merely one type of trackable thing or item represented in a corresponding Cognitive Attention Receiving Space (a.k.a. “CARS”) and upon which users may focus their attentions upon. As used herein, trackable targets of cognition (codings or symbols representing underlying and different kinds of cognitions) have, or have newly created for them, respective data objects uniquely disposed in a corresponding data-objects organizing space, where data signals representing the data objects are stored within the system. One of the ways to uniquely dispose the data objects is to assign them to unique points, nodes or subregions of the corresponding Cognitive Attention Receiving Space (e.g., Topic Space) where such points, nodes, or subregions may be reported on (as long as the to-be-tracked users have given permission that allows for such monitoring, tracking and/or reporting). As will become clearer, the focused-upon top-5 topics, as exemplified by pyramid face 101 t in FIG. 1A, are further represented by topic nodes and/or topic regions defined in a corresponding one or more of topic space defining database records (e.g., area 413 of FIG. 4A) maintained and/or tracked by the STAN_3 system 410. A more rigorous discussion of topic nodes, topic regions, pure and hybrid topic spaces will be provided in conjunction with FIGS. 3D-3E, 3R-3Ta and 3Tb and others as the present disclosure unfolds below.
  • In the simplified example of introductory FIG. 1A, the user of tablet computer 100 (FIG. 1A) has selected a selectable persona of himself (e.g., 431 u 1) to be used as the head entity or “mayor” (or “King-'o-Hill”, KoH, or Pharaoh) of the social entities column 101. The user has selected a selectable set of attributes to be reported on by the status reporting objects (e.g., pyramids) of reporting column 101 r where the selected set of attributes correspond to a topic space usage attributes such as: (a) the current top-5 focused-upon topics of mine, (b) the older top N topics of mine, (c) the recently most “hot” (heated up) top N′ topics of mine, and so on. The user of tablet computer 100 (FIG. 1A) has elected to have one or more such attributes reported on in substantially real time in the subsidiary and radar-like tracking column 101 r disposed adjacent to the social entities listing column 101. The user has also selected an iconic method (e.g., pyramids) by way of which the selected usage attributes will be displayed. It will be seen in FIG. 1D that a rotating pyramid is not the only way.
  • It is to be understood here that the illustrated screen layout of introductory FIG. 1A and the displayed contents of FIG. 1A are merely exemplary and non-limiting. The same tablet computer 100 may display other Layer-Vator (113) reachable floors or layers that have completely different layouts and contain different on-screen objects. This will be clearer when the “Help Grandma” floor is later described as an example in conjunction with FIG. 1N. Moreover, it is to be understood that, although various graphical user interfaces (GUI's) and/or screen touch, swipe click-on, etc. activating actions are described herein as illustrative examples, it is within the contemplation of the disclosure to use user interfaces other than or in addition to GUI's and screen haptic interfacing; these including, but not being limited to; (1) voice only or voice-augmented interfaces (e.g., provided through a user worn head set or earpiece (i.e. a BlueTooth™ compatible earpiece—see FIG. 2); (2) sight-independent touch/tactile interfaces such as those that might be used by visually impaired persons; (3) gesture recognition interfaces such as those where a user's hand gestures and/or other body motions and/or muscle tensionings or relaxations are detected by automated means and converted into computer-usable input signals; and so on; (4) wrist, arm, leg, finger, toe action recognition interfaces such as those where a user wears a wrist-watch like device or an instrumented arm bracelet or an ankle bracelet or an elastic arm band or an instrumented shoe or an instrumented glove or instrumented other garments (or a flexible thin film circuit attached to the user) and the worn device includes acceleration-detecting, location-detecting, temperature-detecting, muscle activation-detecting, perspiration-detecting or like means (e.g., in the form of a MEMs chip) for detecting user body part motions, states, or tensionings or heatings/coolings and means for reporting the same to a corresponding user interface module. More specifically, in one embodiment, the user wears a wrist watch that has a BlueTooth™ interface embedded therein and allows for screen data to be sent to the watch from a host (e.g., as an SMS message) and allows for short replies to be sent from the watch back to the BlueTooth™ host, where here the illustrated tablet computer 100 operates as the BlueTooth™ host and it repeatedly queries the wrist watch (not shown) to respond with telemetry for one or more of detected wrist accelerations, detected wrist locations, detected muscle actuations and detected other biometric attributes (e.g., pulse, skin resistance).
  • In one variation, the insides of a user's mouth are instrumented such that movement of the tip of the tongue against different teeth and/or the force of contact by the tongue against teeth and/or other in-mouth surfaces are used to signal conscious or subconscious wishes of the user. More specifically, the user may wear a teeth-covering and relatively transparent mouth piece that is electronically and/or optically instrumented to report on various inter-oral cavity activities of the user including teeth clenchings, tongue pressings and/or fluid moving activities where corresponding reporting signals are transmitted to the user's local data processing device for possible inclusion in CFi reporting signals, where the latter can be used by the STAN_3 system to determine levels of attentiveness by the user relative to various focused-upon objects.
  • In one embodiment, the user alternatively or additionally wears an instrumented necklace or such like jewelry piece about or under his/her neck where the jewelry piece includes one or more, embedded and forward-pointing video cameras and a wireless short range transceiver for operatively coupling to a longer range transceiver provided nearby. The longer range transceiver couples wirelessly and directly or indirectly to the STAN_3 system. In addition to the forward pointing digital camera(s), the jewelry piece includes a battery means and one or more of sound pickups, biological state transducers, motion detecting transducers and a micro-mirrors image forming chip. The battery means may be repeatedly recharged by radio beams directed to it and/or by solar energy when the latter is available and/or by other recharging means. The embedded biological state transducers may detect various biological states of the wearer such as, but not limited to, heart rate, respiration rate, skin galvanic response, etc. The embedded motion detecting transducers may detect various body motion attributes of the wearer such as being still versus moving and if moving, in what directions and at what speeds and/or accelerations and when. The micro-mirrors image forming chip may be of a type such as developed by the Texas Instruments™ Company which has tiltable mirrors for forming a reflected image when excited by an externally provided, one or more laser beams. In one embodiment, the user enters an instrumented area that includes an automated, jewelry piece tracking mechanism having colored laser light sources within it as well as an optional IR or UV beam source. If an image is to be presented to the user, a tactile buzzer included in the necklace alerts him/her and indicates which way to face so that the laser equipped tracking mechanism can automatically focus in upon the micro-mirrors based image forming device (surrounded by target patterns) and supply excitational laser beams safely to it. The reflected beams form a computer generated image that appears on a nearby wall or other reflective object. Optionally, the necklace may include sound output devices or these can be separately provided in an ear-worn BlueTooth™ device or the like.
  • Informational resources of the STAN_3 system may be provided to the so-instrumented user by way of the projected image wherever a correspondingly instrumented room or other area is present. The user may gesture to the STAN_3 system by blocking part of the projected image with his/her hand or by other means and the necklace supported camera sees this and reports the same back to the STAN_3 system. In one embodiment, the jewelry piece includes two embedded video cameras pointing forward at different angles. One camera may be aimed at a wall mounted mirror (optionally an automatically aimed one which is driven by the system to track the user's face) where this mirror reflects back an image of the user's head while the other camera may be aimed at projected image formed on the wall by the laser beams and the micro-mirrors based reflecting device. Then the user's facial grimaces may be automatically fed back to the STAN_3 system for detecting implicit or explicit voting expressions as well as other user reactions or intentional commands (e.g., tongue projection based commands). In one embodiment, the user also wears electronically driven shutter and/or light polarizing glasses that are shuttered and/or variably polarized in accordance with an over-time changing pattern that is substantially unique to the user. The on-wall projected image is similarly modulated such that only the spectacles-wearing user can see the image intended for him/her. Therefore, user privacy is protected even if the user is in a public instrumented area. Other variations are of course possible, such as having the cameras and image forming devices placed elsewhere on the user's body (e.g., on a hat, a worn arm band near the shoulder, etc.). The necklace may include additional cameras and/or other sensors pointing to areas behind the user for reporting the surrounding environment to the STAN_3 system.
  • Referring still to the illustrative example of FIG. 1A and also to a further illustrative example provided in corresponding FIG. 1B, the user is assumed in this case to have selected a rotating-pyramids visual-radar displaying method for presenting the selected usage attribute(s) (e.g., heat per my now top 5 topics as measured in at least two time periods—two simultaneously showing faces of a pyramid). Here, the two faces of a periodically or sporadically revolving or rotationally reciprocating pyramid (e.g., a pyramid having a square base, and whose rotations are represented by circular arrow 101 u′) are simultaneously seen by the user. One face 101 w′ graphs so-called temperature or heat attributes of his currently focused-upon, top-N topics as determined over a corresponding time period (e.g., a predetermined duration such as over the last 15 minutes). That first period is denoted as “Now”. The other face 101 x′ provides bar graphed temperatures of the identified top topics of “Me” for another time period (e.g., a predetermined duration such as between 2.5 hours ago and 3.5 hours ago) which in the example is denoted as “3 Hours Ago”. The chosen attributes and time periods can vary according to user editing of radar options in an available settings menu. While the example of FIG. 1B displays “heat” per topic node (or per topic region), it is within the contemplation of the present disclosure to alternatively or additionally display “heat” per keyword node (or per keyword region in a corresponding keyword space, where the latter concept is detailed below in conjunction with FIG. 3E) and to alternatively or additionally display “heat” per hybrid node (or per hybrid region in a corresponding hybrid space, where the latter concept is also detailed below in conjunction with FIG. 3E). Although a rotating pyramid having an N-sided base (e.g., N=3, 4, 5, . . . ) is one way of displaying graphed heats, such “heat” temperatures or other user-selectable attributes for different time periods and/or for different user-touchable sub-spaces that include but are not limited to: not only ‘touched’ topic zones, but alternatively or additionally: touched geographic zones or locations, touched context zones, touched habit zones, touched social dynamic zones and so on of a specified user (e.g., the leader or KoH entity), it is also within the contemplation of the present disclosure to instead display such things on respective faces of other kinds of M-faced rotating polyhedrons (where M can be 3 or more, including very large values for M if so desired). These polyhedrons can rotate about different axes thereof so as to display in one or more forward winding or backward winding motions, multiple ones of such faces and their respective attributes.
  • It is also within the contemplation of the present disclosure to use a scrolling reel format such as illustrated in FIG. 1D where the displayed reel winds forwards or backwards and occasionally rewinds through the graph-providing frames of that reel 101 ra′″. In one embodiment, the user can edit what will be displayed on each face of his revolving polyhedron (e.g., 101 ra″ of FIG. 1C) or in each frame of the winding reel (e.g., 101 ra′″ of FIG. 1D) and how the polyhedron/reeled tape will automatically rotate or wind and rewind. The user-selected parameters may include for example, different time ranges for respective time-based faces, different topics and/or different other ‘touchable’ zones of other spaces and/or different social entities whose respective ‘touchings’ are to be reported on. The user-selected parameters may additionally specify what events (e.g., passage of time, threshold reached, desired geographic area reached, check-in into business or other establishment or place achieved, etc.) will trigger an automated rotation to, and a showing off of a given face or tape frame and its associated graphs or its other metering or mapping mechanisms.
  • In FIGS. 1A, 1B, 1D as well as in others, there are showings of so-called, affiliated space flags (101 s, 101 s′, 101 s′″). In general, these affiliated space flags indicate a corresponding one or more of system maintained, data-object organizing spaces of the STAN_3 mechanism which spaces can include a topics space (TS—see 313″ of FIG. 3D), a content space (CS—see 314″ of FIG. 3D), a context space (XS—see 316″ of FIG. 3D), a normalized CFi categorizing space (where normalization is described below—see 302″ and 298″ of FIG. 3D), and other Cognitive Attention Receiving Spaces—a.k.a. “CARS's” and/or other Cognition-Representing Objects Organizing Spaces—a.k.a. “CROOS's”. Each affiliated space flag (e.g., 101 s, 101 s′, etc.) can be displayed as having a respective one or more colors, shape and/or glyphs presented thereon for identifying its respective space. For example, the topic-space representing flags may have a target bull's eye symbol on them. If a user control clicks or otherwise activates the affiliated space flag (e.g., 101 s′ of FIG. 1B), a corresponding menu (not shown) pops open to provide the user with more information about the represented space and/or a represented sub-region of that space and to provide the user with various search and/or navigation functions relating to the represented space. One of the menu-provided options allows the user to pop open a local map of a represented topic space region (TSR) where the map can be in a hierarchical tree format (see for example 185 b of FIG. 1G—“You are here in TS”) or the map can be in a terraced terrain format (see for example plane 413′ of FIG. 4D).
  • Incidentally, as used herein, the term “Cognition-Representing Objects Organizing Space” (a.k.a. CROOS) is to be understood as referring to a more generic form of the species, “Cognitive Attention Receiving Space” (a.k.a. CARS) where both are data-objects organizing spaces represented by data objects stored in system memory and logically inter-linked or otherwise organized according to application-specific details. When a person (e.g., a system user) gives conscious attention to a particular kind of cognition, say to a textual cognition; which cognition can more specifically be directed to a search-field populating “keyword” (which could be a simultaneous collection or a temporal clustering of keywords), then as a counterpart machine operation, a representing portion of a counterpart, conscious Cognition Attention Receiving Space (CARS) should desirably be lit up (focused-upon) in a machine sense to reflect a correct modeling of a lighting up of (energizing of) the corresponding cognition providing region in the user's brain that is metabolically being lit up (energized) when the user is giving conscious attention to that particular kind of cognition (e.g., re a “keyword”). Similarly, when a system user gives conscious attention to a question like, “What are we talking about?” and to its answer (“What are we talking about?”), that is referring to what in the machine counterpart system would be a lighting up of (e.g., activation of) a counterpart point, node or subregion in a system-maintained topic space (TS). Some cognitions however, do not always receive conscious attention. An example might be how a person subconsciously parses (syntactically disambiguates) a phonetically received sentence (e.g., “You too/two[?] should see/sea[?] to it[?]”) and decodes it for semantic sense. That often happens subconsciously. At least one of the data-objects organizing spaces discussed herein (FIG. 3V) will be directed to that aspect and the machine-implemented data-objects organizing space that handles that aspect is referred to herein as a Cognition-Representing Objects Organizing Space (a.k.a. CROOS) rather than as a Cognitive Attention Receiving Space (a.k.a. CARS).
  • The present disclosure, incidentally, does not claim to have discovered how to, nor does it endeavor to represent cognitions within the human mind down to the most primitive neuron and synapse actuations. Instead, and as shall be detailed below, a so-called, primitive expressions (or symbols or codings) layer is contemplated within which is stored machine codes representing corresponding expressions, symbols or codings where the latter represent a meta-level of human cognition, say for example, a semantic sense of what a particular text string (e.g., “Lincoln”) represents. The meta-level cognitions can be combined in various ways to build yet more complex representations of cognitions (e.g., “Lincoln” plus “Abraham”; or “Lincoln” plus “Nebraska”; or “Lincoln” plus “Car Dealership”). Although it is not an absolute requirement of the present disclosure, preferably, the primitive expressions storing (and clustering) layer is a communally created and communally updated layer containing “clusterings” of expressions, symbols or codings where a relevant community of users implicitly determines what cognitive sense each such expression or clustering of expressions represents, where legacy “clusterings” of expressions, etc. are preserved and yet new “clusterings” of such expressions, etc. can be added or inserted as substitutes as community sentiments change with regard to such adaptively updateable, expressions, codings, or other symbols that implicitly represent underlying cognitions. More specifically, and as a brief example, prior to September 2011, the expression string” “911” may have most likely invoked the cognitive sense in a corresponding community of a telephone number that is to be dialed In Case of Emergency (ICE). However, after said date, the same expression string” “911” may most likely invoke the cognitive sense in a corresponding community of an attack on the World Trade Center in New York City. For that brief example, an embodiment in accordance with the present disclosure would seek to preserve the legacy cognitive sense while at the same supplanting it with the more up to date cognitive sense. Details of how this can be done are provided later below.
  • Still referring to FIGS. 1A-1D, some affiliated space flags, such as for example the specially shaped flag 101 sh″ topping the pyramid 101 ra″ of FIG. 1C provide the user with expansion tool (e.g., starburst+) access to a corresponding Cognitive Attention Receiving Space (CARS) or to a corresponding Cognition-Representing Objects Organizing Space (a.k.a. CROOS) directed to social dynamics as may be developing between two or more people or groups of people. (The subject of social dynamics will be explored in greater detail later, in conjunction with FIG. 1M.) For sake of intuitively indicating to the user that the pyramid 101 ra″ relates to interpersonal dynamics, an icon 101 p″ showing two personas and their intertwined discourses may be displayed under the affiliated space flag 101 sh″. If the user clicks or otherwise activates the expansion tool (e.g., starburst+) disposed inside the represented dialog of the one of the represented people (or groups), addition information about the person (or group) and his/her/their current dialogs is automatically provided. In one embodiment, in response to activating the dialog expansion tool (e.g., starburst+), a system maintained profile of the represented persona or group is displayed (where persona does not necessarily mean the real life (ReL) person and/or his/her real life identity and real life demographic details but could instead mean an online persona with limited information about that online identity).
  • Additionally, in one embodiment and in response to activating the dialog expansion tool (e.g., starburst+), a current thread of discourse by the respective persona is displayed, where the thread typically is one inside an on-topic chat or other forum participation session for which a “heat of exchange” indication 101 w″ is displayed on the forward turned (101 u″) face (e.g., 101 t″ or 101 x″) of the heat displaying pyramid 101 ra″. Here the “heat of exchange” indication 101 w″ is not showing “heat” cast by a single person on a particular topic but rather heat of exchange as between two or more personas as it may relate to any corresponding point, node or subregion of a respective Cognitive Attention Receiving Space where the later could be topic space (TS) for example, but not necessarily so. Expansion of the social dynamics tree flag 101 sh″ will show how social dynamics between the hotly involved two or more personas (e.g., debating persons) is changing while the “heat of exchange” indications 101 w″ will show which amount of exchange heat and activation of the expansion tool (e.g., starburst+) on the face (e.g., 101 t″) of the pyramid will indicate which topic or topics (or points, nodes or subregions (a.k.a. PNOS's) of another Cognitive Attention Receiving Space) are receiving the heat of the heated exchange between the two or more persons. It may be that there is no one or more points, nodes or subregions receiving such heat, but rather that the involved personas are debating or otherwise heatedly exchanging all over the map. In the latter case, no specific Cognitive Attention Receiving Space (e.g., topic space) and regions thereof will be pinpointed.
  • If the user of the data processing device of FIG. 1A wants to quickly spot when heated exchanges are developing as between for example, which two or more of his friends as it may or may not relate to one or more of his currently Top 5 Now Topics, the user may command the system to display a social heats pyramid like 101 ra″ (FIG. 1C) in the radar column 101 r of FIG. 1A as opposed to displaying a heat on specific topic pyramid such as 101 ra′ of FIG. 1B. The difference between pyramid 101 ra″ (FIG. 1C) and pyramid 101 ra′ (FIG. 1B) is that the social heats pyramid (of FIG. 1C) indicates when a social exchange between two or more personas is hot irrespective of topic (or it could be limited to a specified subset of topics) whereas the on-topic pyramid (e.g., of FIG. 1B) indicates when a corresponding point, node or subregion of topic space (or another specified Cognitive Attention Receiving Space) is receiving significant “heat” irrespective of whether or not a hot multi-person exchange is taking place. Significant “heat” may be cast for example upon a topic node even if only one persona (but a highly regarded persona, e.g., a Tipping Point Person) is casting the heat and such would show up on an on-topic pyramid such as 101 ra′ of FIG. 1B but not on a social heats pyramid such as that of FIG. 1C. On the other hand, two relatively non-hot persons (e.g., not experts) may be engaged in a hot exchange (e.g., a heated debate) that shows up on the social heats pyramid of FIG. 1C but not on the on-topic pyramid 101 ra′ of FIG. 1B. The user can select which kind of radar he wants to see.
  • Referring to FIG. 1D, the radar like reporting tool are not limited to pyramids or the like and may include the illustrated, scrollable (101 u′″) reel 101 ra′″ of frames where each frame can have a different space affiliation (e.g., as indicated by affiliated space flag 101 s′″) and each frame can have a different width (e.g., as indicated by within-frame scrolling tool 101 y′″ and each frame can have a different number of heat or other indicator bars or the like within it. As was the case elsewhere, each affiliated space flag (e.g., 101 s′″) can have its own expansion tool (e.g., starburst+) 101 s+′″ and each associated frame can have its own expansion tool (e.g., starburst+) so that more detailed information and/or options for each can be respectively accessed. The displayed heats may be social exchange heats as is indicated by icon 101 p′″ of FIG. 1D rather than on-topic heats. The non-heat axis (e.g., 144 of FIG. 1D) may represent different persons of pairs of persons rather than specific topics. The different persons or groups of exchanging persons may be represented by different colors, different ID numbers and so on. In the case of per topic heats, the corresponding non-heat axis (e.g., 143 of FIG. 1D) may identify the respective topic (or other point, node or subregion of a different Cognitive Attention Receiving Space) by means of color and/or ID number and/or other appropriate means (e.g., glowing an adjacent identification glyph when the bar is hovered over by a cursor or equivalent). A vertical axis line 142 may be provided with attached expansion tool information (starburst+ not shown) that indicates specifically how the heats of a focused-upon frame are calculated. More details about possible methods of heat calculation will be provided below in conjunction with FIG. 1F. A control portion 141 of the reel may include tools for advancing the reel forward or rewinding it back or shrinking its unwound length or minimizing (hiding) it.
  • In summary, when a user sees an affiliated space flag (e.g., 101 s′) atop an attributes mapping pyramid (e.g., 101 ra′ of FIG. 1B) or attached (e.g., 101 s′″ of FIG. 1D) to a reeled frame, the user can often quickly tell from looking at the flag, what data-object organizing space (e.g., topic space) is involved, or if not, the flag may indicate another kind of heat mapping; such as for example one relating to heat of exchange between specified persons rather than with regard to a specific topic. On each face of a revolving pyramid, or alike polyhedron, or back and forth winding tape reel (141 of FIG. 1D), etc., the bar graphed (or otherwise graphed) and so-called, temperature parameter (a.k.a. ‘heat’ magnitude) may represent any of a plurality of user-selectable attributes including, but not limited to, degree and/or duration of focus on a topic or on a topic space region (TSR) or on another space node or space sub-region (e.g., keywords space, URL's space, etc) and/or degree of emotional intensity detected as statistically normalized, averaged, or otherwise statistically massaged for a corresponding social entity (e.g., “Me”, “My Friend”, “My Friends” (a user defined group), “My Family Members”, “My Immediate Family” (a user defined or system defined group), etc.) and optionally as the same regards a corresponding set of current top N now nodes of the KOH entity 101 a designated in the social entities column 101 of FIG. 1A.
  • In addition to displaying the so-called “heats” cast by different social entities on respective topic or other nodes, the exemplary screen of FIG. 1A provides a plurality of invitation “serving plates” disposed on a so-called, invitations serving tray 102. The invitations serving tray 102 is retractable into a minimized mode (or into mostly off-screen hidden mode in which only the hottest invitations occasionally protrude into edges of the screen area) by clicking or otherwise activating Hide tool 102 z. In the illustrated example, invitations to chat or other forum participation sessions related to the current top 5 topics of the head entity (KoH) 101 a are found in compacted form on a current top topics serving plate (or listing) 102 a Now displayed as being disposed on the top serving tray 102 of screen 111. If the user hovers a cursor or other pointer object over a compacted invitations object such as over circle 102 i, a de-compacted invitations object such as 102J pops out. In one embodiment, the de-compacted invitations object 102J appears as a 3D, inverted Tower of Hanoi set of rings, where the largest top rings represent the newest, hottest invitations and the lower, smaller and receding toward disappearance rings represent the older, growing colder invitations for a same topic subregion. In other words, there is a continuous top to bottom flow of invitation-representing objects directed to respective subregions of topic space. The so de-compacted invitations object 102J not only has its plurality of stacked and emerging or receding rings, but also a starburst-shaped center pole and a darkened outer base disc platform. Hovering or clicking or otherwise activating these different concentric areas (rings, center post, base) of the de-compacted invitations object 102J provides further functions; including immediately popping open one or more topic-related chat or other forum participation opportunities (not shown in FIG. 1A, but see instead the examples 113 c, 113 d, 113 e of FIG. 1I). In one embodiment, when hovering over a de-compacted invitations object such as a Tower of Hanoi ring in the 3D version of 102J or its more compacted seed 102 i, a blinking of a corresponding spot is initiated in playgrounds column 103. The playgrounds column 103 displays a set of platform-representing objects, 103 a, 103 b, . . . , 103 d to which the corresponding chat or other forum participation sessions belong. More specifically, if one of the chat rooms; for which a join-now invitation (e.g., a Tower of Hanoi Like ring) is available, is maintained by the STAN_3 system, then the corresponding STAN3 playground object 103 a will blink, glow or otherwise make itself apparent. Alternatively or additionally a translucent connection bridge 103 i will appear as extending between the playground representing icon 103 a and the de-compacted invitations object 102J that holds an invitation for immediately joining in on an online chat belonging to that playground 103 a. Thus a user can quickly see which platform an invitation belongs to without actually accepting the invitation. More specifically, if one of the invited-to-it forum opportunities (e.g., Tower of Hanoi Like rings) belongs to the FB playground 103 b, then that playground representing object 103 b will glow and a corresponding translucent connection bridge 103 k will appear as extending between the FB playground 103 b and the de-compacted invitations object 102J. The same holds true for playground representing objects 103 c and 103 d. Thus, even before popping open the forum(s) of an invitations-serving object like 102J or 102 i, the user can quickly find out what one or more playgrounds (103 a-103 d) are hosting corresponding chat or other forum participation sessions relating to the corresponding topic (the topic of bubble 102 i).
  • Throughout the present disclosure, a so-called, starburst+ expansion tool is depicted as a means for obtaining more detailed information. Referring for example to FIG. 1B and more specifically to the “Now” face 101 w′ of that pyramid 101 ra′, at the apex of that face there is displayed a starburst+ expansion tool 101 t+′. By clicking or otherwise activating there, the user activates a virtual magnifying or details-showing and unpacking function that provides the user with an enlarged and more detailed view of the corresponding object and/or object feature (e.g., pyramid face) and its respective components. It is to be understood that in FIGS. 1A-1D as well as others, a plus symbol (+) inside of a star-burst icon (e.g., 101 t+′ of FIG. 1B or 99+ of FIG. 1A) indicates that such is a virtual magnification/unpacking invoking button tool which, when activated (e.g., by clicking or otherwise activating) will cause presentation of a magnified or expanded-into-more detailed (unpacked) view of the object or object portion. The virtual magnification button may be activated by on-touch-screen finger taps, swipes, etc. and/or other activation techniques (e.g., mouse clicks, voice command, toe tap command, tongue command against an instrumented mouth piece, etc.). Temperatures, as a quantitative indicator of cast “heat”; may be represented as length or range of the displayed bar in bar graph fashion and/or as color or relative luminance of the displayed bar and/or flashing rate of a blinking bar where the flashing may indicate a significant change from last state and/or an above-threshold value of a determined “heat” value (e.g., emotional intensity) associated with the now-“hot” item. These are merely non-limiting examples. Incidentally, in FIG. 1A, embracing hyphens (e.g., those at the start and end of a string like:—99+—) are generally used around reference numbers to indicated that these reference symbols are not displayed on the display screen 111.
  • Still referring to FIG. 1B, in one embodiment, a special finger waving flag 101 fw may automatically pop out from the top of the pyramid (or reel frame if the format of FIG. 1D is instead used) at various times. The popped out finger waving flag 101 fw indicates (as one example of various possibilities) that the tracked social entity has three out of five of commonly shared topics (or other types of nodes) with the column leader (e.g., KoH=‘Me’) where the “heats” of the 3 out of 5 exceed respective thresholds or exceed a predetermined common threshold. The heat values may be represented by translucent finger colors, red being the hottest for example. In other words, such a 2-fingered, 3, 4, etc. fingered wave of a virtual hand (e.g., 101 fw) alerts the user that the corresponding non-leader social entity (could be a person or a group) is showing above-threshold heat not just for one of the current top N topics of the leader (of the KoH), but rather for two or more, or three or more shared topic nodes or shared topic space regions (TSR's—see FIG. 3D), where the required number of common topics and level of threshold crossing for the alerting hand 101 fw to pop up is selected by the user through a settings tool (114) and, of course, the popping out of the waving hand 101 fw may also be turned off if the user so desires. The exceeding-threshold, m out of n common topics function may be provided not only for the alert indication 101 fw shown in FIG. 1B, but also for similar alerting indications (not shown) in FIG. 1C, in FIGS. 1D and 1 n FIG. 1K. The usefulness of such an m out of n common topics indicating function (where here m≤n and both are whole numbers) will be further explained below in conjunction with later description of FIG. 1K. Basically, when another user is currently focused-upon a plurality of same or similar topics as is the first user, they are more likely to have much in common with each other as compared to a users who have only one topic node in common with one another.
  • Referring back to the left side of FIG. 1A, it is to be assumed that reporting column 101 r is repeatedly changing (e.g., periodically being refreshed). Each time the header (leader, KoH, Pharaoh's) pyramid 101 ra (or another such “heat” and/or commonality indicating means) rotates or otherwise advances to a next state to thus show a different set of faces thereof, and to therefore show (in one embodiment) a different set of cross-correlated time periods or other context-representing faces; or each time the header object 101 ra partially twists and returns to its original angle of rotation, the follower pyramids 101 rb-101 rd (or other radar objects) below it will follow suite (but perhaps with slight time delay to show that they are mirroring followers, not leaders who define their own top N topics). At this time of pyramid rotation, the displayed faces of each pyramid (or other radar object) are refreshed to show the latest temperature or heats data for the displayed faces (or displayed frames on a reel; 101 ra′″ of FIG. 1D) and optionally where a predetermined threshold level has been crossed by the displayed heat or other attribute indicators (e.g., bar graphs). As a result, the user (not shown in 1A, see instead 201A of FIG. 2) of the tablet computer 100 can quickly see a visual correlation as between the top topics of the header entity 101 a (e.g., KoH=“Me”) and the intensity with which other associated social entities 101 b-101 d (e.g., friends and family) are also focusing-upon those same topic nodes (top topics of mine) during a relevant time period (e.g., Now versus X minutes ago or H hours ago or D days ago). In cases where there is a shared large amount of ‘heat’ with regard to more than one common topic, the social entities that have such multi-topic commonality of concurrently large heats (e.g., 3 out of 5 are above-threshold per for example, what is shown on face 101 w′ of FIG. 1B); such may be optionally flagged (e.g., per waving hand object 101 fw of FIG. 1B) as deserving special attention by the user. Incidentally, the header entity 101 a (e.g., KoH=“Me”) does not have to be the user of the tablet computer 100. Also, the time periods reported by the respective faces of the KoH pyramid 101 ra do not have to be the same as the time periods reported by the respective faces (e.g., 101 t, 101 x of follower pyramid 101 rb) of the subservient pyramids 101 rb-101 rd. It is possible that the KoH=Me entity just began this week to focused-upon topics 3 through 5 with great intensity (large “heat”) whereas two of his early adapter friends were already focused-upon topic 4 two weeks ago (and maybe they have moved onto a brand new topic number 6 this week). Nonetheless, it may be useful to the user to learn that his followed early adapters (e.g., “My Followed Tipping Point Persons”—not explicitly shown in FIG. 1A, could be disc 101 d) were hot about that same one or more topics two weeks ago. Accordingly, while the follower pyramids may mirror the KoH (when a KoH is so anointed) in terms of tracked topic nodes and/or tracked topic space regions (TSR) and/or tracked other nodes/subregions of other spaces; they do not necessarily mirror the time periods of the KoH reporting object (101 ra) in an absolute sense (although they may mirror in a relative sense by having two pyramid faces that are about H hours apart or about D days apart and so on).
  • The tracked social entities of left column 101 do not necessarily have to be friends or family or other well-liked or well-known acquaintances of the user (or of the KoH entity; not necessarily same as the user). Instead of being persons or groups whom the user admires or likes, they can be social entities whom the user despises, or feels otherwise about, or which the first user never knew before, but nonetheless the first user wishes to see what topics are currently deemed to be the “topmost” and/or “hottest” for that user-selected header entity 101 a (where KoH is not equal to “Me”) and further social entities associated with that user-selected KoH entity. Incidentally, in one embodiment, when the user selects a new KoH entity (e.g., new KoH=“Charlie”), the system automatically presents the user with a set of options: (a) Don't change the other discs in column 101; (b) Replace the current discs 101 b-101 d in column 101 with a first set of “Charlie”-associated other entity discs (e.g., “Charlie's Family”, “Charlie's Friends”, etc.); (c) Replace the current discs 101 b-101 d in column 101 with a second set of “Charlie”-associated other entity discs (e.g., “Charlie's Workplace Colleagues”, etc.) and (d) Replace the current discs 101 b-101 d in column 101 with a new third set that the user will next specify. Thus, by changing the designated KoH entity, the user may not only change the identification of the currently “hot” topics whose heats are being watched, but the user may also change, by substantially the same action, the identifications of the follower entities 101 b-101 d.
  • While the far left side column 101 of FIG. 1A is social-entity “centric” in that it focuses on individual personas or groups of personas (or projects associated with those social entities), the upper top row 102 (a.k.a. upper serving tray) is topic “centric” in one sense and, in a more general way, it can be said to be ‘touched’-space centric because it serves up information about what nodes or subregions in topic space (TS); or in another Cognitive Attention Receiving Space (e.g., keyword space (KS)) have been “touched” by others or should be (are automatically recommended by the system to be) “touched” by the user. The term, ‘touching’ will be explained in more detail later below. Basically, there are at least two kinds of ‘touching’, direct and indirect. When a STAN_3 user “touches” a node or subregion (e.g., a topic node (TN) or a topic region (TSR)) of a given, system-supported “space”, that ‘touching’ can add to a heat count associated with the node or subregion. The amount of “heat”, its polarity (positive or negative), its decay rate and so on may depend on who the toucher(s) is/are, how many touchers there are, and on the intensity with which each toucher virtually “touches” that node or subregion (directly or indirectly). In one embodiment, when a node is simultaneously ‘touched’ by many highly ranked users all at once (e.g., users of relatively high reputation and/or of relatively high credentials and/or of relatively high influencing capabilities), it becomes very “hot” as a result of enhanced heat weights given to such highly ranked users.
  • In the exemplary case of FIG. 1A, the upper serving tray 102 is shown to be presenting the user with different sets of “serving plates” (e.g., 102 aNow, 102 a′Earlier, . . . , 102 b (Their Top 5), etc.). As will become more apparent below, the first set 102 a of “serving plates” relate to topics which the “Me” entity (101 a) has recently been focused-upon with relatively large “heat”. Similarly, the second set 102 b of “serving plates” relate to topics which a “Them” entity (e.g., My Friends 101 c) has recently been focused-upon with relatively large “heat”. Ellipses 102 c represent yet other upper tray “serving plates” which can correspond to yet other social entities (e.g., My Others 101 d) and, in one specific case, the topics which those further social entities have recently been focusing-upon with relatively large “heat” (where here, ‘recently’ is a relative term and could mean 1 year ago rather than 1 hour ago). However, in a more generic sense, the further “serving plates” represented by ellipses 102 c can correspond to generic nodes or subregions (e.g., in keyword space, context space, etc.) which those further social entities have recently been ‘touching’ upon with relatively large amounts of “heat”. (It is also within the contemplation of the disclosure to report on nodes or subregions that have been ‘touched’ by respective social entities with minimal or zero “heat” although, often, that information is of limited interest.)
  • In one embodiment, the changing of designation of who (what social entity) is the KoH 101 a automatically causes the system to present the user with a set of upper-tray modification options: (a) Don't change the serving plates on tray 102; (b) Replace the current serving plates 102 a, 102 b, 102 c in row 102 with a first set of “Charlie”-associated other serving plates (e.g., “Charlie's Top 5 Now Topics”, “Charlie's Family's Top 5 Now Topics”, etc. where here the KoH is being changed from being “Me” to being “Charlie”); (c) Replace the current serving plates 102 a, 102 b, 102 c in row 102 with a second set of “Charlie”-associated other serving plates (e.g., “Top N now topics of Charlie's Workplace Colleagues”, “Top M now keywords being used by Charlie's Workplace Colleagues”, etc.); and (d) Replace the current serving plates 102 a, 102 b, 102 c in row 102 with a new third set of serving plates that the user will next specify. Thus, by changing the designated KoH entity, the user may not only change the identification of the currently “hot” topics (or other “hot” nodes) whose heats are being watched in reporting column 101 r, but the user may also change, by substantially the same action, the identifications of the serving plates in the upper tray area 102 and the nature of the “touched” or to-be-“touched” items that they will serve up (where those “touched” or to-be-“touched” items can come in the form of links to, or invitations to, chat or other forum participation sessions that are “on-topic” or links to suggested other kinds of content resources that are deemed to be “on-topic” or links to, or invitations to, chat or other forum participation sessions or other resources that are deemed to be well cross-correlated with other types of ‘touched’ nodes or subregions (e.g., “Top M now keywords being used by Charlie's Workplace Colleagues”). At the same time the upper tray items 102 a-102 c are being changed due to switching of the KoH entity, the identifications of the corresponding follower entities 101 b-101 d may also be changed.
  • The so-called, upper serving plates 102 a, 102 b, 102 c, etc. of the upper serving tray 102 (where 102 c and the extendible others which may be accessible for enlarged viewing with use of a viewing expansion tool (e.g., clicking or otherwise activating the 3 ellipses 102 c)). These upper serving plates are not limited to showing (serving up) an automatically determined set of recently ‘touched’ and “hot” nodes or subregions such as a given social entities' top 5 topics or top N topics (where N can be a number other than 5 here, and where automated determination of the recently ‘touched’ and “hot” nodes or subregions in a selected space (e.g., topic space) can be based on predetermined knowledge base rules). Rather, the user can manually establish how many ‘touched’-topics or to-be-‘touched’/recommended topics serving plates 102 a, 102 b, etc. (if any) and/or other ‘touched’/recommended node serving plates (e.g., “Top U now URL's being hyperlinked to by Charlie's Workplace Colleagues”, —not shown) will be displayed on the “hot” nodes or hot space subregions serving tray 102 (where the tray can also serve up “cold” items if desired and where the serving tray 102 can be hidden or minimized (via tool 102 z)). In other words, instead of relying on system-provided templates (recommended) for determining which topic or collection of topics will be served up by each “hot” now topics serving plate (e.g., 102 a), the user can use the setting tools 114 to establish his own, custom tailored, serving rules and corresponding plates or his own, custom tailored, whole serving trays where the items served up on (or by) such carriers can include, but are not limited to, custom picked topic nodes or subregions and invitations to chat or other forum participation sessions currently or soon to be tethered to such topic nodes and/or links to other on-topic resources suggested by (linked to by and rated highly by) such topic nodes. Alternatively or additionally, the user can use the setting tools 114 to establish his own, custom tailored, serving plates or whole serving trays where the items served on such carriers can include, but are not limited to, custom picked keyword nodes or subregions, custom picked URL nodes or subregions, or custom picked points, nodes or subregions (a.k.a. PNOS's) of another Cognitive Attention Receiving Space. The topics on a given topics serving plate (e.g., 102 a) do not have to be related to one another, although they could be (and generally should be for ease of use).
  • Incidentally, the term, “PNOS's” is used throughout this disclosure as an abbreviation for “points, nodes or subregions”. Within that context, a “point” is a data object of relatively similar data structure to that of a corresponding “node” of a corresponding Cognitive Attention Receiving Space or Cognitions-representing Space (e.g., topic space) except that the “point” need not be part of a hierarchical tree structure whereas a “node” is often part of a hierarchical, data-objects organizing scheme. Accordingly, the data structure of a PNOS “point” is to be understood as being substantially similar to that of a corresponding “node” of a corresponding Cognitions-representing Space except that fields for supporting the data object representing the “point” do not need to include fields for specifying the “point” as an integral part of a hierarchical tree structure and such fields may be omitted in the data structure of the space-sharing “point”. A “subregion” within a given Cognitions-representing Space (e.g., a CARS or Cognitive Attention Receiving Space) may contain one or more nodes and/or one or more “points” belonging to its respective Cognitions-representing Space. A Cognitions-representing Space may be comprised of hierarchically interrelated “nodes” and/or spatially distributed “points” and/or both of such data structures. A “node” may be spatially positioned within its respective Cognitions-representing Space as well as being hierarchically positioned therein.
  • The term, “cognitive-sense-representing clustering center point” also appears numerous times within the present disclosure. The term, “cognitive-sense-representing clustering center point” (or “center point” for short) as used herein is not to be confused with the PNOS type of “point”. Cognitive-sense-representing clustering center points (or COGS's for short) are also data structures similar to nodes that can be hierarchically and/or spatially distributed within a corresponding hierarchical and/or spatial data-objects organizing scheme of a given Cognitions-representing Space except that, at least in one embodiment, system users are not empowered to give names to such center points (COGS's) and chat room or other forum participation sessions do not directly tether to such COGS's and such COGS's do not directly point to informational resources associated with them (with the COGS's) or with underlying cognitive senses associated with the respective and various COGS's. Instead, a COGS (a single cognitive-sense-representing clustering center point) may be thought of as if it were a black hole in a universe populated by topic stars, subtopic planets and chat room spaceships roaming there about to park temporarily in orbit about one planet and then another (or to loop figure eight style or otherwise simultaneously about plural topic planets). Each COGS provides a clustering-thereto cognitive sense kind of force much like the gravitational force of a real world astronomical black hole provides an attracting-thereto gravitational force to nearby bodies having physical mass. One difference though, is that users of the at least one embodiment can vote to move a cognitive-sense-representing clustering center point (COGS) from one location to another within a Cognitions-representing Space (or a subregion there within) that they control. When a COGS moves, the points, nodes or subregions (PNOS's) that were clustered about it do not automatically move. Instead the relative hierarchical and/or spatial distances between the unmoved PNOS's and the displaced COGS change. That change indicates how close in a cognitive sense the PNOS's are deemed to be relative to an unnamed cognitive sense represented by the displaced COGS and vice versa. Just as in the physical astronomical realm where it is not possible (according to current understandings) to see what lies inward of the event horizon of a black hole, according to one aspect of the present disclosure, it is generally not permitted to directly define the cognitive sense represented by a COGS. Instead the represented cognitive sense is inferred from the PNOS's that cluster about and nearby to the COGS. That inferred cognitive sense can change as system users vote to move (e.g., drift) the nearby PNOS's to newer ones of hierarchical and/or spatial locations, thereby changing the corresponding hierarchical and/or spatial distances between the moved PNOS's and the one or more COGS that derive their inferred cognitive senses from their neighboring PNOS's. The inferred cognitive sense can also change if system users vote to move the COGS rather than moving the one or more PNOS's that closely neighbor it. A COGS may have additional attributes such substitutability by way of re-direction and expansion by use of expansion pointers. However, such discussion is premature at this stage of the disclosure and will be picked up much later below. (See for example and very briefly the discussion re COGS 30W.7 p of FIG. 3W.)
  • In one embodiment, different organizations of COGS's may be provided as effective for different layers of cognitive sentiments. More specifically, one layer of cognitive sentiments may be attributed to so-called, central or main-stream ways of thinking by the system user population while a second such layer of cognitive sentiments may be attributed to so-called, left wing extremist ways of thinking and yet a third such layer may be attributed to so-called, right wing extremist ways of thinking (this just being one possible set of examples). If a first user (or first persona) who subscribes to main-stream way of thinking logs in, the corresponding central or main-stream layer of accordingly organized COGS's is brought into effect while the second and third are rendered ineffective. On the other hand, if the logging-in first persona self-identifies him/herself as favoring the left wing extremist ways of thinking, then the second layer of accordingly organized COGS's is brought into effect while the first and third layers are rendered ineffective. Similarly, if the logging-in first persona self-identifies him/herself as favoring the right wing extremist ways of thinking, then the third layer of accordingly organized COGS's is brought into effect while the first and second layers are rendered ineffective. In this way, each sub-community of users, be they left-winged, middle of the road, or right winged (or something else) can have the topical universe presented to them with cognitive-sense-representing clustering center points being positioned in that universe according to the confirmation biasing preferences of the respective user. As mentioned, the left versus right versus middle of the road mindsets are merely examples and it is within the contemplation of the present disclosure to have more or other forms of multiple sets of activatable and deactivatable “layers” of differently organized COGS's where one or more such layers are activated (brought into effect) for a given one mindset and/or context of a respective user. In one embodiment, different governance bodies of respective left, right or other mindsets are given control over the hierarchical and/or spatial postionings of the COGS's of their respectively activatable layers where the controlled postionings are relative to the hierarchically and/or spatially organized points, nodes or subregions (PNOS's) of topic space and/or of another applicable, Cognitions-representing Space. In one embodiment, the respective governance bodies of respective Wikipedia™ like collaboration projects (described below) are given control over the postionings of the COGS's that become effective for their respective B level, C level or other hierarchical tree (described below) and/or semi-privately controlled spatial region within a corresponding Cognitions-representing Space.
  • In one embodiment, in addition to having the so-called, cognitive-sense-representing clustering center points (COGS's) around which, or over which, points, nodes or subregions (PNOS's) of substantially same or similar cognitive sense may cluster, with calculated distance being indicative of how same or similar they in accordance with a not necessarily articulated sense, it is within the contemplation of the present disclosure to have cognitive-sense-representing clustering lines, or curves or closed circumferences where PNOS-types of points, nodes or subregions disposed on a one such line, curve or closed circumference share a same cognitive sense and PNOS's distanced away from such line, curve or closed circumference are deemed dissimilar in accordance with the spacing apart distance calculated along a normal drawn from the spaced apart PNOS to the line, curve of circumference. In one embodiment, and yet alternatively or additionally, so-called, repulsion and/or exclusion center points, lines, curves or closed circumferences may be employed where PNOS-types of points, nodes or subregions are repulsed from (according to a decay factor) and/or are excluded from occupying a part of hierarchical and/or spatial space occupied by a respective, repulsion and/or exclusion type of center point, line, curve or closed circumference. The repulsion and/or exclusion types of boundary defining entities may be used to coerce the governance bodies who control placement of PNOS-types of points, nodes or subregions to distribute their controlled PNOS's more evenly within different bands of hierarchical and/or spatial space rather than clumping all such controlled PNOS's together. For example, if concentric exclusion circles are defined, then governance bodies are coerced into placing their controlled PNOS's into one of several concentric bands or another rather than organizing them as one unidifferentiated clump in the respective Cognitions-representing Space.
  • The topic of COGS, PNOS's, repulsion bands and so forth was raised here because the term PNOS's has been used a number of times above without giving it more of definition and this juncture in the disclosure presented itself as an opportune time to explain such things. The discussion now returns to the more mundane aspects of FIG. 1A and the displayed objects shown therein. Column 101 of FIG. 1A was being described prior to the digression into the topics of PNOS's, COGS and so on.
  • Referring to FIG. 1A, one or more editing functions may be used to determine who or what the header entity (KoH) 101 a is; and in one embodiment, the system (410) automatically changes the identity of who or what is the header entity 101 a at, for example, predetermined intervals of time (e.g., once every 10 minutes) or when special events take place so that the user is automatically supplied over time with a variety of different radar scope like reports that may be of interest. When the header entity (KoH) 101 a is automatically so changed, the leftmost topics serving plate (e.g., 102 a) is automatically also changed to, for example, serve up a representation of the current top 5 topics of the new KoH (King of the Hill) 101 a. As mentioned above, the selection of social entity representing objects in left vertical column 101 (or projects or other attributes cross-correlated with those social entities) including which one will serve as KOH (if there is a KoH) can automatically change based on one or more of a variety of triggering factors including, but not limited to, the current location, speed and direction of facing or traveling of the user, the identity of other personas currently known to the user (or believed by the user) to be in Cognitive Attention Giving Relation to the user based on current physical proximity and/or current online interaction with the user, by the current activity role adopted by the user (user adopted context) and also even based on the current floor that the Layer-Vator™ 113 has virtually brought the user to.
  • The ability to track the top-N topic(s) that the user and/or other social entity is now focused-upon (giving cognitive attention to) or has earlier focused-upon is made possible by operations of the STAN_3 system 410 (which system is represented for example in FIG. 4A as optionally including cloud-based and/or remote-server based and database based resources). These operations include that of automatically determining the more likely topics currently deemed to be on the minds of (receiving most attention from) logged-in STAN users by the STAN_3 system 410. Of course each user, whose topic-related temperatures are shown via a radar mechanism such as the illustrated revolving pyramids 101 ra-101 rd, is understood to have a-priori given permission (or double level permissions—explained below) in one way or another to the STAN_3 system 410 to share such information with others. In one embodiment, each user of the STAN_3 system 410 can issue a retraction command that causes the STAN_3 system to erase all CFi's and/or CVi's collected from that user in the last m minutes (e.g., m=2, 5, 10, 30, 60 minutes) and to erase from sharing, topical information regarding what the user was doing in the specified last m minutes (or an otherwise specified one or more blocks or ranges of time; e.g. from yesterday at 2 pm until today at 1 pm). The retraction command can be specific to an identified region of topic space instead of being global for all of topic space. (Or it can be alternatively or additionally be directed to other or custom picked points, nodes or subregions of other Cognitive Attention Receiving Spaces.) In this way, if the user realizes after the fact that what he/she was focusing-upon is something they do not want to have shared, they can retract the information to the extent it has not yet been seen by, or captured by others.
  • In one embodiment, each user of the STAN_3 system 410 can control his/her future share-out attributes so as to specify one or more of: (1) no sharing at all; (2) full sharing of everything; (3) limited sharing to a limited subset of associated other users (e.g., my trusted, behind-the-wall friends and immediate family); (4) limited sharing as to a limited set of time periods; (5) limited sharing as to a limited subset of areas on the screen 111 of the user's computer; (6) limited sharing as to limited subsets of identified regions in topic space; (7) limited sharing as to limited subsets of identified regions in other Cognitive Attention Receiving Spaces (CARs); (8) limited sharing based on specified blockings of identified points, nodes or regions (PNOS's) in topic space and/or other Cognitive Attention Receiving Spaces; (9) limited sharing based on the Layer-Vator™ (113) being stationed at one of one or more prespecified Layer-Vator™ floors, (10) limited sharing as to limited subsets of user-context identified by the user, and so on. If a given second user has not authorized sharing out of his attribute statistics, such blocked statistics will be displayed as faded out, grayed out screen areas or otherwise indicated as not available areas on the radar icons column (e.g., 101 ra′ of FIG. 1B) of the watching first user. Additionally, if a given second user is currently off-line, the “Now” face (e.g., 101 t′ of FIG. 1B) of the radar icon (e.g., pyramid) of that second user may be dimmed, dashed, grayed out, etc. to indicate the second social entity is not online. If the given second user was off-line during the time period (e.g., 3 Hours Ago) specified by the second face 101 x′ of the radar icon (e.g., pyramid) of that second user, such second face 101 x′ will be grayed out. Accordingly, the first user may quickly tell whom among his friends and family (or other associated social entities) was online when (if sharing of such information is permitted by those others) and what interrelated topics (or other types of points, nodes or subregions) they were focused-upon during the corresponding time period (e.g., Now versus 3 Hrs. Ago). In one embodiment, an encoded time graph may be provided showing for example that the other social entity was offline for 30 minutes of the last 90 minute interval of today and offline for 45 minutes of a 4 hour interval of the previous day. Such addition information may be useful in indicating to the first user, how in tune the second social entity probably is with regard to current events that unfolded in the last hour or last few days. If a second user does not want to share out information about when he/she is online or off, no pyramid (or other radar object) will be displayed for that second user to other users. (Or if the second user is a member of group whose group dynamics are being tracked by a radar object, that second user will be treated as if he or she not then participating in the group, in other words, as if he/she is offline because he/she does not want to then share.) If a pyramid is a group representing one, it can show an indicator that four out of nine people are online, for example by providing on the bottom of the pyramid a line graph like the following that indicates 4 people online, 5 people offline: (4on/5off):
    Figure US20190109810A1-20190411-P00001
    |x x x x x”. If desired, the graphs can be more detailed to show how long and/or with what emotional intensities the various online or offline entities are/were online and/or for how long they in their current offline state.
  • Not all of FIG. 4A has been described thus far. That is because there are many different aspects. This disclosure will be ping ponging between FIGS. 1A and 4A as the interrelation between them warrants. With regard to FIG. 4A, it has already been discussed that a given first user (431) may develop a wide variety of user-to-user associations and corresponding U2U records 411 will be stored in the system based on social networking activities carried out within the STAN_3 system 410 and/or within external platforms (e.g., 441, 442, etc.). Also the real person user 431 may elect to have many and differently identified social personas for himself which personas are exclusive to, or cross over as between two or more social networking (SN) platforms. For example, the user 431 may, while interacting only with the MySpace™ platform 442 choose to operate under an alternate ID and/or persona 431 u 2—i.e. “Stewart” instead of “Stan” and when that persona operates within the domain of external platform 442, that “Stewart” persona may develop various user-to-topic associations (U2T) that are different than those developed when operating as “Stan” and under the usage monitoring auspices of the STAN_3 system 410. Also, topic-to-topic associations (T2T), if they exist at all and are operative within the context of the alternate SN system (e.g., 442) may be different from those that at the same time have developed inside the STAN_3 system 410. Additionally, topic-to-content associations (T2C, see block 414) that are operative within the context of the alternate SN system 442 may be nonexistent or different from those that at the same time have developed inside the STAN_3 system 410. Yet further, Context-to-other attribute(s) associations (L2/(U/T/C), see block 416) that are operative within the context of the alternate SN system 442 may be nonexistent or different from those that at the same time have developed inside the STAN_3 system 410. It can be desirable in the context of the present disclosure to import at least subsets of user-to-user association records (U2U) developed within the external platforms (e.g., FaceBook™ 441, LinkedIn™ 444, etc.) into a user-to-user associations (U2U) defining database section 411 maintained by the STAN_3 system 410 so that automated topic tracking operations such as the briefly described one of columns 101 and 101 r of FIG. 1A can take place while referencing the externally-developed user-to-user associations (U2U). Aside from having the STAN_3 system maintain a user-to-user associations (U2U) data-objects organizing space and a user-to-topic associations (U2T) data-objects organizing space, it is within the contemplation of the present disclosure to maintain a user-to-physical locations associations (U2L) data-objects organizing space and a user-to-events associations (U2E) data-objects organizing space. The user-to-physical locations associations (U2L) space may indicate which users are expected to be at respective physical locations during respective times of day or respective days of the week, month, etc. One use for this U2L space is that of determining user context. More specifically, if a particular one or more users are not at their usual expected locations, that may be used by the system to flag an out-of-normal context. The user-to-events associations (U2E) may indicate which users are expected to be at respective events (e.g., social gatherings) during respective times of day or respective days of the week, month, etc. One use for this U2E space is that of determining user context. More specifically, if a particular one or more users are not at their usual expected events, that may be used by the system to flag an out-of-normal context. Yet more specifically, in the above given example where the system flagged the Superbowl™ Sunday Party attendee that “This is the kind of party that your friends A) Henry and B) Charlie would like to be at”, the U2E space may have been consulted to automatically determine that two usual party attendees are not there and to thereby determine that maybe the third user should message to them that they are “sorely missed”.
  • The word “context” is used herein to mean several different things within this disclosure. Unfortunately, the English language does not offer many alternatives for expressing the plural semantic possibilities for “context” and thus its meaning must be determined based on; please forgive the circular definition, its context. One of the meanings ascribed herein for “context” is to describe a role assigned to or undertaken by an actor and the expectations that come with that role assignment. More specifically, when a person is in the context of being “at work”, there are certain presumed “roles” assigned to that actor while he or she is deemed to be operating within the context of that “at work” activity. More particularly, a given actor may be assigned to the formal role of being Vice President of Social Media Research and Development at a particular company and there may be a formal definition of expected performances to be carried out by the actor when in that role (e.g., directing subordinates within the company's Social Media Research and Development Department). Similarly, the activity (e.g., being a VP while “at work”) may have a formal definition of expected subactivities. At the same time, the formal role may be a subterfuge for other expected or undertaken roles and activities because everybody tends to be called “Vice President” for example in modern companies while that formal designation is not the true “role”. So there can be informal role definitions and informal activity definitions as well as formal ones. Moreover, a person can be carrying out several roles at one time and thus operating within overlapping contexts. More specifically, while “at work”, the VP of Social Media R&D may drop into an online chat room where he has the role of active room moderator and there he may encounter some of the subordinates in his company's Social Media R&D Dept. also participating within that forum. At that time, the person may have dual roles of being their boss in real life (ReL) and also being room moderator over their virtual activities within the chat room. Accordingly, the simple term “context” can very quickly become complex and its meanings may have to be determined based on existing circumstances (another way of saying context). Other meanings for the term context as used herein can include, but are not limited to unless specifically so-stated: (1) historical context which is based on what memories the user currently has of past attention giving activities; (2) social dynamics context which is based on what other social entities the given user is, or believes him/herself to be in current social interaction with; (3) physical context which is based on what physical objects the given user is, or believes him/herself to be in current proximity with; and (4) cognitive state context, which here, is a catch-all term for other states of cognition that may affect what the user is currently giving significant energies of cognition to or recalling having given significant energies of cognition to, where the other states of cognition may include attributes such as, but not limited to, things sensed by the 5 senses, emotional states such as: fear, anxiety, aloofness, attentiveness, happy, sad, angry and so on; cognitions about other people, about geographic locations and/or places in time (in history); about keywords; about topics and so on.
  • One addition provided by the STAN_3 system 410 disclosed here is the database portion 416 which provides “Context” based associations and hybrid context-to-other space(s) associations. More specifically, these can be Location-to-User and/or Location-to-Topic and/or Location-to-Content and/or Place-in-Time-to-Other-Thing associations. The context; if it is location-based for example, can be a real life (ReL) geographic one and/or a virtual one of where the real life (ReL) or virtual user is deemed by the system to be located. Alternatively or additionally, the context can be indicative of what type of Social-Topical situation the user is determined by the machine system to be in, for example: “at work”, “at a party”, at a work-related party, in the school library, etc. The context can alternatively or additionally be indicative of a temporal range (place-in-time) in which the user is situated, such as: time of day, day of week, date within month or year, special holiday versus normal day and so on. Alternatively or additionally, the context can be indicative of a sequence of events that have and/or are expected to happen such as: a current location being part of a sequence of locations the user habitually or routinely traverses through during for example, a normal work day and/or a sequence of activities and/or social contexts the user habitually or routinely traverses through during for example, a normal weekend day (e.g., IF Current Location/Activity=Filling up car at Gas Station X, THEN Next Expected Location/Activity=Passing Car through Car Wash Line at same Gas Station X in next 20 minutes). Moreover, context can add increased definition to points, nodes or subregions in other Cognitive Attention Receiving Spaces; thus defining so-called, hybrid spaces, points, nodes or subregions; including for example IF Context Role=at work and functioning as receptionist AND keyword=“meeting” THEN Hybrid ContextualTopic#1=Signing in and Directing new arrivals to Meeting Room. Much more will be said herein regarding “context”. It is a complex subject.
  • For now it is sufficient to appreciate that database records (e.g., hierarchically organized context nodes and links which connect them to other nodes) in this new section 416 can indicate for the machine system, context related associations (e.g., location and/or time related associations) including, but not limited to, (1) when an identified social entity (e.g., first user) is present (virtually or in real life) at a given location as well as within a cross-correlated time period, and that the following one or more topics (e.g., T1, T2, T3, etc.) are likely to be associated with that location, that time and/or a role that the social entity is deemed by the machine system to probably be engaged in due to being in the given “context’ or circumstances; (2) when a first user is disposed at a given location as well as within a cross-correlated time period, then the following one or more additional social entities (users) are likely to be associated with (e.g., nearby to) the first user: U2, U3, U4, etc.; (3) when a first user is disposed at a given location as well as within a cross-correlated time period, then the following one or more content items are likely to be associated with the first user: C1, C2, C3, etc.; and (4) when a first user is disposed at a given location as well as within a cross-correlated time period, then the following one or more hybrid combinations of social entity, topic, device and content item(s) are likely to be associated with the first user: U2/T2/D2/C2, U3/T2/D4/C4, etc. The context-to-other (e.g., hybrid) association records 416 (e.g., X-to-U/T/C/D association records 416, where X here represents context) may be used to support location-based or otherwise context-based, automated generation of assistance information. In FIG. 4A, box 416 says L-to-U/T/C rather than X-to-U/T/C/D because location is a simple first example of context (X) and thus easier to understand. Incidentally, the “D” in the broader concept of X-to-U/T/C/D stands for Device, meaning user's device. A given user may be automatically deemed to be in a respective different context (X) if he is currently using his hand-held smartphone as opposed to his office desktop computer.
  • Before providing a more concrete example of how a given user (e.g., Stan/Stew 431) may have multiple personas operating in different contexts and how those personas may interact differently based for example on their respective contexts and may form different user-to-user associations (U2U) when operating under their various contexts (currently adopted roles or models) including under the contexts of different social networking (SN) or other platforms, a brief discussion about those possible other SN's or other platforms is provided here. There are many well known dot.COM websites (440) that provide various kinds of social interaction services. The following is a non-exhaustive list: Baidu™; Bebo™; Flickr™; Friendster™; Google Buzz™; Google+™ (a.k.a. Google Plus™), Habbo™, hi5™; LinkedIn™; LiveJournal™; MySpace™; NetLog™; Ning™, Orkut™; PearlTrees™, Qzone™, Squidoo™, Twitter™; XING™; and Yelp™.
  • One of the currently most well known and used ones of the social networking (SN) platforms is the FaceBook™ system 441 (hereafter also referred to as FB). FB users establish an FB account and set up various permission options that are either “behind the wall” and thus relatively private or are “on the wall” and thus viewable by any member of the public. Only pre-identified “friends” (e.g., friend-for-the-day, friend-for-the-hour) can look at material “behind the wall”. FB users can manually “de-friend” and “re-friend” people depending on who they want to let in on a given day or other time period to the more private material behind their wall.
  • Another well known SN site is MySpace™ (442) and it is somewhat similar to FB. A third SN platform that has gained popularity amongst so-called “professionals” is the LinkedIn™ platform (444). LinkedIn™ users post a public “Profile” of themselves which typically appears like a resume and publicizes their professional credentials in various areas of professional activity. LinkedIn™ users can form networks of linked-to other professionals. The system automatically keeps track of who is linked to whom and how many degrees of linking separation, if any, are between people who appear to the LinkedIn™ system to be strangers to each other because they are not directly linked to one another. LinkedIn™ users can create Discussion Groups and then invite various people to join those Discussion Groups. Online discussions within those created Discussion Groups can be monitored (censored) or not monitored by the creator (owner) of the Discussion Group. For some Discussion Groups (private discussion groups), an individual has to be pre-accepted into the Group (for example, accepted by the Group moderator) before the individual can see what is being discussed behind the wall of the members-only Discussion Group or can contribute to it. For other Discussion Groups (open discussion groups), the group discussion transcripts are open to the public even if not everyone can post a comment into the discussion. Accordingly, as is the case with “behind the wall” conversations in FaceBook™, Group Discussions within LinkedIn™ may not be viewable to relative “strangers” who have not been accepted as a linked-in friend or as a contact for whom an earlier member of the LinkedIn™ system sort of vouches for by “accepting” them into their inner ring of direct (1st degree of operatively connection) contacts.
  • The Twitter™ system (445) is somewhat different because often, any member of the public can “follow” the “tweets” output by so-called “tweeters”. A “tweet” is conventionally limited to only 140 characters. Twitter™ followers can sign up to automatically receive indications that their favorite (followed) “tweeters” have tweeted something new and then they can look at the output “tweet” without need for any special permissions. Typically, celebrities such as movie stars output many tweets per day and they have groups of fans who regularly follow their tweets. It could be said that the fans of these celebrities consider their followed “tweeters” to be influential persons and thus the fans hang onto every tweeted output sent by their worshipped celebrity (e.g., movie star).
  • The Google™ Corporation (Mountain View, Calif.) provides a number of well known services including their famous online and free to use search engine. They also provide other services such a Google™ controlled Gmail™ service (446) which is roughly similar to many other online email services like those of Yahoo™, EarthLink™, AOL™, Microsoft Outlook™ Email, and so on. The Gmail™ service (446) has a Group Chat function which allows registered members to form chat groups and chat with one another. GoogleWave™ (447) is a project collaboration system that is believed to be still maturing at the time of this writing. Microsoft Outlook™ provides calendaring and collaboration scheduling services whereby a user can propose, declare or accept proposed meetings or other events to be placed on the user's computerized schedule. A much newer social networking service launched very recently by the Google™ Corporation is the Google Plus™ system which includes parts called: “Circles”, “Hangouts”, “Sparks”, and “Huddle”.
  • It is within the contemplation of the present disclosure for the STAN_3 system to periodically import calendaring and/or collaboration/event scheduling data from a user's Microsoft Outlook™ and/or other alike scheduling databases (irrespective of whether those scheduling databases and/or their support software are physically local within a user's computer or they are provided via a computing cloud) if such importation is permitted by the user, so that the STAN_3 system can use such imported scheduling data to infer, at the scheduled dates, what the user's more likely environment and/or contexts are. Yet more specifically, in the introductory example given above, the hypothetical attendant to the “Superbowl™ Sunday Party” may have had his local or cloud-supported scheduling databases pre-scanned by the STAN_3 system 410 so that the latter system 410 could make intelligent guesses as to what the user is later doing, what mood he will probably be in, and optionally, what group offers he may be open to welcoming even if generally that user does not like to receive unsolicited offers.
  • Incidentally, it is within the contemplation of the present disclosure that essentially any database and/or automated service that is hosted in and/or by one or more of a user's physically local data processing devices, or by a website's web serving and/or mirroring servers and data processing parts or all or part of a cloud computing system or equivalent can be used in whole or in part such that it is accessible to the user through one or more physical data processing and/or communicative mechanisms to which the user has access. In other words, even with a relatively small sized and low powered mobile access device, the user can have access to, not only much more powerful computing resources and much larger data storage facilities but also to a virtual community of other people even if each is on the go and thus can only use a mobile interconnection device. The smaller access devices can be made to appear as each had basically borrowed the greater and more powerful resources of cooperatively-connected-to other mechanisms. And in particular, with regard to the here disclosed STAN_3 system, a relatively small sized and low powered mobile access device can be configured to make use of collectively created resources of the STAN_3 system such as so-called, points, nodes or subregions in various Cognitive Attention Receiving Spaces which the STAN_3 system maintains or supports, including but not limited to, topic spaces (TS), keyword spaces (KwS), content spaces (CS), CFi categorizing spaces, context categorizing spaces, and others as shall be detailed below. More to the point, with net-computers, palm-held convergence devices (e.g., iPhone™, iPad™ etc.) and the like, it is usually not of significance where specifically the physical processes of data processing of sensed physical attributes takes place but rather that timely communication and connectivity and multimedia presentation resources are provided so that the user can experience substantially same results irrespective of how the hardware pieces are interconnected and located. Of course, some acts of data acquisition and/or processing may by necessity have to take place at the physical locale of the user such as the acquisition of user responses (e.g., touches on a touch-sensitive tablet screen, IR based pattern recognition of user facial grimaces and eyeball orientations, etc.) and of local user encodings (e.g., what the user's local environment looks, sounds, feels and/or smells like). And also, of course, the user's experience can be limited by the limitations of the multimedia presentation resources (e.g., image displays, sound reproduction devices, etc.) he or she has access to within a given context.
  • Accordingly, the disclosed system cannot bypass the limitations of the input and output resources available to the user. But with that said, even with availability of a relatively small display screen (e.g., one with embedded touch detection capabilities) and/or minimalist audio interface resources, a user can be automatically connected in short order to on-topic and screen compatible and/or audio compatible chat or other forum participation sessions that likely will be directed to a topic the user is apparently currently casting his/her attention toward such that the user can have a socially-enhanced experience because the user no longer feels as if he/she is dealing “alone” with the user's area of current focus but rather that the user has access to other, like-minded and interaction co-compatible people almost anytime the user wants to have such a shared experience. (Incidentally, just because a user's hand-held, local interface device (e.g., smartphone) is itself relatively small in size that does not mean that the user's interface options are limited to screen touch and voice command alone. As mentioned elsewhere herein, the user may wear or carry various additional devices that expand the user's information input/output options, for example by use of an in-mouth, tongue-driven and wirelessly communicative mouth piece whereby the user may signal in privacy, various choices to his hand-held, local interface device (e.g., smartphone).)
  • A more concrete example of context-driven determination of what the user is apparently focusing-upon may take advantage of the digressed-away method of automatically importing a user's scheduling data to thereby infer at the scheduled dates, what the user's more likely environment and/or other context based attributes is/are. Yet more specifically, if the user's scheduling database indicates that next Friday he is scheduled to be at the Social Networking Developers Conference (SNDC, a hypothetical example) and more particularly at events 1, 3 and 7 in that conference at the respective hours of 10:00 AM, 3:00 PM and 7:00 PM, then when that date and a corresponding time segment comes around, the STAN_3 system may use such information in combination with GPS or like location determining information (if available) as part of its gathered, hint or clue-giving encodings for then automatically determining what likely are the user's current situation, mood, surroundings (especially context of the user and of other people interacting with the user), expectations and so forth. For example, between conference events 1 and 3 (and if the user's then active habit profile—see FIG. 5A—indicates as such), the user may be likely to seek out a local lunch venue and to seek out nearby friends and/or colleagues to have lunch with. This is where the STAN_3 system 410 can come into play by automatically providing welcomed “offers” regarding available lunching resources and/or available lunching partners. One welcomed offer might be from a local restaurant which proposes a discount if the user brings 3 of his friends/colleagues. Another such welcomed offer might be from one of his friends who asks, “If you are at SNDC today or near the downtown area around lunch time, do you want to do lunch with me? I want to let you in on my latest hot project.” These are examples of location specific, social-interrelation specific, time specific, and/or topic specific event offers which may pop up on the user's tablet screen 111 (FIG. 1A) for example in topic-related area 104 t (adjacent to on-topic window 117) or in general event offers area 104 (at the bottom tray area of the screen).
  • In order for the system 400 to appear as if it can magically and automatically connect all the right people (e.g., those with concurrent shared areas of focus in a same Cognitions-representing Space and/or those with social interaction co-compatibilities) at the right time for a power lunch in the locale of a business conference they are attending, the system 400 should have access to data that allows the system 400 to: (1) infer the likely moods of the various players (e.g., did each not eat recently and is each in the mood for and/or in the habit or routine a business oriented lunch when in this sort of current context?), (2) infer the current topic(s) of focus most likely on the mind of each individual at the relevant time; (3) infer the type of conversation or other social interaction each individual will most likely desire at the relevant time and place (e.g., a lively debate as between people with opposed view points, or a singing to the choir interaction as between close and like-minded friends and/or family?); (4) infer the type of food or other refreshment or eatery ambiance/decor each invited individual is most likely to agree to (e.g., American cuisine? Beer and pretzels? Chinese take-out? Fine-dining versus fast-food? Other?); (5) infer the distance that each invited individual is likely to be willing to travel away from his/her current location to get to the proposed lunch venue (e.g., Does one of them have to be back on time for a 1:00 PM lecture where they are the guest speaker? Are taxis or mass transit readily available? Is parking a problem?) and so on. See also FIG. 1J of the present disclosure.
  • Since STAN systems such as the ones disclosed in here incorporated U.S. application Ser. No. 12/369,274 and Ser. No. 12/854,082 as well as in the present disclosure are repeatedly testing for, or sensing for, change of user context, of user mood (and thus change of active PEEP and/or other profiles—see also FIG. 3D, part 301 p), the same results produced by mood and context determining algorithms may be used for automatically formulating group invitations based on user mood, user context and so forth. Since STAN systems are also persistently testing for change of current user location or current surroundings (—See also time and location stamps of CFi's as provided Gif. 2A of here incorporated Ser. No. 12/369,274), the same results produced by the repeated user location/context determining algorithms may be used for automatically formulating group invitations based on current user location and/or other current user surroundings information. Since STAN systems are also persistently testing for change of user's current likely topic(s) of focus (and/or current likely other points, nodes or subregions of focus in other Cognitions-representing Spaces), the same results produced by the repeated user's current topic(s) or other-subregions-of-focus determining algorithms may be used for automatically formulating group invitations based on same or similar user topic(s) being currently focused-upon by plural people and determining if there are areas of overlap and/or synergy. (Incidentally, in one embodiment, sameness or similarity as between current topics of focus—and/or sameness or similarity as between current likely other points, nodes or subregions (PNOS) of focus in other Cognitions-representing Spaces is determined at least in part on hierarchical and/or spatial distances between the tested two or more PNOS.) Since STAN systems are also persistently checking their users' scheduling calendars for open time slots and pressing obligations, the same results produced by the repeated schedule-checking algorithms may assist in the automated formulating of group invitations based on open time slots and based on competing other obligations. In other words, much of the underlying data processing is already occurring in the background for the STAN systems to support their primary job of delivering online invitations to STAN users to join on-topic (or other) online forums that appear to be best suited for what the machine system automatically determines to be the more likely topic(s) of current focus and/or other points, nodes or subregions (PNOS) of current focus in other Cognitions-representing Spaces for each monitored user. It is thus a practical extension to add various other types of group offers to the process, where; aside from an invitation to join in for example on an online chat, the various other types of offers can include invitations to join in on real world social interactions (e.g., lunch, dinner, movie, show, bowling, etc.) or to join in on real world or virtual world business oriented ventures (e.g., group discount coupon, group collaboration project).
  • In one embodiment, users are automatically and selectively invited to join in on a system-sponsored game or contest where the number of participants allowed per game or contest is limited to a predetermined maximum number (e.g., 100 contestants or less, 50 or less, 10 or less, or another contest-relevant number). The game or contest may involve one or more prizes and/or recognitions for a corresponding first place winning user or runner up. The prizes may include discount coupons or prize offerings provided by a promoter of specified goods and/or services. In one embodiment, to be eligible for possible invitation to the game or contest (where invitation may also require winning in a final invitations round lottery), the users who wish to be invited (or have a chance of being invited) need to pre-qualify by being involved in one or more pre-specified activities related to the STAN_3 system and/or by having one or more pre-specified user attributes. Examples of such activities/attributes related to the STAN_3 system include, but are not limited to: (1) participating in a chat or other forum participation session that corresponds to a pre-specified topic space subregion (TSR) and/or to a subregion of another system-maintained space (another CARS); (2) participating in adding to or modifying (e.g., editing) within a system-maintained Cognitive Attention Receiving Space (CARS, e.g., topic space), one or more points, nodes or subregions of that space; (3) volunteering to perform other pre-specified services that may be beneficial to the community of users who utilize the STAN_3 system; (4) having a pre-specified set of credentials that indicate expertise or other special disposition relative to a corresponding topic in the system-maintained topic space and/or relative to other pre-specified points, nodes or subregions of other system-maintained CARS's and agreeing to make oneself available for at least a pre-specified number of invitations and/or queries by other system users in regard to the topic node and/or other such CARS PNOS; (5) satisfying in the user's then active personhood and/or profiles of pre-specified geographic and/or other demographic criteria (e.g., age, gender, income level, highest education level) and agreeing to make oneself available for at least a pre-specified number of invitations and/or queries by other system users in regard to the corresponding demographic attributes, and so on.
  • In one embodiment, user PEEP records (Personal Emotion Expression Profiles) are augmented with user PHAFUEL records (Personal Habits And Favorites/Unfavorites Expression Logs—see FIG. 5A re the latter) which indicate various life style habits and routines of the respective users such as, but not limited to: (1) what types of foods he/she likes to eat, when, in what order and where (e.g., favorite restaurants or restaurant types); (2) what types of sports activities he/she likes to engage in, when, in what order and where (e.g., favorite gym or exercise equipment); (3) what types of non-sport activities he/she likes to engage in, when, in what order and where (e.g., favorite movies, movie houses, theaters, actors, musicians, etc.); (4) what are the usual sleep, eat, work and recreational time patterns of the individuals are (e.g., typically sleeps 11 pm-6am, gym 7-8, then breakfast 8-8:30, followed by work 9-12, 1-5, dinner 7 pm, etc.) during normal work weeks, when on vacation, when on business oriented trips, etc. The combination of such PEEP records and PHAFUEL records can be used to automatically formulate event invitations that are in tune with each individual's life style habits and routines. More specifically, a generic algorithm for generating a meeting promoting invitation based on habits, routines and availability might be of the following form: IF a 30 minute or greater empty time slot coming up AND user is likely to then be hungry AND user is likely to then be in mood for social engagement with like focused other people (e.g., because user has not yet had a socially-fulfilling event today), THEN locate practically-meetable nearby other system users who have an overlapping time slot of 30 minutes of greater AND are also likely to then be hungry and have overlapping food type/venue type preferences AND have overlapping likely desire for socially-fulfilling event, AND have overlapping topics of current focus AND/OR social interaction co-compatibilities with one another; and if at least two such users located, automatically generate lunch meeting proposal for them and send same to them. (In one embodiment, the tongue is used simultaneously as an intentional signaling means and a biological state deducing means. More specifically, the user's local data processing device is configured to respond to the tongue being stuck out to the left and/or right with lips open or closed for example as meaning different things and while the tongue is stuck out, the data processing device takes an IR scan and/or visible spectrum scan of the stuck out tongue to determine various biological states related to tongue physiology including mapping flow of blood along the exposed area of the tongue and determining films covering the tongue and/or moisture state of the tongue (i.e. dried versus moist).)
  • Automated life style planning tools such as the Microsoft Outlook™ product can be used to locate common empty time slots and geographic proximity because tools such as the Microsoft Outlook™ typically provide Tasks tracking functions wherein various to-do items and their criticalities (e.g., flagged as a must-do today, must-do next week, etc.) are recorded. Such data could be stored in a computing cloud or in another remotely accessible data processing system. It is within the contemplation of the present disclosure for the STAN_3 system to periodically import Task tracking data from the user's Microsoft Outlook™ and/or other alike task tracking databases (if permitted by the user, and whether stored in a same cloud or different resource) so that the STAN_3 system can use such imported task tracking data to infer during the scheduled time periods, the user's more likely environment, context, moods, social interaction dispositions, offer welcoming dispositions, etc. The imported task tracking data may also be used to update user PHAFUEL records (Personal Habits And Favorites/Unfavorites Expression Log) which indicate various life style habits of the respective user if the task tracking data historically indicates a change in a given habit or a given routine. More specifically with regard to current user context, if the user's task tracking database indicates that the user has a high priority, high pressure work task to be completed by end of day, the STAN_3 system may use this imported information to deduce that the user would not then likely welcome an unsolicited event offer (e.g., 104 t or 104 a in FIG. 1A) directed to leisure activities for example and instead that the user's mind is most likely sharply focused on topics related to the must-be-done task(s) as their deadlines approach and they are listed as not yet complete. Similarly, the user may have Customer Relations Management (CRM) software that the user regularly employs and the database of such CRM software might provide exportable information (if permitted by the user) about specific persons, projects, etc. that the user will more likely be involved with during certain time periods and/or when present in certain locations. It is within the contemplation of the present disclosure for the STAN_3 system to periodically import CRM tracking data from the user's CRM tracking database(s) (if permitted by the user, and whether such data is stored in a same cloud or different resources) so that the STAN_3 system can use such imported CRM tracking data to, for example, automatically formulate an impromptu lunch proposal for the user and one of his/her customers if they happen to be located close to a nearby restaurant and they both do not have any time pressing other activities to attend to.
  • In one embodiment, the CRM/calendar tool is optionally configured to just indicate to the STAN_3 system when free time is available but to not show all data in CRM/calendar system, thereby preserving user privacy. In an alternate embodiment, the CRM/calendar tool is optionally configured to indicate to the STAN_3 system general location data as well as general time slots of free time thereby preserving user privacy regarding details. Of course, it is also within the contemplation of the present disclosure to provide different levels of access by the STAN_3 system to generalized or detailed information of the CRM/calendar system thereby providing different levels of user privacy. The above described, automated generations and transmissions of suggestions for impromptu lunch proposals and the like may be based on automated assessment of each invitee's current emotional state (as determined by current active PEEP record) for such a proposed event as well as each invitee's current physical availability (e.g., distance from venue and time available and transportation resources). In one embodiment, a first user's palmtop computer (e.g., 199 of FIG. 2) automatically flashes a group invite proposal to that first user such as: “Customers X and Z happen to be nearby and likely to be available for lunch with you, Do you want to formulate a group lunch invitation?”. If the first user clicks, taps or otherwise indicates “Yes”, a corresponding group event offer (e.g., 104 a) soon thereafter pops on the screens of the selected offerees. In one embodiment, the first user's palmtop computer first presents a draft boiler plate template to the first user of the suggested “group lunch invitation” which the first user may then edit or replace with his own before approving its multi-casting to the computer formulated list of invitees (which list the first user can also edit with deletions or additions). In one embodiment, even before proposing a possible lunch meetup to the first user, the STAN_3 system predetermines if a sufficient number of potential lunchmates are similarly available so that likelihood of success exceeds a predetermined probability threshold; and if not the system does not make the suggestion. As a result, when the first user does receive such a system-originated suggestion, its likelihood of success can be made fairly high. By way of example, the STAN_3 system might check to see if at least 3+ people are available first before even sending invitations at all.
  • As a yet better enhancer for likelihood of success, the system originated and corresponding group event offer (e.g., let's have lunch together) may be augmented by adding to it a local merchant's discount advertisement. For example, and with regard to the group event offer (e.g., let's have lunch together) which was instigated by the first user (the one whose CRM database was exploited to this end by the STAN_3 system to thereby automatically suggest the group event to the first user who then acts on the suggestion), that group event offer is automatically augmented by the STAN_3 system 410 to have attached thereto a group discount offer (e.g., “Note that the very nearby Louigie's Italian Restaurant is having a lunch special today”). The augmenting offer from the local food provider automatically attached due to a group opportunity algorithm automatically running in the background of the STAN_3 system 410 and which group opportunity algorithm will be detailed below. Briefly, goods and/or service providers can formulate discount offer templates which they want to have matched by the STAN_3 system with groups of people that are likely to accept the offers. The STAN_3 system 410 then automatically matches the more likely groups of people with the discount offers those people are more likely to accept. It is win-win for both the consumers and the vendors. In one embodiment, after, or while a group is forming for a social gathering plan (in real life and/or online) the STAN_3 system 410 automatically reminds its user members of the original and/or possibly newly evolved and/or added on reasons for the get together. For example, a pop-up reminder may be displayed on a user's screen (e.g., 111) indicating that 70% of the invited people have already accepted and they accepted under the idea that they will be focusing-upon topics T_original, T_added_on, T_substitute, and so on. (Here, T_original can be an initially proposed topic that serves as an initiating basis for having the meeting while T_added_on can be later added topic proposed for the meeting after discussion about having the meeting started.) In the heat of social gatherings, people sometimes forget why they got together in the first place (what was the T_original?). However, the STAN_3 system can automatically remind them and/or additionally provide links to or the actual on-topic content related to the initial or added-on or deleted or modified topics (e.g., T_original, T_added_on, T_deleted, etc.)
  • More specifically and referring to FIG. 1A, in one hypothetical example, a group of social entities (e.g., real persons) have assembled in real life (ReL) and/or online with the original intent of discussing a book they have been reading because most of them are members of the Mystery-History e-book of the month club (where the e-book can be an Amazon Kindle™ compatible electronic book and/or another electronically formatted and user accessible book). However, some other topic is brought up first by one of the members and this takes the group off track. To counter this possibility, the STAN_3 system 410 can post a flashing, high urgency invitation 102 m in top tray area 102 of the displayed screen 111 of FIG. 1A that reminds one or more of the users about the originally intended topic of focus.
  • In response, one of the group members notices the flashing (and optionally red colored) circle 102 m on front plate 102 a_Now of his tablet computer 100 and double clicks or taps the dot 102 m open. In response to such activation, his computer 100 displays a forward expanding connection line 115 a 6 whose advancing end (at this stage) eventually stops and opens up into a previously not displayed, on-topic content window 117 (having an image 117 a of the book included therein). As seen in FIG. 1A, the on-topic content window 117 has an on-topic URL named as www.URL.com/A4 where URL.com represents a hypothetical source location for the in-window content and A4 represents a hypothetical code for the original topic that the group had initially agreed to meet for (as well as meeting for example to have coffee and/or other foods or beverages). In this case, the opened window 117 is HTML coded and it includes two HTML headers (not shown): <H2>Mystery History Online Book Club</H2> and <H3>This Month's Selection: Sherlock Holmes and the Franz Ferdinand Case</H3>. These are two embedded hints or clues that the STAN_3 system 410 may have used to determine that the content in window 117 is on-topic with a topic center in its topic space (413) which is identified by for example, the code name A4. (It is alternatively or additionally within the contemplation of the disclosure that the responsively opened content frame, e.g., 117, be coded with or include XML and XML tags and/or codes and tags of other markup languages.) Other embedded hints or clues that the STAN_3 system 410 may have used include explicit keywords (e.g., 115 a 7) in text within the window 117 and buried (not seen by the user) meta-tags embedded within an in-frame image 117 a provided by the content sourced from source location www.URL.com/A4 (an example). This reminds the group member of the topic the group originally gathered to discuss. It doesn't mean the member or group is required to discuss that topic. It is merely a reminder. The group member may elect to simply close the opened window 117 (e.g., activating the X box in the upper right corner) and thereafter ignore it. Dot 102 m then stops flashing and eventually fades away or moves out of sight. In the same or an alternate embodiment, the reminder may come in the form of a short reminder phrase (e.g., “Main Meetg Topic=Book of the Month”). (Note: the references 102 a_Now and 102 aNow are used interchangeably herein.)
  • In one embodiment, after passage of a predetermined amount of time the My Top-5 Topics Now serving plate, 102 a_Now automatically transforms into a My Top-5 Topics Earlier serving plate, 102 a′_Earlier which is covered up by a slightly translucent but newer and more up to date, My Top Topics Now serving plate, 102 a_Now. In the case where Tower-of-Hanoi stacked rings are used in an inverted cone orientation, the smaller, older ones of the top plate can leak through to the “Earlier” in time plate 102 a′_Earlier where they again become larger and top of the stack rings because in that “Earlier” time frame they are the newest and best invitations and/or recommendations. If, after such an update, the user wants to see the older, My Top Topics Earlier plate 102 a′_Earlier, he may click on, tap, or otherwise activate a protruding-out small portion of that older plate and stacked behind plate. The older plate then pops to the top. Alternatively the user might use other menu means for shuffling the older serving plate to the front. Behind the My Top Topics Earlier serving plate, 102 a′_Earlier there is disposed an even earlier in time serving plate 102 a″ and so on. Invitations (to online and/or real life meetings) that are for a substantially same topic (e.g., book club) line up almost behind one another so that a historical line up of such on-same-topic invitations is perceived when looking through the partly translucent plates. This optional viewing of current and older on-topic invitations is shown for the left side of plates stack 102 b (Their Top 5 Topics). (Note: the references 102 a′_Earlier and 102 a′Earlier are used interchangeably herein.) Incidentally, and as indicated elsewhere herein, the on-topic serving plates, such as those of plate stack 102 b need not be of the meet-up opportunity type, or of the meet-up opportunity only type. The serving plates (e.g., 102 aNow) can alternatively or additionally serve up links to on-topic resources (e.g., content providing resources) other than invitations to chat or other forum participation sessions. The other on-topic resources may include, but not limited to, links to on-topic web sites, links to on-topic books or other such publications, links to on-topic college courses, links to on-topic databases and so on.
  • If the exemplary Book-of the-Month Club member had left window 117 open for more than a predetermined length of time, an on-topic event offering 104 t may have popped open adjacent to the on-topic material of window 117. However, this description of such on-topic promotional offerings has jumped ahead of itself because a broader tour of the user's tablet computer 100 has not yet been supplied here and such a re-tour (return to the main tour) will now be presented.
  • Recall how the Preliminary Introduction above began with a bouncing, rolling ball (108) pulling the user into a virtual elevator (113) that took the user's observed view to a virtual floor of a virtual high rise building. When the doors open on the virtual elevator (113, bottom right corner of screen) the virtual ball (108″) hops out and rolls to the diagonally opposed, left upper corner of the screen 111. This tends to draw the user's eyes to an on-screen context indicator 113 a and to the header entity 101 a of social entities column 101. The user may then note that the header entity has been automatically preset to be “Me”. The user may also note that the on-screen context indicator 113 a indicates the user is currently on a virtual floor named, “My Top 5 Now Topics” (which floor name is not shown in FIG. 1A due to space limitations—the name could temporarily unfurl as the bouncing, rolling ball 108 stops in the upper left screen corner and then could roll back up behind floor/context indicator 113 a as the ball 108 continues to another temporary stopping point 108′). There could be 100s of floors in the virtual building (or other such virtual structure) through which the Layer-Vator™ 113 travels and, in one embodiment, each floor has a respective label or name that is found at least on the floor selection panel inside the Layer-Vator™ 113 and besides or behind (but out-poppable therefrom) the current floor/context indicator 113 a.
  • Before moving on to next stopping point 108′, the virtual ball (also referred to herein as the Magic Marble 108) outputs a virtual spot light from its embedded virtual light sources onto a small topic space flag icon 101 ts sticking up from the “Me” header object 101 a. A balloon icon (not shown) temporarily opens up and displays the guessed-at most prominent (top) topic that the machine system (410) has determined to be the topic likely to be foremost (topmost) in the user's mind. In this example, it says, “Superbowl™ Sunday Party”. The temporary balloon (not shown) collapses and the Magic Marble 108 then shines another virtual spotlight on invitation dot 102 i at the left end of the also-displayed, My Top Topics Now serving plate 102 a_Now. Then the Magic Marble 108 rolls over to the right, optionally stopping at another tour point 108′ to light up, for example, the first listed Top Now Topic for the “Them/Their” social entity of plates stack 102 b. Thereafter, the Magic Marble 108 rolls over further to the right side of the screen 111 and parks itself in a ball parking area 108 z. This reminds the user as to where the Magic Marble 108 normally parks. The user may later want to activate the Magic Marble 108 for performing user specified functions (e.g., marking up different areas of the screen for temporary exclusion from STAN_3 monitoring or specific inclusion in STAN_3 monitoring where all other areas are automatically excluded).
  • Unseen by the user during this exercise (wherein the Magic Marble 108 is rolling diagonally from one corner (113) to the other (113 a) and then across to come to rest in the Ball Park 108 z) is that the user's tablet computer 100 is automatically watching him while he is watching the Magic Marble 108 move to different locations on the screen. Two spaced apart, eye-tracking sensors, 106 and 109, are provided along an upper edge of the exemplary tablet computer 100. (There could be yet more sensors, such as three at three corners.) Another sensor embedded in the computer housing (100) is a GPS one (Global Positioning Satellites receiver, shown to be included in housing area 106). At the beginning of the story (the Preliminary Introduction to Disclosed Subject Matter), the GPS sensor was used by the STAN_3 system 410 to automatically determine that the user is geographically located at the house of one of his known friends (Ken's house). That information in combination with timing and accessible calendaring data (e.g., Microsoft Outlook™) allowed the STAN_3 system 410 to automatically determine one or a few most likely contexts for the user and then to extract best-guess conclusions that the user is now likely attending the “Superbowl™ Sunday Party” at his friend's house (Ken's), perhaps in the context role of being a “guest”. The determined user context (or most likely handful of contexts) similarly provided the system 410 with the ability to draw best-guess conclusions that the user would soon welcome an unsolicited Group Coupon offering 104 a for fresh hot pizza. But again the story given here is leap-frogging ahead of itself. The guessed at, social context of being at “Ken's Superbowl™ Sunday Party” also allowed the system 410 to pre-formulate the layout of the virtual floor displayed by way of screen 111 as is illustrated in FIG. 1A. That predetermined layout includes the specifics of who (what persona or group) is listed as the header social entity 101 a (KoH=“Me”) at the top of left side column 101 and who or what groups are listed as follower social entities 101 b, 101 c, . . . , 101 d below the header social entity (KoH) 101 a. (In one embodiment, the initial sequence of listing of the follower social entities 101 b, 101 c, . . . , 101 d is established by a predetermined sorting algorithm such as which follower entity has greatest commonality of heat levels applied to same currently focused-upon topics as does the header social entity 101 a (KoH=“Me”). In an alternate embodiment, the sorted positionings of the follower social entities 101 b, 101 c, . . . , 101 d may be established based on an urgency determining algorithm; for example one that determines there are certain higher and lower priority projects that are respectively cross-associated as between the KoH entity (e.g., “Me”) and the respective follower social entities 101 b, 101 c, . . . , 101 d. Additionally or alternatively, the sorting algorithm can use some other criteria (e.g., current or future importance of relationship between KoH and the others) to determine relative positionings along vertical column 101. That initially pre-sorted sequence can be altered by the user, for example with use of a shuffle up tool 98+. The predetermined floor layout also includes the specifics of what types of corresponding radar objects (101 ra, 101 rb, . . . , 101 rd) will be displayed in the radar objects holding column 101 r. It also determines which invitations/suggestions serving plates, 102 a, 102 b, etc. (where here 102 a is understood to reference the plates stack that includes serving plate 102 aNow as well as those behind it) are displayed in the top and retractable, invitations serving tray 102 provided near an edge of the screen 111. It also determines which associated platforms will be listed in a right side, playgrounds holding column 103 and in what sequence. In one embodiment, when a particular one or more invitations and/or on-topic suggestions (e.g., 102 i) is/are determined by the STAN_3 system to be directed to an online forum or real life (ReL) gathering associated with a specific platform (e.g., FaceBook™, LinkedIn™ etc.), then; at a time when the user hovers a cursor or other indicator over the invitation(s) (e.g., 102 i) or otherwise inquires about the invitations (e.g., 102 i; or associated content suggestions), the corresponding platform representing icon in column 103 (e.g., FB 103 b in the case of an invitation linked thereto by linkage showing-line 103 k) will automatically glow and/or otherwise indicate the logical linkage relationship between the platform and the queried invitation or machine-made suggestion. The predetermined layout shown in FIG. 1A may also determine which pre-associated event offers (104 a, 104 b) will be initially displayed in a bottom and retractable, offers serving tray 104 provided near the bottom edge of the screen 111. Each such serving tray or side-column/row may include a minimize or hide command mechanism. For sake of illustration, FIG. 1A shows Hide buttons such as 102 z of the top tray 102 for allowing the user to minimize or hide away any one or more respective ones of the automatically displayed trays: 101, 101 r, 102, 103 and 104. In one embodiment, even when metaphorically “hidden” beyond the edge of the screen, exceptionally urgent invitations or recommendations will protrude slightly into the screen from the edge to thereby alert the user to the presence of the exceptionally urgent (e.g., highly scored and above a threshold) invitation or recommendation. Of course, other types of hide/minimize/resize mechanisms may be provided, including more detailed control options in the Format drop down menu of toolbar 111 a.
  • The display screen 111 may be a Liquid Crystal Display (LCD) type or an electrophoretic type or another as may be appropriate. The display screen 111 may accordingly include a matrix of pixel units embedded therein for outputting and/or reflecting differently colored visible wavelengths of light (e.g., Red, Green, Blue and White pixels) that cause the user (see 201A of FIG. 2) to perceive a two-dimensional (2D) and/or three-dimensional (3D) image being projected to him. The display screens 111, 211 of respective FIGS. 1A and 2 also have a matrix of infra red (IR) wavelength detectors embedded therein, for example between the visible light outputting pixels. In FIG. 1A, only an exemplary one such IR detector is indicated to be disposed at point 111 b of the screen and is shown as magnified to include one or more photodetectors responsive to wavelengths output by IR beam flashers 106 and 109. The IR beam flashers, 106 and 109, alternatingly output patterns of IR light that can reflect off of a user's face (including off his eyeballs) and can then bounce back to be seen (detected and captured) by the matrix of IR detectors (only one shown at 111 b) embedded in the screen 111. The so-captured stereoscopic images (represented as data captured by the IR detectors 111 b) are uploaded to the STAN_3 servers (for example in cloud 410 of FIG. 4A). Before uploading to the STAN_3 servers, some partial data processing on the captured image data (e.g., image clean up and compression) can occur in the client machine, such that less data is pushed to the cloud. The uploaded image data is further processed by data processing resources of the STAN_3 system 410. These resources may include parallel processing digital engines or the like that quickly decipher the captured IR imagery and automatically determine therefrom how far away from the screen 111 the user's face is and/or what specific points on the screen (or sub-portions of the screen) the user's eyeballs are focused upon. The stereoscopic reflections of the user's face, as captured by the in-screen IR sensors may also indicate what facial expressions (e.g., grimaces) the user is making and/or how warm blood is flowing to or leaving different parts of the user's face (including, optionally the user's protruded tongue). The point of focus of the user's eyeballs tells the system 410 what content the user is probably focusing-upon. Point of eyeball focus mapped over time can tell the system 410 what content the user is focusing-upon for longest durations and perhaps reading or thinking about. Facial grimaces, tongue protrusions, head tilts, etc. (as interpreted with aid of the user's currently active PEEP file) can tell the system 410 how the user is probably reacting emotionally to the focused-upon content (e.g., inside window 117). Some facial contortions may represent intentional commands being messaged from the user to the system 410.
  • When earlier, in the introductory story, the Magic Marble 108 bounced around the screen after entering the displayed scene (of FIG. 1A) by taking a ride thereto by way of virtual elevator 113, the system 410 was preconfigured to know where on the screen (e.g., position 108′) the Magic Marble 108 was located. It then used that known position information to calibrate its IRB sensors (106, 109) and/or its IR image detectors (111 b) so as to more accurately determine what angles the user's eyeballs are at as they follow the Magic Marble 108 during its flight. In one embodiment, there are many other virtual floors in the virtual high rise building (or other such structure, not shown) where virtual presence on this other floor may be indicated to the user by the “You are now on this floor” virtual elevator indicator 113 a of FIG. 1A (upper left corner). When virtually transported to a special one of these other floors, the user is presented with a virtual game room filled with virtual pinball game machines and the like. The Magic Marble 108 then serves as a virtual pinball in these games. And the IRB sensors (106, 109) and the IR image detectors (111 b) are calibrated while the user plays these games. In other words, the user is presented with one or more fun activities that call for the user to keep his eyeballs trained on the Magic Marble 108. In the process, the system 410 heuristically or otherwise forms a heuristic mapping between the captured IR reflection patterns (as caught by the IR detectors 111 b) and the probable angle of focus of the user's eyeballs (which should be tracking the Magic Marble 108).
  • Another sensor that the tablet computer 100 may include is a housing directional tilt and/or jiggle sensor 107. This can be in the form of an opto-electronically implemented gyroscopic sensor and/or MEMs type acceleration sensors and/or a compass sensor. The directional tilt and jiggle sensor 107 determines what angles the flat panel display screen 111 is at relative to gravity and/or relative to geographic North, South, East and West. The tilt and jiggle sensor 107 also determines what directions the tablet computer 100 is being shaken in (e.g., up/down, side to side, Northeast to Southwest or otherwise). The user may elect to use the Magic Marble 108 as a rolling type of cursor (whose action point is defined by a virtual spotlight cast by the internally lit ball 108) and to position the ball with tilt and shake actions applied to the housing of the tablet computer 100. Push and/or rotate actuators 105 and 110 are respectively located on the left and right sides of the tablet housing and these may be activated by the user to invoke pre-programmed functions associated with the Magic Marble 108. In an embodiment the Magic Marble 108 can be moved with a finger or hand gesture. These functions may be varied with a Magic Marble Settings tool 114 provided in a tools area of the screen 111.
  • One of the functions that the Magic Marble 108 (or alternatively a touch driven cursor 135) may provide is that of unfurling a context-based controls setting menu such as the one shown at 136 when the user depresses a control-right keypad or an alike side-bar button combination. (Such hot key combination activation may alternatively or additionally be invoked with special, predetermined facial contortions which are picked up by the embedded IR sensors.) Then, whatever the Magic Marble 108 or cursor 135 (shown disposed inside window 117 of FIG. 1A) or both is/are pointing to, can be highlighted and indicated as activating a user-controllable menu function (136) or set of such functions. In the illustrated example of menu 136, the user has preset the control-right key press function (or another hot key combination activation) to cause two actions to simultaneously happen. First, if there is a pre-associated topic (topic node) already associated with the pointed-to on-screen item, an icon representing the associated topic (e.g., the invitation thereto) will be pointed to. More specifically, if the user moves cursor 135 to point to keyword 115 a 7 inside window 117 (the key.a5 word of phrase), a connector beam 115 a 6 grows backwards from the pointed-to object (key.a5) to a topic-wise associated and already presented invitation and/or suggestion making object (e.g., 102 m) in the top serving tray 102. Second, if there are certain friends or family members or other social entities pre-associated with the pointed-to object (e.g., key.a5) and there are on-screen icons (e.g., 101 a, . . . , 101 d) representing those social entities, the corresponding icons (e.g., 101 a, . . . , 101 d) will glow or otherwise be highlighted. Hence, with a simple hot key combination (e.g., a control right click or a double tap, a multi-finger swipe or a facial contortion), the user can quickly come to appreciate object-to-topic relations and/or object-to-person relations as between a pointed-to on-screen first object (e.g., key.a5 in FIG. 1A) and on-screen other icons that correspond to the topic of, or the associated person(s) of that pointed-to object (e.g., key.a5).
  • Let it be assumed for sake of illustration and as a hypothetical that when the user control-right clicks or double taps on or otherwise activates the key.a5 object, the My Family disc-like icon 101 b glows (or otherwise changes). That indicates to the user that one or more keywords of the key.a5 object are logically linked to the “My Family” social entity. Let it also be assumed that in response to this glowing, the user wants to see more specifically what topics the social entity called “My Family” (101 b) is now primarily focusing-upon (what are their top now N topics?). This cannot be done using the pyramid 101 rb for the illustrated configuration of FIG. 1A because “Me” is the header entity in column 101. That means that all the follower radar objects 101 rb, . . . , 101 rd are following the current top-5 topics of “Me” (101 a) and not the current top N topics of “My Family” (101 b). However, if the user causes the “My Family” icon 101 b to shuffle up into the header (leader, mayor) position of column 101, the social entity known as “My Family” (101 b) then becomes the header entity. Its current top N topics become the lead topics shown in the top most radar object of radar column 101 r. (The “Me” icon may drop to the bottom of column 101 and its adjacent pyramid will now show heat as applied by the “Me” entity to the top N topics of the new header entity, “My Family”.) In one embodiment, the stack of on-topic serving plates called My Current Top Topics 102 a shifts to the right in tray 102 and a new stack of on-topic serving plates called My Family's Current Top Topics (not shown) takes its place as being closest to the upper left corner of the screen 111. This shuffling in and out of entities to/from the top leader position (101 a) can be accomplished with a shuffle Up tool (e.g., 98+ of icon 101 c) provided as part of each social entity icon except that of the leader social entity. Alternatively or additionally, drag and drop may be used.
  • That is one way of discovering what the top N now topics of the “My Family” entity (101 b) are. Another way involves clicking or otherwise activating a flag tool 101 s provided atop the 101 rb pyramid as is shown in the magnified view of pyramid 101 rb in FIG. 1A.
  • In addition to using the topic flag icon (e.g., 101 ts) provided with each pyramid object (e.g., 101 rb), the user may activate yet another topic flag icon that is either already displayed within the corresponding social entity representing object (101 a, . . . , 101 d) or becomes visible when the expansion tool (e.g., starburst+) of that social entity representing object (101 a, . . . , 101 d) is activated. In other words, each social entity representing object (101 a, . . . , 101 d) is provided with a show-me-more details tool like the tool 99+ (e.g., the starburst plus sign) that is for example illustrated in circle 101 d of FIG. 1A. When the user clicks or otherwise activates this show-me-more details tool 99+, one or more pop-out windows, frames and/or menus open up and show additional details and/or addition function options for that social entity representing object (101 a, . . . , 101 d). More specifically, if the show-me-more details tool 99+ of circle 101 d had been activated, a wider diameter circle 101 dd spreads out (in one embodiment) from under the first circle 101 d. Clicking or otherwise activating one area of the wider diameter circle 101 dd causes a greater details pane 101 de (for example) to pop up on the screen 111. The greater details pane 101 de may show a degrees of separation value used by the system 410 for defining a user-to-user association (U2U) between the header entity (101 a) and the expanded entity (101 d, e.g., “him”). The degrees of separation value may indicate how many branches in a hierarchical tree structure of a corresponding U2U association space separate the two users. Alternatively or additionally (but not shown in FIG. 1A), a relative or absolute distance of separation value may be displayed as between two or more user-representing icons (me and him) where the displayed separation value indicates in relative or absolute terms, virtual distances (traveled along a hierarchical tree structure or traveled as point-to-point) that separate the two or more users in the corresponding U2U association space. The greater details pane 101 de may show flags (F1, F2, etc.) for common topic nodes or subregions as between the represented Me-and-Him social entities and the platforms (those of column 103), P1, P2, etc. from which those topic centers spring. Clicking or otherwise activating one of the flags (F1, F2, etc.) opens up more detailed information about the corresponding topic nodes or subregions. For example, the additional detailed information may provide a relative or absolute distance of separation value representing corresponding distance(s) as between two or more currently focused-upon topic nodes of a corresponding two or more social entities. The provided relative or absolute distance of separation value(s) may be used to determine how close to one another or not (how similar to one another or not) are the respectively focused-upon topic nodes when considered in accordance with their respective hierarchical and/or spatial placements in a system-maintained topic space. It is moreover within the contemplation of the present disclosure that closeness to one another or similarity (versus being far apart or highly dissimilar) may be indicated for two or more of respective points, nodes or subregions (PNOS) in any of the Cognitions-representing Spaces described herein. That aspect will be explained in more detail below.
  • By clicking or otherwise activating one of the platform icons (P1, P2, etc.) of greater details pane 101 de, such action opens up more detailed information about where in the corresponding platform (e.g., FaceBook™, STAN3™, etc.) the corresponding topic nodes or subregions logically link to. Although not shown in the exemplary greater details pane 101 de, yet further icons may appear therein that, upon activation, reveal more details regarding points, nodes or subregions (PNOS's) in other Cognitive Attention Receiving Spaces such as keyword space (KwS), URL space, context space (XS) and so on. And as mentioned above, some of the revealed more details can indicate how similar or dissimilar various PNOS's are in their respective Cognitions-representing Spaces. More specifically, cross-correlation details as between the current KoH entity (e.g., “Me”) and the other detailed social entity (e.g., “My Other” 101 d) may include indicating what common or similar keywords or content sub-portions both social entities are currently focusing significant “heat” upon or are otherwise casting their attention on. These common keywords (as defined by corresponding objects in keyword space) may be indicated by other indicators in place of the “heat” indicators. For example, rather than showing the “heat” metrics, the system may instead display the top 5 currently focused-upon keywords that the two social entities have in common with each other. In addition to or as an alternative to showing commonly shared topic points, nodes or subregions and/or commonly shared keyword points, nodes or subregions, or how similar they are, the greater details pane 101 de may show commonalities/similarities in other Cognitive Attention Receiving Spaces such as, but not limited to, URL space, meta-tag space, context space, geography space, social dynamics space and so on. In addition to or as an alternative to comparatively showing commonly shared points, nodes or subregions in various Cognitive Attention Receiving Spaces (CARS's) which are common to two or more social entities, the greater details pane 101 de may show the top N points, nodes or subregions of just one social entity and the corresponding “heats” cast by that just one social entity (e.g., “Me”) on the respective points, nodes or subregions in respective ones of different Cognitive Attention Receiving Spaces (CARS's; e.g., topic space, URL space, ERL space (defined below), hybrid keyword-context space, and so on).
  • Aside from causing a user-selected hot key combination (e.g., control right click or double tap) to provide more detailed information about one or more of associated topic and associated social entities (e.g., friends), the settings menu 136 may be programmed to cause the user-selected hot key combination to provide more detailed information about one or more of other logically-associated objects, such as, but not limited to, associated forum supporting mechanisms (e.g., platforms 103) and associated group events (e.g., professional conference, lunch date, etc.) and/or invitations thereto and/or promotional offerings related thereto.
  • While a few specific sensors and/or their locations in the tablet computer 100 have been described thus far, it is within the contemplation of the present disclosure for the user-proximate computer 100 to have other or additional sensors. For example, a second display screen with embedded IR sensors and/or touch or proximity sensors may be provided on the other side (back side) of the same tablet housing 100. In addition to or as replacement for the IR beam units, 106 and 109, stereoscopic cameras may be provided in spaced apart relation to look back at the user's face and/or eyeballs and/or to look forward at a scene the user is also looking at. The stereoscopic cameras may be used for creating a 3-dimensional of the user (e.g., of the user's face, including eyeballs) so that the system can determine therefrom what the user is currently focused-upon and/or how the user is reacting to the focused-upon material.
  • More specifically, in the case of FIG. 2, the illustrated palmtop computer 199 may have its forward pointing camera 210 pointed at a real life (ReL) object such a Ken's house 198 (e.g., located on the North side of Technology Boulevard) and/or a person (e.g., Ken). Object recognition software provided by the STAN_3 system 410 and/or by one or more external platforms (e.g., GoogleGoggles™ or IQ_Engine™) may automatically identify the pointed-at real life object (e.g., Ken's house 198). Alternatively or additionally, item 210 may represent a forward pointing directional microphone configured to pick up sounds from sound sources other than the user 201A. The picked out sounds may be supplied, in one embodiment, to automated voice recognition software where the latter automatically identifies who is speaking and/or what they are saying. The picked out semantics may include merely a few keywords sufficient to identify a likely topic and/or a likely context. The voice based identification of who is speaking may also be used for assisting in the automated determination of the user's likely context. Yet alternatively or additionally, the forward pointing directional microphone (210) may pick up music and/or other sounds or noises where the latter are also automatically submitted to system sound identifying means for the purpose of assisting in the automated determination of the user's likely context. For example, a detection of carousel music in combination with GPS or alike based location identifying operations of the system may indicate the user is in a shopping mall near its carousel area. As an alternative, the directional sound pick up means may be embedded in nearby other machine means and the output(s) of such directional sound pick up means may be wirelessly acquired by the user's mobile device (e.g., 199).
  • Aside from GPS-like location identifying means and/or directional sound pick up means being embedded in the user's mobile device (e.g., 199) or being available in, and accessed by way of, nearby other devices and being temporarily borrowed for use by the user's mobile device (e.g., 199), the user's mobile device may include direction determining means (e.g., compass means and gravity tilt means) and/or focal distance determining means for automatically determining what direction(s) one or more of used cameras/directional microphones (e.g., 210) are pointing to and where (how far out) the focal point is of the directed camera(s)/microphones relative to the location of the of camera(s)/microphones. The automatically determined identity, direction and distance and up/down disposition of the pointed to object/person (e.g., 198) is then fed to a reality augmenting server within the STAN_3 system 410. The reality augmenting server (not explicitly shown, but one of the data processing resources in the cloud) automatically looks up most likely identity of the person(s) (based for example on automated face and/or voice recognition operations carried out by the cloud), most likely context(s) and/or topic(s) (and/or other points, nodes or subregions of other spaces) that are cross-associated as between the user (or other entity) and the pointed-at real life object/person (e.g., Ken's house 198/Ken). For example, one context plus topic-related invitation that may pop up on the user's augmented reality side (screen 211) may be something like: “This is where Ken's Superbowl™ Sunday Party will take place next week. Please RSVP now.” Alternatively, the user's augmented reality or augmented virtuality side of the display may suggest something like: “There is Ken in the real life or in a recently inloaded image and by the way you should soon RSVP to Ken's invitation to his Superbowl™ Sunday Party”. These are examples of context and/or topic space augmented presentations of reality and/or of a virtuality. The user is automatically reminded of likely topics of current interest (and/or of other focused-upon points, nodes or subregions of likely current interest in other spaces) that are associated with real life (ReL) objects/persons that the user aims his computer (e.g., 100, 199) at or associated with recognizable objects/persons present in recent images inloaded into the user's device.
  • As another example, the user may point at the refrigerator in his kitchen and the system 410 invites him to formulate a list of food items needed for next week's party. The user may point at the local supermarket as he passes by (or the GPS sensor 106 detects its proximity) and the system 410 invites him to look at a list of items on a recent to-be-shopped-for list. This is another example of topic and context spaces based augmenting of local reality. So just by way of recap here, it becomes possible for the STAN_3 system to know/guess on what objects and/or which persons are being currently pointed at by one or more cameras/microphones under control of, or being controlled on behalf of a given user (e.g., 210A of FIG. 2) by combining local GPS or GPS-like functionalities with one or more of directional camera pickups, directional microphone pickups, compass functionalities, gravity angle functionalities, distance functionalities and pre-recorded photograph and/or voice recognition functionalities (e.g., an earlier taken picture of Ken and/or his house in which Ken and house are tagged plus an earlier recorded speech sample taken from Ken) where the combined functionalities increase the likelihood that the STAN_3 system will correctly recognize the pointed-to object (198) as being Ken's house (in this example) and the pointed-to person is Ken (in this example). Alternatively or additionally a cruder form of object/person recognition may be used. For example, the system automatically performs the following: 1) identifying the object in camera as a standard “house”, 2) using GPS coordinates and using a compass function to determine which “house” on an accessible map the camera is pointing, 3) using a lookup table to determine which person(s) and/or events or activities are associated with the so-identified “house”, and 4) using the system's topic space and/or other space lookup functions to determine what topics and/or other points, nodes or subregions are most likely currently associated with the pointed at object (or pointed at person).
  • Yet other sensors that may be embedded in the tablet computer 100 and/or other devices (e.g., head piece 201 b of FIG. 2) adjacent to the user include sound detectors that operate outside the normal human hearing frequency ranges, light detectors that operate outside the normal human visibility wavelength ranges, further IR beam emitters and odor detectors (e.g., 226 in FIG. 2). The sounds, lights and/or odor detectors may be used by the STAN_3 system 410 for automatically determining various current events such as, when the user is eating, duration of eating, number of bites or chewings taken, what the user is eating (e.g., based on odor 227 and/or IR readings of bar code information) and for estimating how much the user is eating based on duration of eating and/or counted chews, etc. Later, (e.g., 3-4 hours later) the system 410 may use the earlier collected information to automatically determine that the user is likely getting hungry again. That could be one way that the system of the Preliminary Introduction knows that a group coupon offer from the local pizza store would likely be “welcomed” by the user at a given time and in a given context (Ken's Superbowl™ Sunday Party) even though the solicitation was not explicitly pulled by the user. The system 410 may have collected enough information to know that the user has not eaten pizza in the last 24 hours (otherwise, he may be tired of it) and that the user's last meal was small one 4 hours ago meaning he is likely getting hungry now. The system 410 may have collected similar information about other STAN users at the party to know that they too are likely to welcome a group offer for pizza at this time. Hence there is a good likelihood that all involved will find the unsolicited coupon offer to be a welcomed one rather than an annoying and somewhat overly “pushy” one.
  • In the STAN_3 system 410 of FIG. 4A, there is provided within its ambit (e.g., cloud, and although shown as being outside), a general welcomeness filter 426 and a topic-based hybrid router 427. The general welcomeness filter 426 receives user data 417 that is indicative of what general types of unsolicited offers the corresponding user is likely or not likely to now welcome. More specifically, if the recent user data 417 indicates the user just ate a very large meal, that will usually flag the user as not welcoming an unsolicited current offer involving consumption of more food. If the recent user data 417 indicates the user just finished a long business oriented meeting, that will usually flag the user as not welcoming an unsolicited offer for another business oriented meeting. (In one embodiment, stored knowledge base rules may be used to automatically determine if an unsolicited offer for another business oriented meeting would be welcome or not; such as for example: IF Length_of Last_Meeting>45 Minutes AND Number_Meetings_Done_Today>4 AND Current_Time>6:00 PM THEN Next_Meeting_Offer_Status=Not Welcome, ELSE . . . ) If the recent user data 417 indicates the user just finished a long exercise routine, that will usually flag the user as not likely welcoming an unsolicited offer for another physically strenuous activity although, on the other hand, it may additionally, flag the user as likely welcoming an unsolicited offer for a relaxing social event at a venue that serves drinks. These are just examples and the list can of course go on. In one embodiment, the general welcomeness filter 426 is tied to a so-called PHA_FUEL file of the user's (Personal Habits And Favorites/Unfavorites Expression Log—see FIG. 5A) where the latter will be detailed later below. Briefly, known habits and routines of the user are used to better predict what the user is likely to welcome or not in terms of unsolicited offers when in different contexts (e.g., at work, at home, at a party, etc.). (Note: the references PHA_FUEL and PHAFUEL are used interchangeably herein.)
  • If general welcomeness has been determined by the automated welcomeness filter 426 for certain general types of offers, the identification of the likely welcoming user is forwarded to the hybrid topic-context router 427 for more refined determination of what specific unsolicited offers the user (and current friends) are more likely to accept than others based on one or more of the system determined current topic(s) likely to be currently on his/their minds and current location(s) where he/they are situated and/or other contexts under which the user is currently operating. Although, it is premature at this point in the present description to go into greater detail, later below it will be seen that so-called, hybrid topic-context points, nodes or subregions can be defined by the STAN_3 system in respective hybrid Cognitive Attention Receiving Spaces. The idea is that a user is not just merely hungry (as an example of mood/biological state) and/or currently casting attention on a specific topic, but also that the user has adopted a specific role or job definition (as part of his/her context) that will further determine if a specific promotional offering is now more welcome than others. By way of a more specific example, assume that the hypothetical user (you) of the above Superbowl™ Sunday party example is indeed at Ken's house and the Superbowl™ game is starting and that hypothetical user (you) is worried about how healthy Joe-The-Throw Nebraska is, but also that one tiny additional fact has been left out of the story. The left out fact is that a week before the party, the hypothetical user entered into an agreement (e.g., a contract) with Ken that the hypothetical user will be working as a food serving and trash clean-up worker and not as a social invitee (guest) to the party. In other words, the user has a special “role” that the user is now operating under and that assumed role can significantly change how the user behaves and what promotional offerings would be more welcomed or less unwelcomed than others. Yet more specifically, a promotional offering such as, “Do you want to order emergency carpet cleaning services for tomorrow?” may be more welcomed by the user when in the clean-up crew role but not when in the party guest role. The subject of assumed roles will be detailed further in conjunction with FIG. 3J (the context primitive data structure).
  • In the example above, one or more of various automated mechanisms could have been used by the STAN_3 system to learn that the user is in one role (one adopted context) rather than another. The user may have a task-managing database (e.g., Microsoft Outlook Calendar™) or another form of to-do-list managing software plus associated stored to-do data, or the user may have a client relations management (CRM) tool he regularly uses, or the user may have a social relations management (SRM) tool he regularly uses, or the user may have received a reminder email or other such electronic message (e.g., “Don't forget you have clean-up crew job duty on Sunday”) reminding the user of the job role he has agreed to undertake. The STAN_3 system automatically accesses one or more of these (after access permission has been given) and searches for information relating to assumed, or to-be-assumed roles. Then the STAN_3 system determines probabilities as between possible roles and generates a sorted list with the more probable roles and their respective probability scores at the top of the list; and the system prioritizes accordingly.
  • Assumed roles can determine predicted habits and routines. Predicted habits and routines (see briefly FIG. 5A, the active PHAFUEL profile) can determine what specific promotional offerings would more likely be welcomed or not. In accordance with one aspect of the disclosure, the more probable user context (e.g., assumed role) is used for selectively activating a correspondingly more probable PHAFUEL profile (Personal Habits And Favorites/Unfavorites Expression Log) and then the hybrid topic-context router 427 (FIG. 4A) utilizes data and/or knowledge base rules (KBR's) provided in the activated PHAFUEL profile for determining how to route the identity of the potential offeree (user) to one promotion offering sponsor more so than to another. In other words, the so sorted outputs of the Topic/Other Router 427 are then forwarded to current offer sponsors (e.g., food vendors, paraphernalia vendors, clean up service providers, etc.) who will have their own criteria as to which of the pre-sorted users or user groups will qualify for certain offers and these are applied as further match-making criteria until specific users or user groups have been shuffled into an offerees group that is pre-associated with a group offer they are very likely to accept. The purpose of this welcomeness filtering and routing and shuffling is so that STAN_3 users are not annoyed with unwelcome solicitations and so that offer sponsors are not disappointed with low acceptance rates (or too high of an acceptance rate if alternatively that is one of their goals). More will be detailed about this below. Before moving on and just to recap here, the assumed role that a user has likely undertaken (which is part of user “context”) can influence whom he would want to share a given and shareable experience with (e.g., griping about clean-up crew duty) and also which promotional offerings the user will more likely welcome or not in the assumed role. Filter and router modules 426 and 427 are configured to base their results (in one embodiment) on the determined-as-more-likely-by-the-system roles and corresponding habits/routines of the user. This increases the likelihood that unsolicited promotional offerings will not be unwelcomed.
  • Referring still to FIG. 4A, but returning now to the subject of the out-of-STAN platforms or services contemplated thereby, the StumbleUpon™ system (448) allows its registered users to recommend websites to one another. Users can click or tap or otherwise activate a thumb-up icon to vote for a website they like and can similarly click or tap on a thumb-down icon to indicate they don't like it. The explicitly voted upon websites can be categorized by use of “Tags” which generally are one or two short words to give a rough idea of what the website is about. Similarly, other online websites such as Yelp™ allow its users to rate real world providers of goods and services with number of thumbs-up, or stars, etc. It is within the contemplation of the present disclosure that the STAN_3 system 410 automatically imports (with permission as needed from external platforms or through its own sideline websites) user ratings of other websites, of various restaurants, entertainment venues, etc. where these various user ratings are factored into decisions made by the STAN_3 system 410 as to which vendors (e.g., coupon sponsors) may have their discount offer templates matched with what groups of likely-to-accept STAN users. Data imported from external platforms 44X may include identifications of highly credentialed and/or influential persons (e.g., Tipping Point Persons) that users follow when using the external platforms 44X. In one embodiment, persons or platforms that rate external services and/or goods also post indications of what specific contexts the ratings apply to. The goal is to minimize the number of times that STAN-generated event offers (e.g., 104 t, 104 a in FIG. 1A) invite STAN users to establishments whose services or goods are below a predetermined acceptable level of quality and/or suitability for a given context. In other words, fitness ratings are generated as indicating appropriate quality and/or suitability to corresponding contexts as perceived by the respective user. More specifically, and for example, what is more “fitting and appropriate” for a given context such as informal house party versus formal business event might vary from a budget pizza to Italian cuisine from a 5 star restaurant. While the 5 star restaurant may have more quality, its goods/services might not be most “fit” and appropriate for a given context. By rating goods/services relative to different contexts, the STAN_3 system works to minimize the number of times that unsolicited promotional offerings invite STAN users to establishments whose services or goods are of the wrong kinds (e.g., not acceptable relative to the role or other context under which the user is operating and thus not what the user had in mind). Additionally, the STAN_3 system 410 collects CVi's (implied vote-indicating records) from its users when and while they are agreeing to be so-monitored. It is within the contemplation of the present disclosure to automatically collect CVi's from permitting STAN users during STAN-sponsored group events where the collected CVi's indicate how well or not the STAN users like the event (e.g., the restaurant, the entertainment venue, etc.). Then the collected CVi's are automatically factored into future decisions made by the STAN_3 system 410 as to which vendors may have their discount offer templates matched with what groups of likely-to-accept STAN users and under what contexts. The goal again is to minimize the number of times that STAN-generated event offers (e.g., 104 t, 104 a) invite STAN users to establishments whose services or goods are collectively voted on as being inappropriate, untimely and/or below a predetermined minimum level of acceptable quality and monetary fitness to the gathering and its respective context(s).
  • Additionally, it is within the contemplation of the present disclosure to automatically collect implicit or explicit CVi's from permitting STAN users at the times that unsolicited event offers (e.g., 104 t, 104 a) are popped up on that user's tablet screen (or otherwise presented to the user). An example of an explicit CVi may be a user-activateable flag which is attached to the promotional offering and which indicates, when checked, that this promotional offering was not welcome or worse, should not be present again to the user and/or to others ever or within a specified context. The then-collected CVi's may indicate how welcomed or not welcomed the unsolicited event offers (e.g., 104 t, 104 a) are for that user at the given time and in the given context. The goal is to minimize the number of times that STAN-generated event offers (e.g., 104 t, 104 a) are unwelcomed by the respective user. Neural networks or other heuristically evolving automated models may be automatically developed in the background for better predicting when and under which contexts, various unsolicited event offers will be welcomed or not by the various users of the STAN_3 system 410. Parameters for the over-time developed heuristic models are stored in personal preference records (e.g., habit and routine records, see FIG. 5A) of the respective users and thereafter used by the general welcomeness filter 426 and/or routing module 427 of the system 410 or by like other means to block inappropriate-for-the-context and thus unwelcomed solicitations from being made too often to STAN users. After sufficient training time has passed, users begin to feel as if the system 410 somehow magically knows when and under what circumstances (context) unsolicited event offers (e.g., 104 t, 104 a) will be welcomed and when not. Hence in the above given example of the hypothetical “Superbowl™ Sunday Party”, the STAN_3 system 410 had beforehand developed one or more PHAFUEL records (Personal Habits And Favorites/Unfavorites Expression Profiles) for the given user indicating for example what foods he likes or dislikes under different circumstances (contexts), when he likes to eat lunch, when he is likely to be with a group of other people and so on. The combination of the pre-developed PHAFUEL records and the welcome/unwelcomed heuristics for the unsolicited event offers (e.g., 104 t, 104 a) can be used by the STAN_3 system 410 to know when are likely times and circumstances that such unsolicited event offers will be welcome by the user and what kinds of unsolicited event offers will be welcome or not. More specifically, the PHAFUEL records of respective STAN users can indicate what things the user least likes or hates as well what they normally like and accept for a given circumstance (a.k.a. “context fitness”). So if the user of the above hypothecated “Superbowl™ Sunday Party” hates pizza (or is likely to reject it under current circumstances, e.g., because he just had pizza 2 hours ago) the match between vendor offer and the given user and/or his forming social interaction group will be given a low score and generally will not be presented to the given user and/or his forming social interaction group. Incidentally, active PHAFUEL records for different users may automatically change as a function of time, mood, context, etc. Accordingly, even though a first user may have a currently active PHAFUEL record (Personal Habit Expression Profiles) indicating he now is likely to reject a pizza-related offer; that same first user may have a later activated PHAFUEL record which is activated in another context and when so activated indicates the first user is likely to then accept the pizza-related offer.
  • Referring still to FIG. 4A and more of the out-of-STAN platforms or services contemplated thereby, consider the well known social networking (SN) system reference as the SecondLife™ network (460 a) wherein virtual social entities can be created and caused to engage in social interactions. It is within the contemplation of the present disclosure that the user-to-user associations (U2U) portion 411 of the database of the STAN_3 system 410 can include virtual to real-user associations and/or virtual-to-virtual user associations. A virtual user (e.g., avatar) may be driven by a single online real user or by an online committee of users and even by a combination of real and virtual other users. More specifically, the SecondLife™ network 460 a presents itself to its users as an alternate, virtual landscape in which the users appear as “avatars” (e.g., animated 3D cartoon characters) and they interact with each other as such in the virtual landscape. The SecondLife™ system allows for Non-Player Characters (NPC's) to appear within the SecondLife™ landscape. These are avatars that are not controlled by a real life person but are rather computer controlled automated characters. The avatars of real persons can have interactions within the SecondLife™ landscape with the avatars of the NPC's. It is within the contemplation of the present disclosure that the user-to-user associations (U2U) 411 accessed by the STAN_3 system 410 can include virtual/real-user to NPC associations. Yet more specifically, two or more real persons (or their virtual world counterparts) can have social interactions with a same NPC and it is that commonality of interaction with the same NPC that binds the two or more real persons as having second degree of separation relation with one another. In other words, the user-to-user associations (U2U) 411 supported by the STAN_3 system 410 need not be limited to direct associations between real persons and may additionally include user-to-user-to-user-etc. associations (U3U, U4U etc.) that involve NPC's as intermediaries. A very large number of different kinds of user-to-user associations (U2U) may be defined by the system 410. This will be explored in greater detail below.
  • Aside from these various kinds of social networking (SN) platforms (e.g., 441-448, 460), other social interactions may take place through tweets, email exchanges, list-serve exchanges, comments posted on “blogs”, generalized “in-box” messagings, commonly-shared white-boards or Wikipedia™ like collaboration projects, etc. Various organizations (dot.org's, 450) and content publication institutions (455) may publish content directed to specific topics (e.g., to outdoor nature activities such as those followed by the Field-and-Streams™ magazine) and that content may be freely available to all members of the public or only to subscribers in accordance with subscription policies generated by the various content providers. (With regard to Wikipedia™ like collaboration projects, those skilled in the art will appreciate that the Wikipedia™ collaboration project—for creating and updating a free online encyclopedia—and similar other “Wiki”-spaces or collaboration projects (e.g., Wikinews™, Wikiquote™, Wikimedia™, etc.) typically provide user-editable world-wide-web content. The original Wiki concept of “open editing” for all web users may be modified however by selectively limiting who can edit, who can vote on controversial material and so on. Moreover, a Wiki-like collaboration project, as such term is used further below, need not be limited to content encoded in a form that is compatible with early standardizations of HTML coding (world-wide-web coding) and browsers that allow for viewing and editing of the same. It is within the contemplation of the present disclosure to use Wiki-like collaboration project control software for allowing experts within different topic areas to edit and vote (approvingly or disapprovingly) on structures and links (e.g., hierarchical or otherwise) and linked-to/from other nodes/content providers of topic nodes that are within their field of expertise. More detail will follow below.)
  • Since a user (e.g., 431) of the STAN_3 system 410 may also be a user of one or more of these various other social networking (SN) and/or other content providing platforms (440, 450, 455, 460, etc.) and may form all sorts of user-to-user associations (U2U) with other users of those other platforms, it may be desirous to allow STAN users to import their out-of-STAN U2U associations, in whole or in part (and depending on permissions for such importation) into the user-to-user associations (U2U) database area 411 maintained by the STAN_3 system 410. To this end, a cross-associations importation or messaging system 432 m may be included as part of the software executed by or on behalf of the STAN user's computer (e.g., 100, 199) where the cross-associations importation or messaging system 432 m allows for automated importation or exchange of user-to-user associations (U2U) information as between different platforms. At various times the first user (e.g., 432) may choose to be disconnected from (e.g., not logged-into and/or not monitored by) the STAN_3 system 410 while instead interacting with one or more of the various other social networking (SN) and other content providing platforms (440, 450, 455, 460, etc.) and forming social interaction relations there. Later, a STAN user may wish to keep an eye on the top topics (and/or other top nodes or subregions of non-topic spaces) currently being focused-upon by his “friend” Charlie, where the entity known to the first user as “Charlie” was befriended firstly on the MySpace™ platform. (See briefly 484 a under column 487.1C of FIG. 4C.) Different iconic GUI representations may be used in the screen of FIG. 1A for representing out-of-STAN friends like “Charlie” and the external platform on which they were befriended. In one embodiment, when the first user hovers his cursor over a friend icon, highlighting or glowing will occur for the corresponding representation in column 103 of the main platform and/or other playgrounds where the friendship with that social entity (e.g., “Charlie”) first originated. In this way the first user is quickly reminded that it is “that” Charlie, the one he first met for example on the MySpace™ platform. So next, and for sake of illustration, a hypothetical example will be studied where User-B (432) is going to be interacting with an out-of-STAN_3 subnet (where the latter could be any one of outside platforms like 441, 442, 444, etc.; 44X in general) and the user forms user-to-user associations (U2U) in those external playgrounds that he would like to later have tracked by columns 101 and 101 r at the left side of FIG. 1A as well as reminded of by column 103 to the right.
  • In this hypothetical example, the same first user 432 (USER-B) employs the username, “Tom” when logged into and being tracked in real time by the STAN_3 system 410 (and may use a corresponding Tom-associated password). (See briefly 484.1 c under column 487.1A of FIG. 4C.) On the other hand, the same first user 432 employs the username, “Thomas” when logging into the alternate SN system 44X (e.g., FaceBook™—See briefly 484.1 b under column 487.1B of FIG. 4C.) and he then may use a corresponding Thomas-associated password. The Thomas persona (432 u 2) may favor focusing upon topics related to music and classical literature and socially interacting with alike people whereas the Tom persona (432 u 1) may favor focusing on topics related to science and politics (this being merely a hypothesized example) and socially interacting with alike science/politics focused people. Accordingly, the Thomas persona (432 u 2) may more frequently join and participate in music/classical literature discussion groups when logged into the alternate SN system 44X and form user-to-user associations (U2U) therein, in that external platform. By contrast, the Tom persona (432 u 1) may more frequently join and participate in science/politics topic groups when logged into or otherwise being tracked by the STAN_3 system 410 and form corresponding user-to-user associations (U2U) therein which latter associations can be readily recorded in the STAN_3 U2U database area 411. The local interface devices (e.g., CPU-3, CPU-4) used by the Tom persona (431 u 1) and the Thomas persona (432 u 2) may be a same device (e.g., same tablet or palmtop computer) or different ones or a mixture of both depending on hardware availability, and moods and habits of the user. The environments (e.g., work, home, coffee house) used by the Tom persona (432 u 1) and the Thomas persona (432 u 2) may also be same or different ones depending on a variety of circumstances.
  • Despite the possibilities for such difference of persona and interests, there may be instances where user-to-user associations (U2U) and/or user-to-topic associations (U2T) developed by the Thomas persona (432 u 2) while operating exclusively under the auspices of the external SN system 44X environment (e.g., FaceBook™) and thus outside the tracking radar of the STAN_3 system 410 may be of cross-association value to the Tom persona (432 u 1). In other words, at a later time when the Tom/Thomas person is logged into the STAN_3 system 410, he may want to know what topics, if any, his new friend “Charlie” is currently focusing-upon. However, “Charlie” is not the pseudo-name used by the real life (ReL) personage of “Charlie” when that real life personage logs into system 410. Instead he goes by the name, “Chuck”. (See briefly item 484 c under column 487.1A of FIG. 4C.)
  • It may not be practical to import the wholes of external user-to-user association (U2U) maps from outside platforms (e.g., MySpace™) because, firstly, they can be extremely large and secondly, few STAN users will ever demand to view or otherwise interact with all other social entities (e.g., friends, family and everyone else in the real or virtual world) of all external user-to-user association (U2U) maps of all platforms. Instead, STAN users will generally wish to view or otherwise interact with only other social entities (e.g., friends, family) whom they wish to focus-upon because they have a preformed social relationship with them and/or a preformed, topic-based relationship with them. Accordingly, the here disclosed STAN_3 system 410 operates to develop and store only selectively filtered versions of external user-to-user association (U2U) maps in its U2U database area 411. The filtering is done under control of so-called External SN Profile importation records 431 p 2, 432 p 2, etc. for respective ones of STAN_3's registered members (e.g., 431, 432, etc.). The External SN Profile importation records (e.g., 431 p 2, 432 p 2) may reflect the identification of the external platform (44X) where the relationship developed as well as user social interaction histories that were externally developed and user compatibility characteristics (e.g., co-compatibilities to other users, compatibilities to specific topics, types of discussion groups etc.) and as the same relates to one or more external personas (e.g., 431 u 2, 432 u 2) of registered members of the STAN_3 system 410. The external SN Profile records 431 p 2, 432 p 2 may be automatically generated or alternatively or additionally they may be partly or wholly manually entered into the U2U records area 411 of the STAN_3 database (DB) 419 and optionally validated by entry checking software or other means and thereafter incorporated into the STAN_3 database.
  • An external U2U associations importing mechanism is more clearly illustrated by FIG. 4B and for the case of second user 432. In one embodiment, while this second user 432 is logged-in into the STAN_3 system 410 (e.g., under his STAN_3 persona as “Tom”, 432 u 1), a somewhat intrusive and automated first software agent (BOT) of system 410 invites the second user 432 to reveal by way of a survey his external UBID-2 information (his user-B identification name, “Thomas” and optionally his corresponding external password) which he uses to log into interfaces 428 a/428 b of specified Out-of-STAN other systems (e.g., 441, 442, etc.), and if applicable; to reveal the identity and grant access to the alternate data processing device (CPU-4) that this user 432 uses when logged into the Out-of STAN other system 44X. The automated software agent (not explicitly shown in FIGS. 4A-4B) then records an alias record into the STAN_3 database (DB 419) where the stored record logically associates the user's UAID-1 of the 410 domain with his UAID-2 of the 44X external platform domain. Yet another alias record would make a similar association between the UAID-1 identification of the 410 domain with some other identifications, if any, used by user 432 in yet other external domains (e.g., 44Y, 44Z, etc.) Then the agent (BOT) begins scanning that alternate data processing device (CPU-4) for local friends and/or buddies and/or other contacts lists 432L2 and their recorded social interrelations as stored in the local memory of CPU-4 or elsewhere (e.g., in a remote server or cloud). The automated importation scan may also cover local email contact lists 432L1 and Tweet following lists 432L3 (or lists for other blogging or microblogging sites) held in that alternate data processing device (CPU-4). If it is given, the alternate site password for temporary usage, the STAN_3 automated agent also logs into the Out-of-STAN domain 44X while pretending to be the alternate ego, “Thomas” (with user 432's permission to do so) and begins scanning that alternate contacts/friends/followed tweets/etc. listing site for remote listings 432R of Thomas's email contacts, Gmail™ contacts, buddy lists, friend lists, accepted contacts lists, followed tweet lists, and so on; depending on predetermined knowledge held by the STAN_3 system of how the external content site 44X is structured. (The remote listings 432R may include cloud hosted ones of such listings.) Different external content sites (e.g., 441, 442, 444, etc.) may have different mechanisms for allowing logged-in users to access their private (behind the wall) and public friends, contacts and other such lists based on unique privacy policies maintained by the various external content sites. In one embodiment, database 419 of the STAN_3 system 410 stores accessing know-how data (e.g., knowledge base rules) for known ones of the external content sites. In one embodiment, a registered STAN_3 user (e.g., 432) is enlisted to serve as a sponsor into the Out-of STAN platform for automated agents output by the STAN_3 system 410 that need vouching for. Aside from scanning and importing external user-to-user association data (U2U; e.g., 432L1-432L3), the STAN_3 system may at repeated times use its access permissions to collect external data relating to current and future roles (contexts) that the user is likely to undertake. The context related data may include, but is not limited to, data from a local client relations management module 432L5 the user regularly uses and data from a local task management module 432L6 the user regularly uses. As explained above, a user's likely context at different times and places may be automatically determined based on scheduled to-do items in his/her task management and/or calendaring databases. It will also become apparent below that a user's context can be a function of the people who are virtually or physically proximate to him/her. For example, if the user unexpectedly bumps into some business clients within a chat or other forum participation session (or in a live physical gathering), the STAN_3 system can automatically determine that there is a business oriented user-to-user association (U2U) present in the given situation based on data garnered from the user's CRM or task tools (432L5-432L6) and the system can automatically determine, based on this that it is likely the user has switched into a client interfacing or other business oriented role. In other words, the user's “context” has changed. When this happens, the STAN_3 system may automatically switch to context-appropriate and alternate user profiles as well as context-appropriate knowledge base rules (KBR's) when determining what augmentations or normalizations should be applied to user originated CFi's and CVi's and what points, nodes or subregions in various Cognitive Attention Receiving Spaces (e.g., topic space) are to next receive user ‘touchings’ (and corresponding “heat”). The concept of context-based CFi augmentations and/or normalizations will be further explicated below in conjunction with FIG. 3R.
  • In one embodiment, and for the case of accessing data of external sources (e.g., 432L1-432L6), cooperation agreements may be negotiated and signed as between operators of the STAN_3 system 410 and operators of one or more of the Out-of STAN other platforms (e.g., external platforms 441, 442, 444, etc.) or tools (e.g., CRM) that permit automated agents output by the STAN_3 system 410 or live agents coached by the STAN_3 system to access the other platforms or tool data stores and operate therein in accordance with restrictions set forth in the cooperation agreements while creating filtered submaps of the external U2U association maps and thereafter causing importation of the so-filtered submaps (e.g., reduced in size and scope; as well as optionally compressed by compression software) into the U2U records area 411 of the STAN_3 database (DB) 419. An automated format change may occur before filtered external U2U submaps are ported into the STAN_3 database (DB) 419.
  • Referring to FIG. 4C, shown as a forefront pane 484.1 is an example of a first stored data structure that may be used for cross linking between pseudonames (alter-ego personas) used by a given real life (ReL) person when operating under different contexts and/or within the domains of different social networking (SN) platforms, 410 as well as 441, 442, . . . , 44X. The identification of the real life (ReL) person is stored in a real user identification node 484.1R of a system maintained, “users space” (a.k.a. user-related data-objects organizing space). Node 484.1R is part of a hierarchical data-objects organizing tree that has all users as its root node (not shown). The real user identification node 484.1R is bi-directionally linked to data structure 484.1 or equivalents thereof. In one embodiment, the system blocks essentially all other users from having access to the real user identification nodes (e.g., 484.1R) of a respective user unless the corresponding user has given written permission (or explicit permission, can be given orally and recorded or transcribed as such after automated voice recognition authentication of the speaker) for his or her real life (ReL) identification to be made public. The source platform (44X) from which each imported U2U submap is logical linked (e.g., recorded alongside) is listed in