WO2021133997A1 - Systèmes et procédés de prédiction d'emplacement étalonnée - Google Patents

Systèmes et procédés de prédiction d'emplacement étalonnée Download PDF

Info

Publication number
WO2021133997A1
WO2021133997A1 PCT/US2020/066982 US2020066982W WO2021133997A1 WO 2021133997 A1 WO2021133997 A1 WO 2021133997A1 US 2020066982 W US2020066982 W US 2020066982W WO 2021133997 A1 WO2021133997 A1 WO 2021133997A1
Authority
WO
WIPO (PCT)
Prior art keywords
campaign
prediction
mobile device
historical
information
Prior art date
Application number
PCT/US2020/066982
Other languages
English (en)
Inventor
Can LIANG
Yilin Chen
Jingqi Huang
Shun JIANG
Amit Goswami
Original Assignee
xAd, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US16/726,056 external-priority patent/US11134359B2/en
Application filed by xAd, Inc. filed Critical xAd, Inc.
Publication of WO2021133997A1 publication Critical patent/WO2021133997A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/02Services making use of location information
    • H04W4/021Services related to particular areas, e.g. point of interest [POI] services, venue services or geofences
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/52Network services specially adapted for the location of the user terminal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/535Tracking the activity of the user

Definitions

  • the present disclosure is related to location-based information technologies, and more particularly to systems and methods for calibrated location prediction.
  • a system for location prediction includes one or more databases storing therein datasets associated with mobile devices, wherein a respective dataset identifies an associated mobile device, and includes a respective time stamp and at least one respective event involving the associated mobile device at a time indicated by the respective time stamp.
  • the datasets include datasets derived from information requests related to mobile devices communicating with the packet-based network.
  • the datasets further include datasets derived from feedbacks about actions on the mobile devices, such as impressions, clicks, calls and/or secondary actions.
  • the system further includes a feature engineering module configured to construct a training feature space including a plurality of training feature sets corresponding, respectively, to a plurality of mobile devices, and to extract a plurality of labels corresponding, respectively, to the plurality of mobile devices.
  • a respective training feature set corresponding to a respective mobile device includes features constructed using datasets associated with the respective mobile device and having time stamps in a training time period.
  • a specific label corresponding to a specific mobile device indicates whether the specific mobile device has had at least one location event at any of one or more predefined places of a location group during a training time frame.
  • the system further includes a machine learning module configured to train a location prediction model corresponding to the location group using the feature space and the plurality of labels.
  • the system further includes a front-end server configured to receive and process information requests and to store processed information requests as datasets in the one or more databases, and a prediction unit configured to predict probabilities of mobile devices to have location events at any of one or more target points of interest (POIs) during certain time frames of an information campaign.
  • a mobile device has a location event at a certain location when an information request associated with the mobile device (or location data included in the information request) triggers a geo-fence of the certain location.
  • the front-end server is configured to receive a first information request associated with a first mobile device, and to determine if the first information request qualifies for the information campaign based on request data associated with the first information request and a set of campaign parameters of the information campaign, the set of campaign parameters including parameters specifying one or more target places of interest (POIs).
  • POIs target places of interest
  • the front-end server is further configured to receive a first information request associated with a first mobile device, and to determine if the first information request qualifies for the information campaign based on data associated with the first information request and a set of campaign parameters of the information campaign.
  • the set of campaign parameters include parameters specifying one or more target places of interest (POIs).
  • POIs target places of interest
  • the front-end server is further configured to, in response to the first information request qualifying for the information campaign, present the first information request for fulfillment based at least in part on a first predicted probability.
  • the prediction unit includes a prediction module configured to apply the prediction model to a first feature set corresponding to the first mobile device to obtain a first prediction result.
  • the first feature set is constructed using datasets associated with (e.g., identifying) the first mobile device and having time stamps in a first prediction time period.
  • the prediction unit further includes a calibration module configured to train a calibration model corresponding to the information campaign, and a calibrated prediction module configured to apply the calibration model to the first prediction result to obtain the first predicted probability.
  • the calibration model is trained in-flight after the information campaign has been running for a predetermined period of time
  • the calibration module is configured to: select a group of mobile devices, each of the group of mobile devices has been impressed with information associated with the information campaign during the predetermined period of time; obtain predicted probabilities of the group of mobile devices to have location events at the one or more target POIs during a prediction time frame; and allocate the group of mobile devices to a plurality of probability brackets corresponding, respectively, to a plurality of ranges of predicted probabilities, such that mobile devices allocated to a specific probability bracket have predicted probabilities in a specific range.
  • the calibration module is further configured to: after the prediction time frame, determine a respective visitation rate corresponding to each respective probability bracket, the respective visitation rate corresponding to a proportion of the mobile devices in the respective probability bracket having had at least one location event at any of the one or more target POIs during the prediction time frame; and machine train a calibration model for the information campaign based at least in part on the visitation rates corresponding, respectively, to the plurality of probability brackets.
  • the calibration model is trained at cold start (e.g., shortly before or soon after (e.g. within a day before or after) a start of the information campaign), and the calibration module is configured to: select a first plurality of historical campaigns, each of the first plurality of historical campaigns having a corresponding historical calibration model; for each respective historical campaign of the first plurality of historical campaigns, estimate a respective difference measure corresponding to the respective historical campaign, the respective difference measure indicating an estimated difference between the information campaign and the respective historical campaign; select one or more similar campaigns from the first plurality of historical campaigns, wherein one or more estimated different measures corresponding to the one or more similar campaigns are less than any of the difference measures corresponding to other historical campaigns of the first plurality of historical campaigns; and obtain the prediction calibration model for the information campaign using one or more historical calibration models corresponding to the one or more similar campaigns.
  • the calibration module is further configured to: obtain a calibration model for each of a second plurality of historical campaigns; and form a plurality of distinct campaign pairs from the second plurality of historical campaigns, each campaign pair of the plurality of campaign pairs including two distinct historical campaigns.
  • the calibration module is further configured to construct a campaign feature space, which includes, for each specific campaign pair of the plurality of campaign pairs, a set of historical campaign features from campaign parameters associated with the specific campaign pair.
  • the calibration module is further configured to: determine campaign difference labels corresponding, respectively, to the plurality of distinct campaign pairs using calibration models of the plurality of distinct campaign pairs; and machine train the campaign comparison model using the campaign feature space and campaign difference labels.
  • a method is performed at one or more computer systems coupled to a packet-based network.
  • Each of the one or more computer systems includes at least one processor, and at least one of the one or more computer systems includes or has access to one or more databases storing therein datasets associated with mobile devices.
  • a respective dataset identifies an associated mobile device, and includes a respective time stamp and at least one respective event involving the associated mobile device at a time indicated by the respective time stamp.
  • the datasets include datasets derived from information requests associated with mobile devices communicating with the packet-based network.
  • the datasets further include datasets derived from feedbacks about actions on the mobile devices, such as impressions, clicks, calls and/or secondary actions.
  • the at least one respective event includes a location event. In some embodiments, the at least one respective event includes one or more of an impression of certain information on a mobile device, a click on a link to additional information made on the mobile device, and/or a call or secondary action taken in response to the certain information.
  • the method comprises constructing a training feature space including a plurality of training feature sets corresponding, respectively, to a plurality of mobile devices, wherein a respective training feature set corresponding to a respective mobile device includes features constructed using datasets associated with the respective mobile device and having time stamps in a training time period, and obtaining a plurality of labels corresponding, respectively, to the plurality of mobile devices.
  • a specific label corresponding to a specific mobile device indicates whether the specific mobile device has had at least one location event at any of one or more predefined places of a location group during a training time frame.
  • the method further comprises machine training a location prediction model corresponding to the location group using the training feature space and the plurality of labels.
  • the method further comprises receiving a first information request during an information campaign, the first information request identifies a first mobile device and includes a first time stamp and a first location of the first mobile device at a time indicated by the first time stamp, and determining if the first information request qualifies for the information campaign based on data associated with the first information request and a set of campaign parameters of the information campaign, the set of campaign parameters including parameters specifying one or more target places of interest (POIs).
  • POIs target places of interest
  • the method proceeds to predicting a first probability of the first mobile device to have a location event at any of the one or more target POIs during a first prediction time frame after the first time, and presenting the first information request for fulfillment based at least in part on the first predicted probability.
  • presenting the first information request for fulfillment based at least in part on the first predicted probability comprises placing a bid for the first information request with a bid price based at least in part on the first predicted probability.
  • the method further comprises, before receiving the first information request, training a prediction calibration model corresponding to the information campaign.
  • predicting the first probability of the first mobile device includes applying the location prediction model to a first feature set corresponding to the first mobile device to obtain a first prediction result; and applying the prediction calibration model to the first prediction result to obtain the first probability.
  • the first feature set is constructed using datasets associated with the first mobile device and having time stamps in a first prediction time period before the first time.
  • the prediction calibration model corresponding to the information campaign is trained in-flight, after the information campaign has been running for a predetermined amount of time, and training the prediction calibration model includes: selecting a group of mobile devices, each of the group of mobile devices has been impressed with information associated with the information campaign; obtaining predicted probabilities of the group of mobile devices to have location events at the one or more target POIs during a prediction time frame; allocating the group of mobile devices to a plurality of probability brackets corresponding, respectively, to a plurality of ranges of predicted probabilities, such that mobile devices allocated to a specific probability bracket have predicted probabilities in a specific range; after the prediction time frame, for each respective probability bracket, determining a respective visitation rate corresponding to the respective probability bracket, the respective visitation rate corresponding to a proportion of mobile devices allocated to the respective probability bracket having had at least one location event at any of the one or more target POIs during the prediction time frame; and machine training a calibration model for the information campaign based at least in part on the visitation
  • obtaining predicted probabilities of the group of mobile devices includes: constructing a prediction feature space including a plurality of prediction feature sets corresponding, respectively, to the group of mobile devices, wherein a specific prediction feature set corresponding to a specific mobile device includes features constructed using datasets associated with the specific mobile device and having time stamps in the prediction time period; and applying the prediction model to the prediction feature space to obtain the predicted probabilities of the group of mobile devices.
  • the respective visitation rate is determined based at least in part on information requests associated with the mobile devices allocated to the respective probability bracket and having time stamps in the prediction time frame.
  • the prediction calibration model corresponding to the information campaign is trained at cold start (e.g., shortly before or shortly after (e.g., within a day before or after) a start of the information campaign), and training the prediction calibration model includes: selecting a first plurality of historical campaigns, each of the first plurality of historical campaigns having a corresponding historical calibration model; for each respective historical campaign of the first plurality of historical campaigns, estimating a respective difference measure corresponding to the respective historical campaign, the respective difference measure indicating an estimated difference between the information campaign and the respective historical campaign; selecting one or more similar campaigns from the first plurality of historical campaigns, wherein one or more estimated different measures corresponding to the one or more similar campaigns are less than any of the difference measures corresponding to other historical campaigns of the first plurality of historical campaigns; and obtaining the prediction calibration model for the information campaign using one or more historical calibration models corresponding to the one or more similar campaigns.
  • the one or more similar campaigns include multiple similar campaigns, and wherein the prediction calibration model is obtained as an average of multiple historical calibration models corresponding to the multiple similar campaigns.
  • estimating the respective difference measure comprises: applying a campaign comparison model to a set of campaign features constructed using campaign parameters of the information campaign and the respective historical campaign to obtain the respective estimated difference measure corresponding to the respective historical campaign.
  • the method further comprises: obtaining a calibration model for each of a second plurality of historical campaigns; forming a plurality of distinct campaign pairs from the second plurality of historical campaigns, each campaign pair of the plurality of campaign pairs including two distinct historical campaigns; constructing a campaign feature space, including, for each specific campaign pair of the plurality of campaign pairs constructing a set of historical campaign features from campaign parameters associated with the specific campaign pair; determining campaign difference labels corresponding, respectively, to the plurality of distinct campaign pairs using calibration models of the plurality of distinct campaign pairs; and machine training the campaign comparison model using the campaign feature space and campaign difference labels.
  • obtaining a calibration model for each of the second plurality of historical campaigns comprises, for a first historical campaign: selecting a set of mobile devices, each of the set of mobile devices has been impressed with information associated with the first historical campaign; obtaining predicted probabilities of the set of mobile devices to be any of one or more POIs associated with the first historical campaign during a historical time frame; allocating the set of mobile devices to a plurality of prediction brackets corresponding, respectively, to a plurality of ranges of prediction probabilities, such that mobile devices allocated to a specific prediction bracket have prediction probabilities in a specific range; after the historical time frame, for each respective prediction bracket, determining a respective historical visitation rate corresponding to the respective prediction bracket, the respective historical visitation rate corresponding to a proportion of mobile devices allocated to the respective prediction bracket having had at least one location event at any of the one or more POIs associated with the first historical campaign during the historical time frame; and machine training a calibration model for the first historical campaign based at least in part on the historical visitation rates corresponding
  • the datasets associated with the respective mobile device and having time stamps in a training time period include datasets derived from information requests associated with the respective mobile device during the training time period, and may further include datasets derived actions on the respective mobile device during the training time period, such as impressions, clicks, calls and/or secondary actions.
  • the respective training feature set corresponding to the respective mobile device includes location features and non-location features. The location features are based on location events triggered by information requests associated with the respective mobile device during the training time period, which may include location events based on predefined geographical regions and location events based on predefined geo-fences associated with points of interest.
  • the first feature set corresponding to the first mobile device includes first location features and first non-location features.
  • the first location features are based on location events triggered by information requests associated with the first mobile device during the first prediction time period, which may include location events based on predefined geographical regions and location events based on predefined geo-fences associated with points of interest.
  • each of the predefined geographical regions borders at least one public road or natural boundary.
  • FIG. 1 is a diagrammatic representation of a system for location prediction according to embodiments.
  • FIG. 2 is a diagrammatic representation of an environment in which the system for location prediction operates according to embodiments.
  • FIG 3 is a diagrammatic representation of a computer/server that performs one or more of the methodologies and/or to provide part or all of the system for location prediction according to embodiments.
  • FIG 4 is a map overlay illustrating exemplary polygonal premise geo-fences in a geographical region according to certain embodiments.
  • FIG 5 is a table illustrating some of the content in a geo-fence database according to certain embodiments.
  • FIG 6 is a map overlay illustrating exemplary geo-blocks in a geographical region according to certain embodiments.
  • FIG. 7 is a table illustrating some of the content in a geo-block database according to certain embodiments.
  • FIGS. 8A-8C are block diagrams illustrating request data at different stages of processing according to certain embodiments.
  • FIG. 9 is a table illustrating some of the content in a request log according to certain embodiments.
  • FIG 10 is a table illustrating some of the content in a feedback log according to certain embodiments.
  • FIG 11 is a flowchart illustrating a method of machine training a location prediction model according to come embodiments.
  • FIGS. 12A-12B are diagrams illustrating time periods and time frames for training and applying location prediction models and calibration models according to certain embodiments.
  • FIG 13 is a block diagram illustrating a data manager for extracting mobile device data according to certain embodiments.
  • FIG 14 includes tables illustrating extracted mobile device data corresponding to a time period according to certain embodiments.
  • FIG 15 is a flowchart illustrating a method for building a feature space according to certain embodiments.
  • FIG 16 includes tables of a set of features in a feature space according to certain embodiments.
  • FIG 17A is a table illustrating a set of labels corresponding to a time frame according to certain embodiments.
  • FIG 17B is a table illustrating examples of location prediction results according to certain embodiments.
  • FIG. 18A is a flowchart illustrating a method of machine training a calibration model according to certain embodiments.
  • FIG 18B is a flowchart illustrating a method of machine training a campaign comparison model according to certain embodiments.
  • FIG 18C is a flowchart illustrating a method of obtaining a calibration model from calibration model(s) of similar historical campaign(s) according to certain embodiments.
  • FIG 19 is a table illustrating impression and visitation events associated with various mobile devices according to certain embodiments.
  • FIG 20 is a table illustrating detected visitation events and predicted probability of visitation for a set of mobile devices according to certain embodiments.
  • FIG 21 is a plot illustrating calibration model training according to certain embodiments.
  • FIG 22 is a plot illustrating calibration models of a pair of historical campaigns according to certain embodiments.
  • FIG 23 is a table illustrating campaign parameters according to certain embodiments.
  • FIG 24 is a table illustrating campaign features according to certain embodiments.
  • FIG 25 is a table illustrating a feature space for training a campaign comparison model according to certain embodiments.
  • FIG 26 is a table illustrating a feature space for applying the campaign comparison model to obtain predicted difference values between a current campaign and a set of historical campaigns according to certain embodiments.
  • FIG 27 is a table illustrating selection of one or more similar historical campaigns based on the predicted difference values according to certain embodiments.
  • FIG 28 is a flowchart illustrating a method for conducting the current campaign using location prediction according to certain embodiments.
  • a location prediction based information service system 100 is shown in FIG. 1.
  • the system 100 is coupled to a packet-based network and configured to serve information to mobile devices communicating with the packet-based network based on predicted probabilities of the mobile devices to have location events at certain locations within certain time frames.
  • the system 100 comprises, among other things, a request processing unit 110, a model training unit 120, and a prediction unit 130.
  • the system 100 is configured to perform one or more methods for calibrated location prediction, as described below.
  • the request processing unit 110 includes a front end server 111, a document server 112, a location module 113 and a geo-fencing module 114.
  • the request processing unit 110 further includes or has access to a geo database 115 storing therein data associated with geo-places, and a campaign database 116 storing therein data associated with a plurality of information campaigns.
  • the request processing unit 110 is configured to receive requests associated with mobile devices communicating with the packet-based network, to process the requests with respect to the geo places in the geo database 115 to detect location events of the mobile devices at any of the geo places, and to buffer and store the processed requests in a request log or database 117.
  • a processed request may include a detected location event corresponding to a time stamp and identifying a geo-place.
  • the request processing unit 110 is further configured to retrieve information or documents associated with information campaigns from the campaign database 116 for delivering to mobile devices based on the processed requests associated with the mobile devices and predictions of the mobile devices to have location events at certain locations within certain time frames.
  • the request processing unit 110 is further configured to receive feedbacks on impression, click/calls, and secondary actions made on the mobile devices in response to the information or documents, and to buffer and store the feedbacks in a feedback log or database 118.
  • the request processing unit 110 further includes a campaign database server 119 to provide information sponsors access to the campaign database 116.
  • the model training unit 120 includes a data manager 121 configured to generate structured mobile device data from the request log 117 and the feedback log 118 for storing in a mobile device database 124, a feature engineering module 122 configured to construct features and extract labels using the structured mobile device data and to store the features and labels in a features/labels database 125, and a model training module 113 configured to train prediction models using some of the features and labels and to store the trained models in a prediction models database 126.
  • the model training unit 120 is further configured to determine a relevance measure (or performance measure) for each of a plurality of geo blocks with respect to the location group (or with respect to an information campaign) and to assign the plurality of geo-blocks into a number of geo-block brackets each corresponding to a distinct range of relevance measures (or performance measures).
  • the model training unit 120 is further configured to construct the features related to a mobile device by generating features related to each of the number of geo-block brackets, generating features related to each of one or more most frequently visited geo-blocks for the mobile device, generating features related to each of a plurality of brands, and/or generating features related to each of the one or more retail geo-blocks, in additional to generating other features.
  • the geo-places include geo-blocks and geo-fences.
  • Each of the geo-blocks correspond to a geographical region having at least one border defined by a public road or natural boundary.
  • Each of the geo-fences correspond to a plurality of points of interest.
  • the geo database 115 includes a geo-block database storing therein data associated with the geo-blocks and a geo-fence database storing therein data associated with the geo fences.
  • the location events include geo-block-based location events and geo-fence-based location events.
  • Each geo-block-based location event is related to a geo-block in the geo-block database, and each geo-fence-based location event is related to a name or brand of a point of interest (POI) having a geo-fence in the geo-fence databases.
  • POI point of interest
  • the prediction unit 130 includes a prediction module 131 configured to apply the prediction models to feature sets to obtain prediction results, which are stored in a general predictions database 134.
  • the prediction results from the prediction module 131 are general prediction results and may need calibration with respect to specific information campaigns.
  • the prediction unit 130 further includes a model calibration module 132 configured to train calibration models, which are then stored in a calibration models database 135.
  • the prediction unit 130 further includes a prediction calibration module 133 configured to calibrate the general predictions to specific campaigns to generate calibrated predictions, which are stored in a calibrated predictions database 136 and are used by the front end server 111 to determine whether and how to present certain processed requests for fulfillment.
  • embodiments of the present disclosure directly improve computer functionality. For instance, embodiments of the present disclosure achieve faster location prediction with smaller memory and processing requirements by translating raw location data into location events with respect geo-fences and geo-blocks and by filtering and aggregating the location events across time and space for machine learning processes.
  • measures of relevance are computed for the geo-blocks using mobile device signals, and the measures of relevance are used to assigne geo-blocks to geo-block brackets for proper dimension reduction and data clustering, resulting in efficient use of computer resources and improved location prediction performance.
  • calibration models are machine trained and used to calibrate general prediction results to specific campaigns, reducing the need to train different location prediction models for different campaigns, and saving time and computer resources.
  • a campaign similarity model is machine trained and used to determine similar historical campaigns for a current campaign so that the calibration models of the similar historical campaigns can be used to generate a calibration model for the current campaign. This reduces the need to machine train the calibration model for the current campaign, and allows the current campaign to have a location prediction model and a calibration model ready at the start.
  • FIG. 2 is a schematic diagram illustrating an overview of an environment 201 in which some embodiments of the disclosed technology may operate.
  • Environment 201 can include one or more computer systems 220 coupled to a packet-based network 200.
  • the packet-based network 200 in certain embodiments includes the Internet 210 and part or all of a cellular network 211 coupled to the Internet 210 via an Internet Gateway.
  • the computers/servers 220 can be coupled to the Internet 210 using wired Ethernet and optionally Power over Ethernet (PoE), WiFi, and/or cellular connections via the cellular network 211 including a plurality of cellular towers 211a.
  • PoE Power over Ethernet
  • the network may also include one or more network attached storage (NAS) systems 221, which are computer data storage servers connected to a computer network to provide data access to a heterogeneous group of clients.
  • NAS network attached storage
  • mobile devices 230 such as smart phones or tablet computers are also coupled to the packet-based network via cellular connections to the cellular network 211.
  • a WiFi hotspot such as hotspot 235
  • a mobile device 230 may connect to the Internet 210 via a WiFi hotspot 235 using its built-in WiFi connection.
  • the mobile devices 230 may interact with computers/servers 220 coupled to the Internet 210.
  • a mobile device 230, or its user, or anyone or anything associated with it, or any combination thereof, is sometimes referred to herein as a mobile entity.
  • FIG. 3 illustrates a diagrammatic representation of a computer/server 220 according to certain embodiments.
  • the computer/server 220 may operate as a standalone device or as a peer computing device in a peer-to-peer (or distributed) network computing environment.
  • the computer/server 220 includes one or more processors 302 (e.g., a central processing unit (CPU), a graphic processing unit (GPU), and/or a digital signal processor (DSP)) and a system or main memory 304 coupled to each other via a system bus 300.
  • processors 302 e.g., a central processing unit (CPU), a graphic processing unit (GPU), and/or a digital signal processor (DSP)
  • main memory 304 coupled to each other via a system bus 300.
  • the computer/server 220 may further include static memory 306, a network interface device 308, a storage unit 310, one or more display devices 330, one or more input devices 334, and a signal generation device (e.g., a speaker) 336, with which the processor(s) 302 can communicate via the system bus 300.
  • static memory 306 a network interface device 308, a storage unit 310, one or more display devices 330, one or more input devices 334, and a signal generation device (e.g., a speaker) 336, with which the processor(s) 302 can communicate via the system bus 300.
  • the display device(s) 330 include one or more graphics display units (e.g., a plasma display panel (PDP), a liquid crystal display (LCD), a projector, or a cathode ray tube (CRT)).
  • the input device(s) 334 may include an alphanumeric input device (e.g., a keyboard), a cursor control device (e.g., a mouse, trackball, joystick, motion sensor, or other pointing instrument).
  • the storage unit 310 includes a machine-readable medium 312 on which is stored instructions 316 (e.g., software) that enable anyone or more of the systems, methodologies or functions described herein.
  • the storage unit 310 may also store data 318 used and/or generated by the systems, methodologies or functions.
  • the instructions 316 e.g., software
  • the instructions 316 may be loaded, completely or partially, within the main memory 304 or within the processor 302 (e.g., within a processor's cache memory) during execution thereof by the computer/server 220.
  • the main memory 304 and the processor 302 also constitute machine-readable media.
  • the procedures, devices, and processes described herein constitute a computer program product, including a non-transitory computer-readable medium, e.g., a removable storage medium such as one or more DVD-ROM's, CD-ROM's, diskettes, tapes, etc., that provides at least a portion of the software instructions for the system.
  • a computer program product can be installed by any suitable software installation procedure, as is well known in the art.
  • at least a portion of the software instructions may also be downloaded over a cable, communication and/or wireless connection.
  • the location prediction system 100 can be implemented using one or more computers/servers 220 executing programs to carry out the functions and methods disclosed herein. It should be understood that the example embodiments described herein may be implemented in many different ways. In some instances, the various servers and/or modules in FIG 1, e.g., the front end server 111, the document server 112, the location module, the campaign database server 119, the data manager 121, the feature engineering module 122, the model training module 123, the prediction module 131, the model calibration module 132, and/or the prediction calibration module 133, as described herein, may be implemented by one or more computers/servers 220, respectively or in various combinations.
  • the various servers and/or modules in FIG 1, e.g., the front end server 111, the document server 112, the location module, the campaign database server 119, the data manager 121, the feature engineering module 122, the model training module 123, the prediction module 131, the model calibration module 132, and/or the prediction calibration module 133, as described herein may be implemented
  • the one or more computers/servers may be general purpose computers/servers that are transformed into the machines that execute the methods described herein, for example, by loading software instructions into one or more data processors, and then causing execution of the instructions to carry out the functions, processes, and/or methods described herein.
  • some of the computers/servers 220 are coupled to each other via a local area network (LAN) 210, which in turn is coupled to the Internet 210.
  • LAN local area network
  • each computer/server 220 referred herein can include any collection of computing devices.
  • the system 100 includes a front-end server that receives requests from the packet-based network 200. These requests may be generated by one or more computers/servers 220 in the packet-based network as they provide mobile services to the mobile devices.
  • the system 100 further includes a location module 113 coupled to the front-end server and configured to detect the location of a mobile device associated with each of the requests.
  • the location module 113 is further configured to examine the location data in each received request to determine whether they include a reliable latitude/longitude (LL) pair, and if the request does not include a reliable LL pair, the location module 113 would proceed to derive the location of the associated mobile device from other information in the location data, as described in more detail in commonly owned U.S. Patent No. 9,886,703, issued on February 6, 2018, which is incorporated herein by reference in its entirety.
  • the system 100 further includes a geo-fencing module 114 coupled to the location module 113 and configured to determine if the detected mobile device location triggers any geo-place(s) in a geo-database and returns the triggered geo-place(s) to the front-end server.
  • the geo-places include geo-fences and geo-blocks
  • the geo database 115 is a spatial database optimized for storing and querying data that represent geographical areas or spaces and may include spatial data and meta data associated with each of the geographical areas or spaces.
  • the geo-fences in the geo database 115 include spatial data representing virtual perimeters of defined areas or places that mirror real-world geographical areas associated with various entities and/or brands.
  • a defined area can be a static circle around a business location, e.g. a fence obtained using offline index databases such as InfoUSA (www.infousa.com), which provides a list of POIs and their locations, or areas specified by marketers using predefined boundaries, such as neighborhood boundaries, school attendance zones, or parcel boundaries, etc.
  • the defined areas include one or more geo-fences for each of a plurality of points of interests in consideration of the map data around the POI.
  • one or more polygons are defined for the Costco Almaden store 401 to be in conformity with the real-world geographical structure and boundaries of the store and its surroundings, such as a first polygon 410 around the building of the store, a second polygon 420 around the building and its parking lot, and/or a third polygon 430 around a shopping area or business region including the store and other points of interests (POIs).
  • these different types of geo-fences are defined for a point of interest (POI) to indicate different levels of intentions, interests, and/or behavior, etc., of a mobile user with respect to the POI, which can be used for location prediction purposes.
  • POI point of interest
  • different types of geo-fences are associated with a business and may include, for example, (1) a business center (BC) represented by, for example, a polygon corresponding to the perimeter of the building of the business (e.g., the first polygon 410 in FIG. 4); (2) a business premise (BP) represented by a polygon corresponding to the perimeter of the business building and the neighboring parking lots (e.g., the second polygon 420 in FIG 4); and (3) a business region (BR) or area represented by a polygon corresponding to the perimeter of a shopping center or business or commercial area in which this business is located (e.g., the third polygon 430 in FIG 4).
  • BC business center
  • BP business premise
  • BR business region
  • a business center is triggered by a mobile device location, it can be reliably inferred that the user of the mobile device is interested in the business by actually visiting it. Triggering of a business premise provides good indication of an intent to visit the business, but not as strong as triggering the business center. If a user triggers a business region, the intent may be regarded as valid but weaker than that from triggering a business premise.
  • FIG 5 illustrates examples of some of the geo-fences in the geo database 115, according to certain embodiments.
  • the site Costco in Almaden has three different types of geo-fences associated with it - geo-fence with Fence ID 19-35175 corresponds to a business center (BC) , which is defined by a polygon around the store building and represented by spatial index al, a2, ... , ai; geo-fence with Fence ID 19-35176 corresponds to a polygon around the site’s larger premise including its parking lot and represented by spatial index bl, b2, ...
  • BC business center
  • geo-fence with Fence ID 19-35177 corresponds to polygon around the shopping center including the store and other POIs and represented by spatial index cl, c2, ... , ck. Note that geo-fence with Fence ID 19-35177 is also associated with the names/brands of other POIs in the shopping center, as well as name of the shopping center itself.
  • FIG 5 also shows that the site T.J. Maxx is associated with Fence ID 19-35177 and also has two other types of fences associated with it, and the site Trader Joe’s is also associated with Fence ID 19-35177 and has at least a business center place associated with it.
  • each geo-fence entry in the geo database 115 includes the spatial data associated with the respective place together with some meta data about the respective place, such as, for example, one or more names/brands associated with the place, a category of the place, a place identifier identifying a particular locale (e.g., city, district, etc.) for the place, the place type, and/or one or more doc IDs identifying one or more information documents (e.g., one or more html/JavaScript files) associated with the names/brands or the place.
  • a POI’s name is established as its brand, so they are used interchangeably.
  • the brand of a POI is referred to hereafter as either the name or the brand of the POI, whichever is associated with the POI in the geo-fence database.
  • the geo-blocks in the geo database 115 represent geographical regions with natural boundaries such as major roads, shorelines, mountain ranges, etc., as described in further detail below.
  • FIG. 6 illustrates graphically exemplary geo-blocks according to certain embodiments.
  • the geo-blocks are shown as outlined in boundaries overlaid on top of a map for the area, taken from, for example, Google Map, and the boundaries of the geo blocks are mostly aligned with major roads and natural boundaries, taking into account the road width so as to exclude mobile signals from travelers on the major roads.
  • geo-block 601 containing the Moonlite Shopping Center is shown to be bordered on three sides by major roads, El Camino Real, Bowers Ave, and Benton St., respectively, and on another side by the Saratoga Creek.
  • Each of the geo-blocks shown in FIG 6 can be further partitioned into more granular blocks bordered by smaller roads.
  • Real world entities present in these geo-blocks tend to serve common functional purposes (residential, retail etc.), and these blocks form the foundation for the construction of boundaries that are highly indicative of location dependent attributes such as intention and demographics.
  • FIG. 7 illustrates examples of some of the geo-blocks in the geo database 115, according to certain embodiments.
  • each geo-block includes spatial data defining the boundary of the geo-block and meta data including, for example, the city/state in which the geo-block is located, the functionality of the geo-block (e.g., residential, retail, recreational, educational, etc.), one or more major POIs in the geo-block, as well as other information such as demographic of the residents or visitors of the geo-block, and inventory of requests with location data in the geo-block, etc., which can be derived from logged request data.
  • the functionality of the geo-block e.g., residential, retail, recreational, educational, etc.
  • major POIs in the geo-block e.g., as well as other information such as demographic of the residents or visitors of the geo-block, and inventory of requests with location data in the geo-block, etc., which can be derived from logged request data.
  • the front-end server receives a request 801 presented by, for example, a server run by a mobile service provider (MSP or MSP server), or an ad exchange (or exchange), via the network 200.
  • the request 801 includes a request ID, mobile device data such as mobile device ID, maker/model, operating system, etc., mobile entity data such as user ID (UID), age, gender, income bracket, education level, etc., mobile device location data including a plurality of location components, such as latitude and longitude coordinates (LL), IP addresses (IP), postal or zip codes (ZC), and/or city-state names (CS), etc.
  • the request may further include other information.
  • the front-end server validates the location information by checking the validity and consistency of the location components and by weeding out any invalid location component(s).
  • the LL is usually believed to be the most useful location component.
  • mobile applications at the MSP server typically provide only coarse location data in the form of, for example, an IP address, a ZC (e.g. entered by the user at the time of registration), or CS.
  • mobile applications at the MSP server frequently provide LLs obtained from geo-coding software, which translates ZC, CS, and other points of interests into one representative LL.
  • such representative LLs are categorized as "bad LLs".
  • a bad LL can be, for example, a centroid of a ZC/CS, or any fixed point on a map (e.g. (0,0) or an arbitrary location).
  • the location module 113 is configured to weed out the bad LL’s, so that location data with bad LL’s are not provided to the next stage processing, by using the techniques disclosed in commonly owned U.S. Patent Application No. 14/716,816, entitled “System and Method for Estimating Mobile Device Locations,” filed on May 19, 2015, which is incorporated herein by reference in its entirety.
  • the location module 113 is further configured to estimate the location of the mobile device from the request 801 and generate location data to represent an estimated mobile device location, which may be a geographical point represented by a lat/long pair or one or more probable areas or regions the mobile device is estimated to be in, as shown in processed request 802 with generated location data in FIG. 8B.
  • the geo-fencing module 114 queries the geo database 115 with the lat/long pair or the one or more probable regions to determine whether the location data triggers one or more geo-places in the geo database 115, and returns the triggered geo-place(s) to the front-end server.
  • the front end server 111 annotates the request 801 with the triggered geo-place(s) to generate an annotated (or processed) request 810, and outputs the annotated request 810 to buffer 1, which buffers and outputs the annotated request 810 to a request log 117.
  • the triggered geo- place ⁇ ) may include a geo-block (if the mobile device is in a place that has been geo- blocked) and may further include one or more geo-fences if the estimated location or probable area or region is in or overlaps with the one or more geo-fences, as shown in FIG. 8C.
  • the triggering of a geo-place e.g., a geo-block or a geo-fence
  • an annotated request may include one or more location events.
  • FIG. 9 is a table illustrating exemplary entries (or datasets) in the request log 117, according to certain embodiments.
  • Each dataset in the request log 117 corresponds to a respective processed (or annotated) request and includes at least some of the data in the respective annotated request, such as request (or packet) ID, mobile device information such as mobile device ID, make/model, mobile user information such as UID, age, gender, eduction, etc., a latitute/longitude pair, data related to location events involving triggered geo-fence(s) and/or a geo-block, mobile application used at the time of the request, and the time stamp of the request, etc.
  • the annotated requests in the request log 117 are organized in the order of their respective time stamps, as shown in FIG. 9.
  • the request 801 received by the front-end server includes other information as well as the location information, such as an application program running on the mobile device, a time stamp indicating the time of the request (e.g., day, hour, minute, etc.), one or more keywords suggesting types of information for returning to the mobile device, and/or other information associated with the mobile user, the mobile device, and/or the MSP.
  • the location data can trigger multiple places. For example, as shown in FIG.
  • a request that triggers the BC place 410 of Costco Almaden also triggers the BR place 430 of any of the POIs in the same business region as well as the geo-fence for the business region (e.g., a retail center or shopping mall).
  • the request may be annotated with the BR place of one or more other POIs in the same business region, as well as the BC place of Costco Almaden.
  • the BR place may be the only geo-fenced place associated therewith, so the business region is triggered as long as the associated BR place is triggered.
  • Each location event involving a triggered geo-fence or a triggered geo-block is included in the annotated request together with information about the triggered geo-fence or geo-block. If a request triggers multiple places associated with a brand, only the smallest of the places (e.g., the BC or the BP place) is included as a location event. As shown in FIG 8C, some or all of the meta data 812 of a triggerd geo-fence and some or all of the meta data 814 of the triggered geo-block can be included in the annotated request 810.
  • the front end server 111 in system 100 is further configured to evaluate the annotated request 810 with respect any of one or more information campaigns running in the system 100, and with respect to related location prediction generated by the prediction unit 130, and to determine whether to present the request 801 for fulfillment.
  • the front end server 111 presents the request for fulfillment by placing a bid for the request 801 at the exchange or MSP, and, in the case the bid is accepted, transmitting the annotated request 810 to the document (or information) server in the system 100.
  • the front end server 111 has access to the location predictions stored in a calibrated predictions database 136 in the prediction unit 130.
  • the document (or information) server is configured to receive the annotated request 810 output from the front-end server and to determine which document to select based on the location data and non-location data in the annotated request. The selected document is then transmitted by the document server 112 to the MSP server (or ad exchange) via the network 200.
  • the information server is a computer server, e.g., a web server, backed by a campaign database server 119 that information sponsors use to periodically update the content thereof and may store information documents.
  • Each of the information documents may be stored in a campaign database 116 in the form of, for example, an html/JavaScript file or a link thereto, which, when loaded on a mobile device, displays information in the form of, for examples, a banner (static images/animation) or text.
  • the document server 112 evaluates the annotated request 810 based on the location predictions generated by the prediction module 130.
  • the document selected for transmission to the MSP can be provided in the form of, for example, an html/JavaScript file, or a link to a universal resource location (URL), which can be used by the MSP or a mobile device to fetch the html/JavaScript file.
  • the html/JavaScript file once displayed or impressed on a mobile device, may also include one or more links that an interested user can click to access a webpage or place a call using the mobile device.
  • the webpage enables the user of the mobile device to take secondary actions such as downloading an app or make an on-line purchase.
  • the html/JavaScript file is designed such that when it is displayed or impressed on a mobile device, a signal is sent by the MSP server or the mobile device automatically to the document server 112 either directly or via another server (e.g., the MSP server so that the document server 112 can keep track of whether the file has really been impressed on the mobile device.
  • mechanism are also put in place such that when any of the one or more links are clicked, or when the mobile user download an app or make a purchase from a linked webpage, a signal is also sent from the mobile device to the document server 112 in the background either directly or indirectly so that the document server 112 can keep track of the clicks/calls or secondary actions made in response to the impression.
  • the document server 112 provides data of such feedback events (i.e., impressions, clicks/calls, and secondary actions) to buffer 2, which buffers and outputs the data to a feedback log 118.
  • FIG. 10 is a table illustrating exemplary entries (or datasets) in the feedback log 118 in the ordered of their respective time stamps, according to certain embodiments.
  • raw location data in requests are converted into brands and geo-blocks in processed requests.
  • the logged data in the requests log and the feedback log 118 collected over a period of time form a large collection of mobile device data (e.g., millions and millions of annotated requests and impression/click/call events).
  • the dimensions of these data are usually too large to be used directly for meaningful location prediction.
  • the model training unit 120 is configured to reduce the dimmensions of the logged data by extracting features and labels from the location data, to train one or more prediction models using the features and labels, and to apply the prediction models to an appropriate feature space to obtain off-line predictions.
  • the model training unit 120 includes a data manager 121, a mobile device data database 124, a feature engineering module 122, a features/labels database 125, a model training module 123, and a prediction models library.
  • the model training unit 120 is configured to perform a method 1100 for training a location prediction model off line, while the front end server 111 continues to receive and process incoming requests.
  • method 1100 includes extracting (1110) mobile device data from datasets in the request log and feedback log having time stamps in a training time period, constructing (1120) a training feature space from at least the mobile device data, extracting (1130) a set of labels from datasets in the request log having time stamps in a training time frame, and training (1140) a prediction model using a machine learning routine with the training feature space and the set of labels as input.
  • the data manager 121 is configured to extract mobile device data corresponding to location events in a certain time period (e.g., training time period TTP shown in FIG 12A) from datasets in the request log 117 having time stamps in the corresponding time period.
  • the data manager 121 includes a search engine configured to search, for each mobile device of a plurality of mobile devices, processed requests associated with the mobile device and having time stamps in a time period (e.g., TTP), and a buffer configured to buffer the output from the search engine.
  • the data manager 121 further includes a compressor configured to compress the buffered data, such that multiple location events corresponding to consecutive time stamps triggered by the mobile device at the same place within a preset time interval (e.g. 1 hour), are reduced to a single location event.
  • a preset time interval e.g. 1 hour
  • the multiple events are reduced to a single event to indicate a single visit by the mobile devise to the triggered geo-fence or geo-block at one of the time stamps (e.g., the earliest time stamp) with a duration computed using the the difference between the earliest time stamp and a latest time stamp within an hour from the earliest time stamp.
  • the time stamps e.g., the earliest time stamp
  • Another location event by the mobile device at the same place but with a time stamp beyond the preset time interval from the earliest time stamp would be considered another visit to the place by the mobile device, even though the mobile device may have stayed at the same place during the whole time.
  • the search engine can be configured to only search for location events with certain types of geo-fences depending on the associated brands. For example, for certain brands, only location events with triggered BC places are considered as visits to these brands/names, for certain other brands, location events with triggered BP places are sufficient to be considered as visits to these brands, and for some brands, such as retail centers or shopping malls, location events with triggered BR places are considered as visits to these brands.
  • the data manager 121 further includes another buffer that stores the compressed location events for the mobile device, and an aggregator configured to aggregate the location events to form a set of mobile device data corresponding to location events in the time period TTP for the mobile device. As shown in FIG.
  • the mobile device data 1410 for the mobile device with the UID 36***412 may include, for example, device and user meta data such as age, gender, education level, and other information such as maker/model, operating system, etc., aggregated location events associated with each geo-block triggered by the mobile device during the time period TTP, aggregated location events associated with each brand triggered by the mobile device during the time period TTP, aggregated feedback events associated with one or more documents impressed on the mobile device during the time period TTP, and aggregated usage data associated with mobile applications used on the mobile device during the time period TTP.
  • a brand is triggered when a location event 812 includes the brand.
  • the brand uses multiple types of fences, as described above, the brand is triggered when a location event 812 includes the brand and the fence type specified for the prediction model to be trained.
  • aggregated location events associated with each triggered geo-block (e.g., GBx) or brand (e.g., Bx) includes, for example, a number of visits to the geo-block or brand during the time period TTP, time of last visit during TTP, average length of stay per visit, etc.
  • the number of visits to the geo-block or brand is divided among a plurality of time blocks (shown as TB1, TB2, TB3) during a day, such as morning (6:00 am to 12:00 pm), afternoon (12:00 pm to 6:00 pm) and evening (6:00 pm to 6:00 am).
  • usage data associated with each mobile application used on the mobile device during the time period TTP are aggregated likewise.
  • the aggregated feedback events associated with each of one or more documents (e.g., Docx) impressed on the mobile device during the time period TTP may include, for example, a number of impressions of the document made on the mobile device during TTP, a number of click/calls the mobile device made on the impressed document, and a number of secondary actions taken with the mobile device in response to the impressed document. These numbers can also be divided among the different time blocks.
  • the data manager 121 is configured to perform the above searching, compression, and aggregation processes for each of the plurality of mobile devices and to store the compressed and aggregated data for the plurality of mobile devices in the mobile device database 124.
  • the compressed and aggregated data for each mobile device e.g., mobile device with UID 36***412
  • each time period e.g., time period TTP
  • the model training unit 120 further includes a feature engineering module 122 configured to engineer a set of features for a location group corresponding to each of the plurality of time periods according to a feature engineering process 1500 illustrated in FIG. 15.
  • the feature engineering module 122 is configured to construct a training feature space for the location group using at least the mobile device data corresponding to the training time period TTP.
  • the location group includes one or more locations selected for prediction. The one or more locations may correspond to, for example, one or more geo-fences associated with, for example, one or more brands or one or more categories in the geo database 115.
  • the feature engineering module 122 is configured to select a set of geo-blocks and determine a relevance measure for each of the geo-blocks with respect to the location group, as recited in block 1505 of process 1500 illustrated in FIG 15.
  • the relevance measure (or performance measure) for a geo-block can be determined using a variety of relevance or performance metrics, as discussed in further detail below.
  • the relevance measure for the geo-block with respect to a location group can be determined using a first number of distinct mobile devices that triggered the geo-block during a pre-selected time period and a second number of the mobile devices that triggered the geo-block and also triggered the brand or category associated with any of the locations in the location group during the same time period.
  • the relevance measure can simply be a ratio of the second number to the first number or some other combination of the two numbers and/or other factors.
  • the first number and the second number can be determined using logged request data or extracted mobile device data.
  • Other examples of determining the relevance measures (or performance measures) are provided in commonly -owned U.S. Patent Application No. 15/999,330, filed August 17, 2018, now U.S. Patent 10,278,014, which is incorporated herein by reference in its entirety.
  • the feature engineering module 122 is further configured to assign the selected geoblocks into a number of geo-block (GB) brackets according to their respective relevance measures (or performance measures), as recited in block 1510 of process 1500 illustrated in FIG 15.
  • the selected geo-blocks are divided into a number of geo-block brackets each corresponding to a distinct range of relevance measures (or performance measures). For example, suppose there are 2000 selected geo-blocks, which can be all of the geo-blocks in one or more targeted geo graphical regions, and there are 20 geo-block brackets (e.g., GBB1, GBB2, ...
  • the geo-blocks in GBB1 could include 100 geo-blocks with the highest relevance or performance measures
  • the geo-blocks in GBB2 could include 100 geo-blocks with the next highest relevance or performance measures, and so on. Assigning the geo-blocks into geo block brackets largely reduces the data dimmensions.
  • the feature engineering module 122 is further configured to select a set of brands, which may be, for example a set of relatively big brands (BB) that have sizable visits by mobile users to allow sufficient density of data, as recited in block 1505 of process 1500 illustrated in FIG 15.
  • BB relatively big brands
  • the feature engineering module 122 may be further configured to assign the selected big brands into a number of big brand (BB) brackets according to their respective measures of relevance, as recited in block 1520 of process 1500 illustrated in FIG 15.
  • the feature engineering module 122 may be configured to determine the relevance measure for each selected brand with respect to the location group.
  • the relevance measure for a brand can be determined using a third number of distinct mobile devices that triggered the brand during a pre-selected time period and a fourth number of the mobile devices that triggered the brand and also triggered a brand associated with any of the locations in the location group during the same time period.
  • the relevance measure can simply be a ratio of the fourth number to the third number or some other combination of the two numbers and/or other factors.
  • the third number and the fourth number can be determined using logged request data or extracted mobile device data.
  • the selected brands are divided into a number of big brand brackets (BBB) each corresponding to a distinct range of relevance measures. For example, suppose there are 1000 selected brands, and there are 20 big brand brackets (e.g., BBB1, BBB2, ... , BBB20), the brands in BBB1 could include 50 brands with the highest performance measures, the brands in BBB2 could include 50 geo-blocks with the next highest performance measures, and so on. Selecting the big brands and optionally assigning them to the big brand brackets further reduces the data dimmensions.
  • BBB big brand brackets
  • the feature engineering module 122 is further configured to construct a set of features for each of the plurality of mobile devices using the mobile device data associated with the mobile device and corresponding to the training time period TTP. As shown in FIG. 15, the feature engineering module 122 is configured to: select a user from among a plurality of users (block 1525 of process 1500), generate features related to each geo-block bracket for the user (block 1530 of process 1500), generate features related to retail geo-blocks for the user (block 1535 of process 1500), identify a set of most frequently visited geo-blocks (MFVGB-1, ...
  • the set of MFVGBs may include a MFVGB from each of a plurality of geo-block functionalities, such as retail, residential, industrial, etc.
  • the MFVGBs may include a residential MFVGB, which could be the home of the mobile user, an industrial MFVGB, which could be the workplace of the mobile user, and a retail MFVGB, which may be where the individual does most of the shopping, etc.
  • FIG. 16 illustrates as examples a set of features for a mobile device with UID 52**256 for the time period TTP.
  • the set of features may include device/user meta data.
  • the features related to each GBB of the geo-block brackets include a number of visits to any geo-block in the GBB during the time period TTP, time of last visit to any geo-block in the GBB during TTP, an average length of stay per visit to any geo-block in the GBB during TTP, etc.
  • the number of visits to the GBB is divided among a plurality of time blocks (shown as TB1, TB2, TB3) during a day, such as morning (6:00 am to 12:00 pm), afternoon (12:00 pm to 6:00 pm) and evening (6:00 pm to 6:00 am).
  • time blocks shown as TB1, TB2, TB3 during a day, such as morning (6:00 am to 12:00 pm), afternoon (12:00 pm to 6:00 pm) and evening (6:00 pm to 6:00 am).
  • the features related to each BB or BBB of the big brands (BBi, BB2, ... , BBm) or big brand brackets (BBBi, BBB2, ... , BBB m ) include a number of visits to the BB or any brand in the BBB during the time period TTP, time of last visit to the BB or any brand in the BBB during TTP, an average length of stay per visit to the BB or any brand in the BBB during TTP, etc.
  • the number of visits to the BB or any brand in the BBB is divided among a plurality of time blocks (shown as TB1, TB2, TB3) during a day, such as morning (6:00 am to 12:00 pm), afternoon (12:00 pm to 6:00 pm) and evening (6:00 pm to 6:00 am).
  • time blocks shown as TB1, TB2, TB3 during a day, such as morning (6:00 am to 12:00 pm), afternoon (12:00 pm to 6:00 pm) and evening (6:00 pm to 6:00 am).
  • a retail geo-block is a geo-block having a retail functionality, as indicated by its associated meta data.
  • the features related to the retail geo-blocks among the selected geo-blocks include a number of weighted visits and a number of net visits.
  • the number of net visits is the number of visits made with the mobile device to any of the retail blocks within the time period TTP, as determined using the mobile device data associated with the mobile device and corresponding to the TTP.
  • the number of weighted visits is the weighted sum of a number of visits to each of the retail geo-blocks multiplied by a weight of the retail geo-block.
  • the weight of the retail geo-block can be computed based on a fifth number of distinct mobile devices that triggered the retail geo-block during a pre selected time period and a sixth number of the mobile devices that triggered the retail geo block and also triggered a brand associated with any of the locations in the location group during the same time period.
  • the weight can simply be a ratio of the sixth number to the fifth number or some other combination of the two numbers and/or other factors.
  • the sixth number and the fifth number can be determined using logged request data or extracted mobile device data.
  • the most frequently visited geo-block is the geo block that has the most number of visits from the mobile device compared to the other selected geo-blocks.
  • the features associated with the MFVGB can include for example, a number of visits to the MFVGB by the mobile device during the time period TTP, and the distance from the MFVGB to a nearest location among the locations in the location group.
  • the number of visits to the MFVGB is divided among a plurality of time blocks (shown as TB1, TB2, TB3) during a day, such as morning (6:00 am to 12:00 pm), afternoon (12:00 pm to 6:00 pm) and evening (6:00 pm to 6:00 am).
  • the set of features for the mobile device may include other features, such as mobility features and feedback features.
  • the mobility features may include, for example, a number of distinct brands triggered by the mobile device during the time period TTP, and a GB ratio of a number of distinct geo-blocks triggered by the mobile device to the sum of visits to all of the triggered geo-blocks during the time period TTP.
  • the feedback features may include, for example, pre-exposure feature, which may be a number impressions of one or more documents related to the location group on the mobile device during the time period TTP, a click/call ratio, which may be the ratio of a number of times a click is made on the mobile device in response to the one or more documents to the number of impressions of the one or more documents, and a secondary action ratio, which may be the ratio of a number of times secondary actions are made on the mobile device in response to the one or more documents to the number of impressions of the one or more documents.
  • pre-exposure feature which may be a number impressions of one or more documents related to the location group on the mobile device during the time period TTP
  • a click/call ratio which may be the ratio of a number of times a click is made on the mobile device in response to the one or more documents to the number of impressions of the one or more documents
  • secondary action ratio which may be the ratio of a number of times secondary actions are made on the mobile device in response to the one or more documents
  • the feature space corresponding to the time period TTP is referred to herein as the training feature space.
  • the feature generator is further configured to extract a set of labels corresponding to a time frame (e.g., TTF shown in FIG 12A) immediately or shortly after the training time period TTP, for the plurality of mobile devices.
  • TTP can be the past 1-3 months from the day on which the training feature space is constructed
  • TTF can be the next 1-2 weeks from the day on which the training feature space is constructed. Note that the beginning of TTF and the end of TTP do not have to coincide. For example, they can be up to a day or a few days apart from each other.
  • the set of labels can be extracted from location events in the processed requests having time stamps in the time frame.
  • the set of labels include one label for each of the plurality of mobile devices and the label is a “1” or “0” depending on whether the mobile device has triggered a geo-fence or brand associated with the location group during the time frame, as shown in FIG 17 A.
  • the model training unit 120 further includes a model training module 123 configured to employ machine learning approaches to train a prediction model using the training feature space and the set of labels.
  • the training module uses, for example, the Hadoop® Hive machine learning platform.
  • the machine learning approaches used to train the off-line prediction models may include random forest, decision trees, and/or boosting trees.
  • the machine learning approaches used in to train the on-line prediction model may include any one or more of: supervised learning using, for example, logistic regression, back propagation neural networks, etc., unsupervised learning using, for example, an Apriori algorithm, K-means clustering, semi-supervised learning, reinforcement learning using, for example, a Q-leaming algorithm, temporal difference learning, and/or any other suitable learning style.
  • the model training module 123 may train a plurality of prediction models for different location groups associated with different information campaigns, respectively.
  • the model training unit 120 further includes a prediction models database 126 storing therein a model library including prediction models trained by the training module, each prediction model in the model library corresponds to a combination of a location group, a training time period and a training time frame.
  • the data manager 121 is configured to extract mobile device data corresponding to a prediction time period (e.g., time period PTP, as shown in FIG 12A) periodically (e.g., once per day) for each of a plurality of mobile devices, and the feature generator is further configured to construct a set of features corresponding to the prediction time period PTP for each of a plurality of mobile devices.
  • a prediction time period e.g., time period PTP, as shown in FIG 12A
  • the feature generator is further configured to construct a set of features corresponding to the prediction time period PTP for each of a plurality of mobile devices.
  • the prediction time period has the same duration as the training time period and its start time (or end time) may be reset periodically (e.g., every time when structured mobile device data and the features corresponding to the PTP are updated), so that its end time is at or shortly before (e.g., within a day from) the time when the mobile device data for the prediction time period is recalculated.
  • the prediction time period has a start time T p sometime after a start time Tt of the training time period and an end time at or shortly before a time when prediction is being made.
  • both TTP and PTP have about the same duration, which can be for example, several weeks to several months (e.g., 1-3 months) to allow sufficient size of the data pool for feature engineering.
  • the mobile devices associated with the aggregated data corresponding to different time periods do not have to be the same.
  • the mobile devices associated with the aggregated data corresponding to time period PTP may be the same plurality of mobile devices associated with the aggregated data corresponding to the time period TTP or a different set of mobile devices.
  • the prediction unit 130 includes a prediction module 131 configured to select and apply a prediction model to a prediction feature set to obtain a corresponding prediction result.
  • each prediction result corresponds to a mobile device, a location group and a prediction time frame, and indicates a predicted probability of the mobile device having at least one location event at any of the locations in the location group during the prediction time frame.
  • FIG. 17B illustrates a set of prediction results generated by the prediction module 131 after applying the prediction model corresponding to a location group to the feature sets of a plurality of mobile devices according to some embodiments.
  • the set of prediction results include predicted probabilities of respective mobile devices to have location events at any of the location group within a prediction time frame PTF after the prediction time period PTP.
  • the set of prediction results shown in FIG 17B are then stored in a general predictions database 134, as they are general prediction results not tailored (or calibrated) to a specific information campaign.
  • the prediction time frame e.g., PTF shown in FIG. 12A
  • the prediction time frame is immediately or shortly after the prediction time period PTP.
  • PTP can be the past 1-3 months from the day on which the prediction feature space is constructed
  • PTF can be the next 1-2 weeks from the time or the day in which the prediction is made.
  • the beginning of PTF can be immediately or up to a day or a few days from the end of PTP.
  • the prediction unit 130 further includes a model calibration module 132 configured to train calibration models and a prediction calibration module 133 configured to calibrate the general predictions to suit specific campaigns.
  • FIG. 18A is a flowchart illustrating a method 1801 performed by system 100 to machine train a calibration model for an information campaign in-flight (e.g., after the information campaign has been running for a certain amount of time) according to some embodiments.
  • method 1801 includes obtaining (1801) a general prediction model suitable for the information campaign using, for example, method 1100 and the prediction unit 120.
  • the general prediction model is trained with respect to one or more predefined locations or places associated with the information campaign and can be used to predict probabilities of mobile devices in a general audience that may or may not have been qualified for the information campaign and/or impressed with information associated with the information campaign to visit the one or more predefined locations or places.
  • method 1801 further includes applying (1820) the prediction model to feature sets of qualified mobile devices (e.g., mobile devices associated with information requests qualified for the information campaign) to obtain predicted visitation probabilities.
  • a plurality of information requests are successively received (1822) by the front end server 111 during a time period TP shortly after the campaign has started (e.g., in the first day of the campaign), as shown in FIG 12 A.
  • Each respective request of the plurality of information requests includes a respective time stamp, respective location data indicating a location of a respective mobile device at a time indicated by the respective time stamp, and non-location data.
  • the plurality of information requests are successively processed by the front end server 111 as they are received, resulting in at least some of mobile devices identified in the plurality of information requests being qualified for the information campaign.
  • a mobile device is qualified for the information campaign when an information request identifying the mobile device is qualified for the information campaign, and the information request is qualified for the information campaign when the non-location data and/or location data meet a set of criteria defined by a set of campaign parameters of the information campaign.
  • the mobile device associated with a qualified information request is referred to here as a qualified mobile device.
  • a feature set corresponding to a mobile device associated with the qualified request can be retrieved from the features/labels database 125 shown in FIG. 1.
  • the feature set may be constructed from processed requests associated with the mobile device in the request database and feedback data associated with the mobile device from the feedback database.
  • the feature set is updated periodically (e.g., daily) so that it is constructed from such data with time stamps in a most recent prediction time period PTP that is close to the time t c of the request (e.g., 1-3 months before start of the information campaign), as shown in FIGS. 9, 10 and 12 A.
  • the feature set can be used to generally predict a probability of the mobile device to have at least one location event at any of the one or more predefined locations or places associated with the information campaign during a corresponding prediction time frame PTF (e.g., 1-2 weeks after start of the information campaign), as shown in FIGS. 9, 10, and 12A.
  • PTF prediction time frame
  • Method 1801 further includes applying (1824) by the prediction module 131 the general prediction model to respective feature sets corresponding to respective qualified mobile devices associated with respective qualified requests among the plurality of requests to obtain predicted probabilities for the respective qualified mobile devices to have location events at any of the one or more predefined locations or places associated with the information campaign during the prediction time frame PTF.
  • a respective feature set is constructed using data in respective datasets stored in the one or more databases. The respective datasets are associated with a respective qualified mobile device and including time stamps in the corresponding prediction time period (e.g., time period PTP).
  • Method 1801 further includes presenting the qualified information requests for fulfilment.
  • the front end server 111 presents a qualified information request for fulfillment by first placing a bid for the qualified information request at the source of the information request (e.g., an ad exchange or a MSP), and in response to the bid being accepted, sending the corresponding processed request to the ad server for fulfillment.
  • the bid includes a bid price that is dependent on a predicted probability corresponding to the mobile device associated with the qualified request.
  • method 1801 further includes receiving (1830) by the document server 112 feedback data indicating at least a portion of the qualified mobile devices having been impressed with information associated with the information campaign, responsive to at least a portion of the qualified requests among the plurality of information requests.
  • Method 1801 further includes determining (1840) by the model calibration module 132 detected visitation rates corresponding to respective ranges of predicted probabilities of the impressed mobile devices.
  • the model calibration module 132 searches in the mobile device database 124 to determine which of the qualified mobile devices have been impressed with information associated with the information campaign and which of the impressed mobile devices have had location events (or visitation events) at any of the one or more predefined locations or places during relevant time frames.
  • FIG. 19 is a table illustrating impression events and detected visitation events associated with certain mobile devices.
  • the impressed mobile devices are then divided based on their respective predicted probabilities, as determined in process 1820, into a plurality of probability brackets. For example, as shown in FIG 20, a first portion of the impressed mobile devices having predicted probabilities larger than 0% and smaller than or equal to 1% are allocated to a first probability bracket, a second portion of the impressed mobile devices having predicted probabilities larger than 1% and smaller than or equal to 2% are allocated to a second probability bracket, ... , and a 100th portion of the impressed mobile devices having predicted probabilities larger than 99% and smaller than or equal to 100% are allocated to a 100th probability bracket.
  • a detected visitation rate for the mobile devices in each of the plurality of probability brackets are then determined using, for example, subsequent request data associated with the mobile devices received during corresponding time frames.
  • each of the mobile devices in a probability bracket is labeled with “0” if there is no location event associated with the mobile device at any of the one or more predefined locations or places associated with the information campaign, or labeled as “1” if there is at least one location event associated with the mobile device at any of the one or more predefined locations or places, based on subsequently processed requests with time stamps during the respective time frame.
  • the detected visitation rate for a probability bracket can be, for example, a ratio of a number of mobile devices labeled with “1” in the probability bracket to a total number of the mobile device allocated to the probability bracket.
  • FIG. 21 includes a plot of detected visitation rates (thick solid line) and a plot of general predicted probabilities (thin solid line) corresponding to mobile devices in respective probability brackets. As shown, the predicted probabilities are significantly lower than the corresponding detected visitation rates for mobile devices in the 30 th to 75 th probability brackets, and significantly higher than the corresponding detected visitation rates for mobile devices in the 80 th to 100 th probability brackets.
  • the detected visitation rates and the general predicted probabilities are used to train (1850) a calibration model using, for example, isotonic regression, by the model calibration module 132.
  • the trained calibration model (plotted in thick dotted line) is stored in the calibration models database 135.
  • the prediction unit 130 further includes a prediction calibration module 133 configured to apply the calibration model to the general predicted probabilities to generate calibrated predictions (plotted as dashed line), which are much more in line with the detected visitation rates than the general predicted probabilities.
  • the calibration model trained using mobile device data in the prediction time period PTP and in the prediction time frame PTF is used by the prediction calibration module 133 to calibrate the general predictions generated by the prediction module 131 to obtain calibrated predictions, which are stored in the calibrated prediction data base.
  • the calibrated predictions are much more accurate in predicting visitations associated with mobile devices qualified in the information campaign and/or impressed with information associated with the information campaign.
  • a calibration model for a current information campaign can be cold-start trained before or shortly after the start of the current information campaign.
  • a campaign comparison model is first trained, which can be used to determine which historical campaign(s) is similar to the current information campaign.
  • FIG 18B illustrates a method 1802 of training a campaign comparison model according to some embodiments. As shown in FIG 18B, method 1802 includes obtaining a calibration model for each respective historical campaign of a plurality of historical campaigns. For example, the calibration model can be obtained for the respective historical campaign using method 1801 discussed above. As shown in FIG.
  • a prediction model can be trained for the historical campaign using a feature space constructed from mobile device data in a training period TTPh and a set of labels extracted from mobile device data in a training time frame TTFh before the start of the historical campaign, and the calibration model can be trained using a feature space constructed from mobile device data in a prediction period PTPh and mobile device data in a prediction time frame PTFh, as discussed above with reference to FIGS. 18A and 19-21.
  • FIG. 22 is a table illustrating campaign parameters of various historical campaigns.
  • the campaign parameters include identification of a pricing model (e.g., Cost-per-Click or CPC, Cost-per-Visit or CPV, Cost-per-Mille or Cost-per-Thousant-Impression or CPM, Cost-Per-Action, Cost-Per-Acquisition (CPA), Cost-Per-Lead (CPL), and Cost-Per- Installation (CPI)etc.).
  • the campaign parameters further include parameters related to a target audience, such as age range, gender type, education level, customer segment (e.g., loyalty or conquest), and geographical area (e.g., city, state, zip code, etc.).
  • the campaign parameters further include parameters defining a type of mobile devices (e.g., iPhone 8+, Galaxy S9+, etc.), a type of advertisement (e.g., banner, animation, interactive, video, etc.), a mobile apps in which the ad is displayed, etc.
  • the campaign parameters further include parameters related to one or more target locations, such as a brand, a type of geo-fence (e.g., radial, premise, etc.), a geographical area, etc.
  • method 1802 further includes constructing (1865) a set of campaign features for each of the plurality of historical campaigns.
  • the set of campaign features for a particular historical campaign can include features related to one or more target points of interest (POIs), such as a number of visits by mobile users per a specific length of time (e.g., one week) (e.g., as indicated in mobile device data), a percentage of the one or more POIs among all the POIs in the same brand/category, the type of geo-fences used for the target POIs, and a category of the POIs (e.g., general stores, department stores, grocery stores, apparel shops, sports shops, restaurants, shopping malls, theatres, etc.).
  • POIs target points of interest
  • the set of campaign features further includes features related to a target audience (e.g., age range, gender type, education level, customer segment, geographical area, etc.), and features related to mobile device type(s), ad type(s), mobile app(s), etc., that are specified in the historical campaign.
  • a target audience e.g., age range, gender type, education level, customer segment, geographical area, etc.
  • features related to mobile device type(s), ad type(s), mobile app(s), etc. that are specified in the historical campaign.
  • Method 1802 then proceeds to form (1870) campaign pairs, each campaign pair including two distinct historical campaigns among the plurality of historical campaigns.
  • FIG 24 which lists various campaign pairs formed using the plurality of historical campaigns and the campaign feature sets for each campaign pair.
  • the campaign pairs include some or all of the distinct campaign pairs that can be formed with the plurality of historical campaigns.
  • n is sufficiently large to provide sufficient data to train the campaign comparison model. For example, n > 100, or n > 1000.
  • method 1802 further comprises, for each campaign pair, constructing (1880) a set of campaign comparison features from the campaign features associated with the campaign pair, and extracting (1885) a difference value indicating a difference between the calibration models of the historical campaigns in the campaign pair.
  • the set of campaign comparison features can simply be the combination of the sets of campaign features of the campaign pair.
  • the difference value for a campaign pair including campaigns A and B can be measured by plotting the calibration models as curves CFi and CF2 associated with campaign A and campaign B, respectively, and sum up the areas between the two curves.
  • the campaign comparison features and the measured difference values of respective campaign pairs can then be used as features and labels to train (1890) a campaign comparison model by the model calibration module 132.
  • the trained campaign comparison model is stored in the calibration models database 135 and is used to determine a difference between the current campaign and each of a set of historical campaigns.
  • FIG. 18C is a flowchart illustrating a method 1803 of obtaining a calibration model for a current campaign using calibration model(s) of one or more similar historical campaigns according to some embodiments. As shown in FIG 18C, method 1803 includes forming (1890) a plurality of campaign pairs, each of the plurality of campaign pairs including the current campaign and the historical campaign of a set of historical campaigns.
  • the set of historical campaigns can be the same or a different set of historical campaigns used to train the comparison model.
  • Method 1803 further includes, for each respective campaign pair including the current campaign and a respective historical campaign, constructing (1892) a set of campaign features for the campaign pair, and applying (1894) the campaign comparison model to the set campaign features for the respective campaign pair to obtain a respective difference measure corresponding to the respective historical campaign, the respective difference measure indicating a predicted difference between the calibration model of the current campaign the calibration model of the respective historical campaign.
  • the set of campaign features can be the combination of campaign features associated with the historical campaign and campaign features associated with the current campaign.
  • the campaign features associated with the historical campaign can include features shown in FIG 23, and so can the campaign features associated with the current campaign.
  • FIG 26 also illustrates the predicted difference measures corresponding to respective historical campaigns.
  • method 1803 further includes selecting (1896) one or more similar campaigns from the set of historical campaigns based on the difference measures corresponding to the set of historical campaigns. For example, as shown in FIG 27, the set of historical campaigns can be ranked based on their predicted different measures with respect to the current campaign. The historical campaign(s) with the lowest similarity measure(s) is then selected as the similar campaign(s). For example, as shown in FIG. 27, campaigns C25682, C65248 and C32658 each has predicted difference measures that can be rounded up to 1% and is significantly lower than the next lowest difference measure of 2.09%. So, campaigns C25682, C65248 and C32658 in this example are selected as the similar campaigns. As shown in FIG.
  • method 1803 further includes obtaining (1898) the calibration model for the current campaign from one or more calibration models of the one or more similar campaigns. For example, if the one or more similar campaigns include just one similar campaign, the calibration model of the one similar campaign can be used as the calibration model for the current campaign. On the other hand, if the one or more similar campaigns includes multiple similar campaigns, the calibration model for the current campaign can be obtained by taking an average of the calibration models of the multiple similar campaigns.
  • FIG. 28 is a flowchart illustrating a method 2800 of conducting the current campaign according to some embodiments.
  • method 2800 includes obtaining (2810) a prediction model for the current campaign, and obtaining (2820) a calibration model for the current campaign based on a set of campaign parameters of the current campaign.
  • the prediction model can be trained with respect to one or more predefined locations or places associated with the current campaign, as described above with references to FIGS. 11-17.
  • the calibration model for the current campaign can be trained before the start of the current campaign, as described above with references to FIGS. 18B-18C and 22-27.
  • the calibration model can also be trained during the campaign, as described above with references to FIGS. 18A and 19-21.
  • method 2800 proceeds to process information requests during the current campaign, and to determine (2830), for each respective request received during the current campaign, if the respective request qualifies for the current campaign.
  • method 2800 includes, in response to the respective request qualifying for the information campaign, presenting (2850) the respective request for fulfillment.
  • presenting the respective request for fulfillment may include, for example, predicting a probability for a respective mobile device associated with the respective request to have at least one location event at any of the one or more locations associated with the information campaign during a specific time frame.
  • the specific time frame can be a time frame of a predefined duration (e.g., 1-2 weeks) from the time of the request.
  • the specific time frame PTFi can be 1-2 weeks from time ti, or from a preset time (e.g., midnight) on the day of time ti.
  • the prediction model is applied to a set of features associated with the mobile device and constructed from mobile device data associated with the mobile device with time stamps in a specific time period PTPi, which can be, for example, the past three months before the day of time ti.
  • PTPi a specific time period
  • the specific time frame PTF2 and the specific time period PTP2 can be different for another request with time stamp t2 on a different day, as shown in FIG 12B.
  • method 2800 further includes calibrating (2864) the prediction result by applying the calibration model to the prediction result to obtain a calibrated prediction probability, and presenting (2870) the respective request for fulfillment based at least on the calibrated prediction probability.
  • the front end server 111 presents the respective request for fulfillment by placing a bid for the respective request at the exchange or MSP, and, in the case the bid is accepted, transmitting the annotated request to the document (or information) server in the system 100.
  • the front end server 111 has access to the location predictions stored in a calibrated predictions database 136 in the prediction unit 130.
  • the bid can include a bid price that is dependent on the calibrated prediction probability corresponding to the mobile device associated with the qualified request.

Abstract

L'invention concerne un système comprenant un module d'apprentissage automatique configuré pour apprendre un modèle de prédiction d'emplacement à l'aide des caractéristiques construites à partir des données de dispositif mobile avec des horodatages durant une période d'apprentissage, et des étiquettes extraites des données de dispositif mobile avec des horodatages durant une période d'apprentissage. Le système comprend également un module de prédiction configuré pour appliquer le modèle de prédiction à un ensemble de caractéristiques construit à l'aide des données de dispositifs mobiles associées à un dispositif mobile avec des horodatages durant une période de prédiction afin d'obtenir un résultat de prédiction correspondant au dispositif mobile. Le système comprend également un module d'étalonnage configuré pour obtenir un modèle d'étalonnage correspondant à une campagne d'informations, ainsi qu'un module de prédiction étalonnée configuré pour appliquer le modèle d'étalonnage au résultat de prédiction afin d'obtenir une probabilité étalonnée que le dispositif mobile ait au moins un événement de localisation à un emplacement quelconque d'un ou de plusieurs emplacements associés à la campagne d'informations pendant une période de prédiction.
PCT/US2020/066982 2019-12-23 2020-12-23 Systèmes et procédés de prédiction d'emplacement étalonnée WO2021133997A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US16/726,056 US11134359B2 (en) 2018-08-17 2019-12-23 Systems and methods for calibrated location prediction
US16/726,056 2019-12-23

Publications (1)

Publication Number Publication Date
WO2021133997A1 true WO2021133997A1 (fr) 2021-07-01

Family

ID=76575141

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2020/066982 WO2021133997A1 (fr) 2019-12-23 2020-12-23 Systèmes et procédés de prédiction d'emplacement étalonnée

Country Status (1)

Country Link
WO (1) WO2021133997A1 (fr)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120276928A1 (en) * 2011-04-29 2012-11-01 Jon Shutter Method and System for Providing Location Targeted Advertisements
US10349208B1 (en) * 2018-08-17 2019-07-09 xAd, Inc. Systems and methods for real-time prediction of mobile device locations

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120276928A1 (en) * 2011-04-29 2012-11-01 Jon Shutter Method and System for Providing Location Targeted Advertisements
US10349208B1 (en) * 2018-08-17 2019-07-09 xAd, Inc. Systems and methods for real-time prediction of mobile device locations

Similar Documents

Publication Publication Date Title
JP7084970B2 (ja) マーケティング上のモバイル広告供給に関するシステムと方法
US10715962B2 (en) Systems and methods for predicting lookalike mobile devices
US10939233B2 (en) System and method for real-time prediction of mobile device locations
US11683655B2 (en) Systems and methods for predicting mobile device locations using processed mobile device signals
US11134359B2 (en) Systems and methods for calibrated location prediction
US10278014B2 (en) System and method for using geo-blocks and geo-fences to predict mobile device locations
JP7214085B2 (ja) オンライン及びオフライン予測を使用してモバイル装置への情報伝達を制御する処理
US8788328B1 (en) Location affinity based content delivery systems and methods
US11743679B2 (en) Systems and methods for pacing information delivery to mobile devices
JP7285521B2 (ja) 類似のモバイル装置を予測するためのシステムと方法
WO2017062912A2 (fr) Procédé et appareil pour mesurer l'effet d'informations fournies à des dispositifs mobiles
US11146911B2 (en) Systems and methods for pacing information campaigns based on predicted and observed location events
JP2022079394A (ja) ターゲテッド位置イベント予測システム及び方法
WO2021133997A1 (fr) Systèmes et procédés de prédiction d'emplacement étalonnée
JP2021120855A (ja) 予測及び観測された位置イベントに基づいて情報キャンペーンをペーシングするシステム及び方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20906633

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20906633

Country of ref document: EP

Kind code of ref document: A1