CN111564200A - Old people diet feature extraction device and method based on rapid random gradient descent - Google Patents

Old people diet feature extraction device and method based on rapid random gradient descent Download PDF

Info

Publication number
CN111564200A
CN111564200A CN202010383353.XA CN202010383353A CN111564200A CN 111564200 A CN111564200 A CN 111564200A CN 202010383353 A CN202010383353 A CN 202010383353A CN 111564200 A CN111564200 A CN 111564200A
Authority
CN
China
Prior art keywords
data
diet
model
training
initializing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010383353.XA
Other languages
Chinese (zh)
Inventor
张能锋
秦雯
许明
罗辛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Wanjia'an Artificial Intelligence Data Technology Co ltd
Original Assignee
Shenzhen Wanjia'an Artificial Intelligence Data Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Wanjia'an Artificial Intelligence Data Technology Co ltd filed Critical Shenzhen Wanjia'an Artificial Intelligence Data Technology Co ltd
Priority to CN202010383353.XA priority Critical patent/CN111564200A/en
Publication of CN111564200A publication Critical patent/CN111564200A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H20/00ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
    • G16H20/60ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to nutrition control, e.g. diets
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/30ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for calculating health indices; for individual health risk assessment

Landscapes

  • Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Public Health (AREA)
  • Engineering & Computer Science (AREA)
  • Primary Health Care (AREA)
  • General Health & Medical Sciences (AREA)
  • Epidemiology (AREA)
  • Nutrition Science (AREA)
  • Biomedical Technology (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Pathology (AREA)
  • Medical Treatment And Welfare Office Work (AREA)

Abstract

The invention discloses a device and a method for extracting diet characteristics of old people based on rapid random gradient descent, wherein the prediction device comprises a data receiving module, a data processing module and a data processing module, wherein the data receiving module is configured to collect old people-diet scoring data; a data storage module configured to store senior-diet scoring data and senior-diet feature data; the characteristic rapid extraction module is configured to train and extract diet characteristics through known scores of the old people on daily diet and transmit the extracted characteristic data to the characteristic output module; the rapid feature extraction module comprises a data initialization unit, a data division unit and a rapid data training unit. The prediction device and the prediction method provided by the invention execute the data block by block in parallel through the momentum combined with the parallel random gradient descent of the random gradient descent strategy, thereby not only accelerating the convergence speed of the model, but also reducing the training time of the model, and being widely applied to a recommendation platform for providing personalized services.

Description

Old people diet feature extraction device and method based on rapid random gradient descent
Technical Field
The invention relates to the technical field of computer data mining, in particular to a device and a method for extracting diet characteristics of old people based on rapid random gradient descent.
Background
With the development of social science and technology and economy, the population is aging seriously, the burden of the labor age population is increased, the number of the elderly living alone is increased, and the healthy diet management platform for the elderly attracts social attention. The number of the old people is increasing, the nutrition diet collocation types are rich, and the old people or nursing personnel for taking care of the diet of the old people are difficult to select the most suitable diet collocation among a plurality of diet collocations. Therefore, an old-person-diet scoring matrix can be formed by scoring different diets by the old person, the preference degree of the old person to different diets is predicted by utilizing the scoring height, the characteristics of different diets are obtained, and a healthy diet matching package is recommended according to the preference. The daily eating types are more and more, the old people have many diets and do not try, the evaluation is not carried out, the characteristic can be rapidly extracted, and the problem of providing healthy diet matching according to the preference of the old people becomes the concern of each family.
Disclosure of Invention
The invention aims to provide a device and a method for extracting the diet characteristics of old people with rapid random gradient decline, which are used for extracting different diet characteristics of the old people according to the historical evaluation of the old people on daily diet.
In order to solve the above-mentioned technical problem, according to a first aspect of the present invention, there is provided an old people diet feature extraction device based on fast random gradient descent, the device comprising:
the data receiving module is used for acquiring old people-diet scoring data from the old people health management platform and transmitting the acquired data into the data preprocessing module;
the data preprocessing module is used for processing the received data into a data form which can be directly used and transmitting the data into the data storage module;
the data storage module is used for storing the acquired known grading data and storing a hidden feature matrix of the old and the daily diet types;
the characteristic rapid extraction module is used for initializing data, constructing an old people diet characteristic extraction model with rapid random gradient descent and performing data training;
and the characteristic output module outputs the prediction data of the unknown old people-diet scores and the implicit characteristics of the old people and diet.
Further, the rapid feature extraction module comprises a data initialization unit, a model component unit and a rapid data training unit, wherein the data initialization unit is used for initializing relevant parameters related to the old people diet feature extraction process based on rapid random gradient descent; the model construction unit constructs an old people diet characteristic extraction model with rapid random gradient descent, which can be executed in parallel, by using a strategy of combining momentum and random gradient descent; the data rapid training unit is used for training to obtain old people diet characteristic data;
according to the preferable scheme of the old people diet feature extraction device based on the rapid stochastic gradient descent, the data preprocessing module specifically comprises:
the collected data is processed into a triplet form, which may be expressed as P ═ x, y, r, where l denotes geriatric, y denotes diet, and r denotes geriatric x scores diet y. Storing the processed data into a data storage module;
according to the preferable scheme of the old people diet characteristic extraction device based on the rapid random gradient descent, the data storage module comprises a diet grading data storage unit and an old people diet characteristic storage unit.
The diet scoring data storage unit is configured to establish a matrix of | X | rows and | Y | columns as an old people-diet scoring matrix R for storing known scoring data, wherein | X | is a set of old people and | Y | is a set of diet.
The elderly diet feature storage unit is configured to store elderly diet feature data. Storing feature data of old people and diet in a matrix form, and storing the feature data by using U, I two hidden feature matrices, wherein the feature dimensions of the two hidden feature matrices are the same as d.
According to the preferred scheme of the old people diet feature extraction device based on the rapid stochastic gradient descent, the feature rapid extraction module comprises a data initialization unit, a model construction unit and a data rapid training unit.
The data initialization unit is configured to initialize elderly people and mealsTwo hidden feature matrices U and I; initializing a hidden feature space dimension d; initializing a regularization factor λ2The method comprises the steps of initializing a node number N of parallel calculation, initializing a data division segment number b, initializing a maximum iteration round number T, initializing an iteration round number control variable T in a training process, initializing a momentum acceleration factor gamma, initializing a model training learning step η, and initializing a convergence termination threshold value a, wherein d determines a feature space dimension of each hidden feature matrix and initializes the hidden feature matrix to be a positive integer, in the hidden feature matrices U and I, U is a hidden feature matrix with | L | row d column and I is a hidden feature matrix with | Y | row d column and initializes the hidden feature matrices with random smaller positive numbers respectively, the maximum training iteration round number T is a variable for controlling an upper limit of training times and initializes to be a larger positive integer, the iteration round number control variable T is initialized to be 1, and a regularization factor lambda is normalized2The method comprises the steps of determining the number of nodes of parallel calculation of a model, initializing the nodes to be positive integers, determining the number b of data division sections, determining the number b of data sections into which a data set is divided, wherein each data section is provided with b data blocks which are independent of each other and do not share a row or a column, initializing the data sections to be positive integers, determining the acceleration effect of momentum on gradient descent, accelerating the convergence speed of the model, initializing the data sections to be small positive numbers, determining the learning step size of model training by the model training learning step size η, initializing the data sections to be small positive numbers, and judging whether the training is finished or not by the convergence termination threshold value a which is a parameter and is initialized by small positive numbers.
The model construction unit is configured to construct an old people diet characteristic extraction model with rapid random gradient descent, and an objective function RLY of the model is constructed according to a known data set Q in an old people diet scoring matrix RQThe formula is as follows:
Figure BDA0002482886770000041
wherein Q represents a known scoring data set of daily diet for the elderly in the elderly-daily diet scoring matrix R; r isx,yThe meaning is x-to-daily drinking for the agedThe preference degree of the food y is the preference score of the old x on the daily food y; u. ofx′Representing a hidden eigenvalue corresponding to the xth old man in the old man hidden eigenvalue matrix U; i.e. iy′And (4) representing the implicit characteristic value corresponding to the y-th diet in the daily diet implicit characteristic matrix I.
To prevent overfitting of the model, the generalization performance of the model is enhanced, usually in the objective function RLYQAdding L2The regularization term constrains the model, which can be given by the following equation:
Figure DEST_PATH_3
wherein λ2Is a regularization term control parameter, λ2Represents the measure L2The limiting effect of the regularization term on the model.
The data rapid training unit is configured to combine the target function constructed by the model construction unit, rapidly train data and extract characteristic values. In order to satisfy U and I for accumulating error RLY on setQAt a minimum, the cumulative error RLY is computed using a fast converging stochastic gradient descent optimization algorithmQAnd training the matrixes U and I to obtain the global optimal solution of the matrixes U and I. In the training process, the selection of the hidden characteristics of the old and the daily diet types is adopted to quickly update the direction, the linear combination of the current gradient and the last update gradient of the parameters is realized according to the learning step length and the momentum acceleration factor, and the current update u is calculatedx′And iy′The formula expression of the gradient value corresponding to the fastest convergence speed is as follows:
Figure BDA0002482886770000051
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e., the kth implicit feature of geriatric x and diet y.
Figure BDA0002482886770000052
And
Figure BDA0002482886770000053
respectively, the current gradient value of the t time. The expression is as follows:
Figure BDA0002482886770000054
selecting the calculated gradient value with the fastest convergence according to the fast convergence direction, and obtaining a parameter updating rule of an old people diet characteristic extraction model based on fast random gradient decline:
about parameter ux,kThe update rule of (2) is as follows:
Figure BDA0002482886770000055
with respect to parameter iy,kThe update rule of (2) is as follows:
Figure BDA0002482886770000056
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e., the kth implicit feature of geriatric x and diet y.
Figure BDA0002482886770000061
And
Figure BDA0002482886770000062
respectively, the current gradient value of the t time.
The following new rule can accelerate the convergence speed of the model, in order to further accelerate the calculation efficiency of the model, the data is divided into b data segments before the model is updated, b blocks are arranged in each data segment, the blocks are irrelevant, and meanwhile, b calculation nodes are used for respectively and simultaneously training b independent blocks, so that the high-efficiency calculation speed is obtained. And executing a parameter updating rule in each block until all data block training is finished, which represents that the iteration is finished.
Repeating the training process on the known data set until the model is converged, wherein the convergence judgment condition is that the iteration number T reaches the maximum training iteration number T or RLY obtained by calculation of the current iteration and the last iteration isQThe absolute value of the difference in values is less than the convergence termination threshold a.
According to the preferred embodiment of the device for extracting the dietary characteristics of the elderly based on rapid stochastic gradient descent, the data output module comprises:
the output unit outputs a target loss function RLYQAnd when the minimum two implicit feature matrixes are reached, the unknown score predicting unit predicts the unknown score in the old man-diet matrix by using the implicit feature matrixes obtained by the output unit.
According to a second aspect of the present invention, there is provided a method for extracting dietary characteristics of elderly people with rapid stochastic gradient descent, comprising the following steps:
s1: and receiving old people diet scoring data of the information acquisition terminal, and sending the old people diet scoring data to an old people diet feature extraction device based on rapid random gradient descent. The old people diet scoring data is real scoring data of diets contained in a nutritional package after the old people select the nutritional package matching on the old people health management platform.
S2: for known data, it is stored in the form of a triplet, i.e. P ═ x, y, r, where l denotes geriatric, y denotes diet, and r denotes geriatric x scores diet y.
S3: initializing relevant parameters in the process of extracting the diet characteristics of the old.
S4: and combining the known acquired data and the initialized related parameters to construct an objective loss function.
S5: and updating the hidden characteristics of the old and the diet according to the constructed target loss function, and obtaining two characteristic matrixes when the model reaches convergence.
S6: two feature matrices are output.
According to the preferred embodiment of the method for extracting dietary characteristics of elderly people with fast random gradient descent, step S3 is used for initializing relevant parameters in the process of extracting dietary characteristics of elderly people, and includes:
initializing hidden feature matrixes U and I; initializing a hidden feature space dimension d; initializing a regularization factor λ2The method comprises the steps of initializing a node number N of parallel calculation, initializing a data division segment number b, initializing a maximum iteration round number T, initializing an iteration round number control variable T in a training process, initializing a momentum acceleration factor gamma, initializing a model training learning step length η, and initializing a convergence termination threshold value a;
wherein:
the hidden feature dimension d determines the spatial dimensions of the feature matrices U and I, and is initialized to a positive integer.
Size of hidden feature matrices U and I: the implicit feature matrix with U being | L | row and d column and the implicit feature matrix with I being | Y | row and d column are initialized by random small positive numbers.
The maximum training iteration round number T is a variable for controlling the upper limit of the training times and is initialized to be a larger positive integer. Initializing an iteration round number control variable t to 1;
regularization factor lambda2The method is used for measuring the limiting effect of the regularization term on the model and is initialized to be a smaller positive number.
Determining the number of nodes of the model parallel computation by the number N of the nodes of the parallel computation, and initializing the number of the nodes of the model parallel computation into a positive integer;
the data dividing segment number b is used for determining that a data set is divided into b data segments, wherein b data blocks are arranged in each data segment, the b data blocks are independent from each other, do not share rows or columns, and are initialized to be positive integers;
the momentum acceleration factor gamma is used for determining the acceleration effect of momentum on gradient reduction, so that the convergence speed of the model is accelerated and the model is initialized to be a smaller positive number;
the learning step length eta of the model training determines the learning step length of the model training and is initialized to a smaller positive number;
the convergence termination threshold a is a parameter for determining whether or not training is terminated, and is initialized with a very small positive number.
According to the preferred embodiment of the method for extracting the dietary characteristics of the elderly with fast stochastic gradient descent, step S4 combines the known acquired data and the initialized relevant parameters to construct an objective loss function, which includes:
constructing an objective function RLY of the model according to a known data set Q in the old people-diet scoring matrix RQThe formula is as follows:
Figure BDA0002482886770000081
wherein Q represents a known scoring data set of daily diet for the elderly in the elderly-daily diet scoring matrix R; r isx,yThe meaning is that the preference degree of the old people x to the daily diet y is the preference score of the old people x to the daily diet y; u. ofx′Representing a hidden eigenvalue corresponding to the xth old man in the old man hidden eigenvalue matrix U; i.e. iy′And (4) representing the implicit characteristic value corresponding to the y-th diet in the daily diet implicit characteristic matrix I.
To prevent overfitting of the model, the generalization performance of the model is enhanced, usually in the objective function RLYQAdding L2The regularization term constrains the model, which can be given by the following equation:
Figure 893902DEST_PATH_3
wherein λ2Is a regularization term control parameter, λ2Represents the measure L2The limiting effect of the regularization term on the model.
According to the preferred scheme of the old people diet characteristic extraction method with the rapid random gradient descent, step S5 is used for solving an objective function extraction characteristic matrix. The method comprises the following steps:
and S51, dividing the data into b data segments, wherein each data segment contains b data blocks and the b data blocks are not related to each other. And selecting one data segment every time, distributing the b data blocks to different computing nodes until all data are distributed and trained, and finishing one iteration.
And S52, different nodes acquire data blocks which are not related to each other, and simultaneously train and optimize the target loss function. In order to satisfy U and I for accumulating error RLY on setQAt a minimum, the cumulative error RLY is computed using a fast converging stochastic gradient descent optimization algorithmQAnd training the matrixes U and I to obtain the global optimal solution of the matrixes U and I. In the training process, the selection of the hidden characteristics of the old and the daily diet types is adopted to quickly update the direction, the linear combination of the current gradient and the last update gradient of the parameters is realized according to the learning step length and the momentum acceleration factor, and the current update u is calculatedx′And iy′The formula expression of the gradient value corresponding to the fastest convergence speed is as follows:
Figure BDA0002482886770000091
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e., the kth implicit feature of geriatric x and diet y.
Figure BDA0002482886770000092
And
Figure BDA0002482886770000093
respectively, the current gradient value of the t time. The expression is as follows:
Figure BDA0002482886770000094
selecting the calculated gradient value with the fastest convergence according to the fast convergence direction, and obtaining a parameter updating rule of an old people diet characteristic extraction model based on fast random gradient decline:
about parameter ux,kThe update rule of (2) is as follows:
Figure BDA0002482886770000101
with respect to parameter iy,kThe update rule of (2) is as follows:
Figure BDA0002482886770000102
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e., the kth implicit feature of geriatric x and diet y.
Figure BDA0002482886770000103
And
Figure BDA0002482886770000104
respectively, the current gradient value of the t time.
And S53, judging whether the training of all the data blocks is finished, if the data is not completely processed in the current iteration, returning to S51, and acquiring the unassigned data blocks to continue the training. If the training is completed, the process proceeds to S54.
S54, judgmentJudging whether the iteration end condition is met, wherein the convergence judgment condition is that the iteration round number T reaches the maximum training iteration round number T or RLY obtained by the calculation of the current iteration and the last iterationQThe absolute value of the difference in values is less than the convergence termination threshold a.
The invention discloses an old people diet feature extraction device and method based on rapid random gradient descent, and aims to obtain a feature matrix of old people diet by utilizing a fast-calculated reduction matrix factorization method, and provide healthy and nutritional diet matching service which accords with the favor of the old people for the old people.
Drawings
Fig. 1 is a prediction device for elderly people diet feature extraction based on rapid stochastic gradient descent in the invention.
Fig. 2 is a schematic flow chart of the process for extracting and updating the dietary characteristics of the elderly based on rapid stochastic gradient descent.
Detailed Description
The present invention will be described in further detail with reference to test examples and specific embodiments. It should be understood that the scope of the above-described subject matter is not limited to the following examples, and any techniques implemented based on the disclosure of the present invention are within the scope of the present invention.
Example 1:
referring to fig. 1, there is shown an apparatus for senior diet feature extraction based on fast stochastic gradient descent of the present invention, comprising:
the data receiving module 210 acquires elderly-eating scoring data from the elderly health management platform and transmits the acquired data to the data preprocessing module;
the data preprocessing module 220 is used for processing the received data into a data form which can be directly used and transmitting the data into the data storage module;
the data storage module 230 is used for storing the acquired known grading data and storing a hidden feature matrix of the old and the daily diet types;
the characteristic fast extraction module 240 is used for initializing data, constructing an old people diet characteristic extraction model with fast random gradient descent and performing data training;
and the characteristic output module 250 outputs the prediction data of the unknown old people-diet score and the implicit characteristics of the old people and diet.
Further, the rapid feature extraction module comprises a data initialization unit 241, a model component unit 242 and a data rapid training unit 243, wherein the data initialization unit 241 is used for initializing relevant parameters involved in the elderly diet feature extraction process based on rapid stochastic gradient descent; the model construction unit 242 constructs an old people diet feature extraction model with rapid random gradient descent, which can be executed in parallel, by using a strategy of combining momentum and random gradient descent; the data fast training unit 243 is used for training to obtain old people diet characteristic data;
in a specific embodiment, the data preprocessing module 220 specifically includes:
the collected data is processed into a triplet form, which may be expressed as P ═ x, y, r, where l denotes geriatric, y denotes diet, and r denotes geriatric x scores diet y. Storing the processed data into a data storage module;
in a specific embodiment, the data storage module 230 includes a diet scoring data storage unit 231 and an elderly diet feature storage unit 232.
The diet scoring data storage unit 231 is configured to establish a matrix of | X | rows and | Y | columns as the senior-diet scoring matrix R, and is configured to store known scoring data, where | X | is a set of senior citizens and | Y | is a set of diets.
The elderly diet feature storage unit 232 is used for storing elderly diet feature data. Storing feature data of old people and diet in a matrix form, and storing the feature data by using U, I two hidden feature matrices, wherein the feature dimensions of the two hidden feature matrices are the same as d.
In a specific embodiment, the feature fast extraction module 240 includes a data initialization unit 241, a model construction unit 242, and a data fast training unit 243.
The data initialization unit 241 is used for initializing two hidden feature matrixes U and I of the old and the diet; initializing a hidden feature space dimension d; initializing a regularization factor λ2The method comprises the steps of initializing a node number N of parallel calculation, initializing a data division segment number b, initializing a maximum iteration round number T, initializing an iteration round number control variable T in a training process, initializing a momentum acceleration factor gamma, initializing a model training learning step η, and initializing a convergence termination threshold value a, wherein d determines a feature space dimension of each hidden feature matrix and initializes the hidden feature matrix to be a positive integer, in the hidden feature matrices U and I, U is a hidden feature matrix with | L | row d column and I is a hidden feature matrix with | Y | row d column and initializes the hidden feature matrices with random smaller positive numbers respectively, the maximum training iteration round number T is a variable for controlling an upper limit of training times and initializes to be a larger positive integer, the iteration round number control variable T is initialized to be 1, and a regularization factor lambda is normalized2The method comprises the steps of determining the number of nodes of parallel calculation of a model, initializing the nodes to be positive integers, determining the number b of data division sections, determining the number b of data sections into which a data set is divided, wherein each data section is provided with b data blocks which are independent of each other and do not share a row or a column, initializing the data sections to be positive integers, determining the acceleration effect of momentum on gradient descent, accelerating the convergence speed of the model, initializing the data sections to be small positive numbers, determining the learning step size of model training by the model training learning step size η, initializing the data sections to be small positive numbers, and judging whether the training is finished or not by the convergence termination threshold value a which is a parameter and is initialized by small positive numbers.
The model construction unit 242 is used for constructing an old people diet feature extraction model with fast random gradient descent, and constructing a target function RLY of the model according to a known data set Q in an old people-diet scoring matrix RQThe formula is as follows:
Figure BDA0002482886770000131
wherein Q represents a known scoring data set of daily diet for the elderly in the elderly-daily diet scoring matrix R; r isx,yThe meaning is that the preference degree of the old people x to the daily diet y is the preference score of the old people x to the daily diet y; u. ofx′Representing hidden characteristic moments of the agedHidden characteristic values corresponding to the x-th old man in the array U; i.e. iy′And (4) representing the implicit characteristic value corresponding to the y-th diet in the daily diet implicit characteristic matrix I.
To prevent overfitting of the model, the generalization performance of the model is enhanced, usually in the objective function RLYQAdding L2Regularization constrains the model, which can be given by the following equation:
Figure 811042DEST_PATH_3
wherein λ2Is a regularization control parameter, λ2Represents the measure L2Regularization limiting the effect on the model.
The data fast training unit 243 is configured to fast train data and extract feature values in combination with the objective function constructed by the model construction unit. In order to satisfy U and I for accumulating error RLY on setQAt a minimum, the cumulative error RLY is computed using a fast converging stochastic gradient descent optimization algorithmQAnd training the matrixes U and I to obtain the global optimal solution of the matrixes U and I. In the training process, the selection of the hidden characteristics of the old and the daily diet types is adopted to quickly update the direction, the linear combination of the current gradient and the last update gradient of the parameters is realized according to the learning step length and the momentum acceleration factor, and the current update u is calculatedx′And iy′The formula expression of the gradient value corresponding to the fastest convergence speed is as follows:
Figure BDA0002482886770000142
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e. the kth implicit of geriatric x and diet yAnd (5) characterizing.
Figure BDA0002482886770000143
And
Figure BDA0002482886770000144
respectively, the current gradient value of the t time. The expression is as follows:
Figure BDA0002482886770000151
selecting the calculated gradient value with the fastest convergence according to the fast convergence direction, and obtaining a parameter updating rule of an old people diet characteristic extraction model based on fast random gradient decline:
about parameter ux,kThe update rule of (2) is as follows:
Figure BDA0002482886770000152
with respect to parameter iy,kThe update rule of (2) is as follows:
Figure BDA0002482886770000153
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e., the kth implicit feature of geriatric x and diet y.
Figure BDA0002482886770000154
And
Figure BDA0002482886770000155
respectively, the current gradient value of the t time.
The following new rule can accelerate the convergence speed of the model, in order to further accelerate the calculation efficiency of the model, the data is divided into b data segments before the model is updated, b blocks are arranged in each data segment, the blocks are irrelevant, and meanwhile, b calculation nodes are used for respectively and simultaneously training b independent blocks, so that the high-efficiency calculation speed is obtained. And executing a parameter updating rule in each block until all the data blocks are trained to end, which represents the end of the iteration.
Repeating the training process on the known data set until the model is converged, wherein the convergence judgment condition is that the iteration number T reaches the maximum training iteration number T or RLY obtained by calculation of the current iteration and the last iteration isQThe absolute value of the difference in values is less than the convergence termination threshold a.
In a specific embodiment, the data output module 250 specifically includes:
output target loss function RLYQAnd when the minimum two implicit feature matrixes are reached, the unknown score predicting unit predicts the unknown score in the old man-diet matrix by using the implicit feature matrixes obtained by the output unit.
Example 2:
referring to fig. 2, fig. 2 shows the method for extracting dietary characteristics of the elderly based on rapid stochastic gradient descent according to the present invention, wherein the prediction method comprises the following steps:
s1: and receiving old people diet scoring data of the information acquisition terminal, and sending the old people diet scoring data to an old people diet feature extraction device based on rapid random gradient descent. The old people diet scoring data is real scoring data of diets contained in a nutritional package after the old people select the nutritional package matching on the old people health management platform.
S2: for known data, it is stored in the form of a triplet, i.e. P ═ x, y, r, where l denotes geriatric, y denotes diet, and r denotes geriatric x scores diet y.
S3: initializing relevant parameters in the process of extracting the diet characteristics of the old.
S4: and combining the known acquired data and the initialized related parameters to construct an objective loss function.
S5: and updating the hidden characteristics of the old and the diet according to the constructed target loss function, and obtaining two characteristic matrixes when the model reaches convergence.
S6: two feature matrices are output.
In a specific embodiment, step S3 is used to initialize relevant parameters in the process of extracting dietary features of the elderly, including:
initializing hidden feature matrixes U and I; initializing a hidden feature space dimension d; initializing a regularization factor λ2The method comprises the steps of initializing a node number N of parallel calculation, initializing a data division segment number b, initializing a maximum iteration round number T, initializing an iteration round number control variable T in a training process, initializing a momentum acceleration factor gamma, initializing a model training learning step length η, and initializing a convergence termination threshold value a;
wherein:
the hidden feature dimension d determines the spatial dimensions of the feature matrices U and I, and is initialized to a positive integer.
Size of hidden feature matrices U and I: the implicit feature matrix with U being | L | row and d column and the implicit feature matrix with I being | Y | row and d column are initialized by random small positive numbers.
The maximum training iteration round number T is a variable for controlling the upper limit of the training times and is initialized to be a larger positive integer. Initializing an iteration round number control variable t to 1;
regularization factor lambda2The method is used for measuring the limiting effect of regularization on the model and is initialized to be a small positive number. Determining the number of nodes of the model parallel computation by the number N of the nodes of the parallel computation, and initializing the number of the nodes of the model parallel computation into a positive integer;
the data dividing segment number b is used for determining that a data set is divided into b data segments, wherein b data blocks are arranged in each data segment, and the b data blocks are independent from each other, namely, the data segments do not share rows or columns and are initialized to be positive integers;
the momentum acceleration factor gamma is used for determining the acceleration effect of momentum on gradient reduction, so that the convergence speed of the model is accelerated and the model is initialized to be a smaller positive number;
the learning step length eta of the model training determines the learning step length of the model training and is initialized to a smaller positive number;
the convergence termination threshold a is a parameter for determining whether or not training is terminated, and is initialized with a very small positive number.
In a specific embodiment, step S4 combines the known acquired data and initialized relevant parameters to construct an objective loss function, which includes:
constructing an objective function RLY of the model according to a known data set Q in the old people-diet scoring matrix RQThe formula is as follows:
Figure BDA0002482886770000181
wherein Q represents a known scoring data set of daily diet for the elderly in the elderly-daily diet scoring matrix R; r isx,yThe meaning is that the preference degree of the old people x to the daily diet y is the preference score of the old people x to the daily diet y; u. ofx′Representing a hidden eigenvalue corresponding to the xth old man in the old man hidden eigenvalue matrix U; i.e. iy′And (4) representing the implicit characteristic value corresponding to the y-th diet in the daily diet implicit characteristic matrix I.
To prevent overfitting of the model, the generalization performance of the model is enhanced, usually in the objective function RLYQAdding L2Regularization constrains the model, which can be given by the following equation:
Figure 528463DEST_PATH_3
wherein λ2Is a regularization control parameter, λ2Represents the measure L2Regularization limiting the effect on the model.
In a specific embodiment, step S5 is implemented to solve the objective function extraction feature matrix, and includes:
and S51, dividing the data into b data segments, wherein each data segment contains b data blocks and the b data blocks are not related to each other. And selecting one data segment every time, distributing the b data blocks to different computing nodes until all data are distributed and trained, and finishing one iteration.
And S52, different nodes acquire data blocks which are not related to each other, and simultaneously train and optimize the target loss function. In order to satisfy U and I for accumulating error RLY on setQAt a minimum, the cumulative error RLY is computed using a fast converging stochastic gradient descent optimization algorithmQAnd training the matrixes U and I to obtain the global optimal solution of the matrixes U and I. In the training process, the selection of the hidden characteristics of the old and the daily diet types is adopted to quickly update the direction, the linear combination of the current gradient and the last update gradient of the parameters is realized according to the learning step length and the momentum acceleration factor, and the current update u is calculatedx′And iy′The formula expression of the gradient value corresponding to the fastest convergence speed is as follows:
Figure BDA0002482886770000191
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e., the kth implicit feature of geriatric x and diet y.
Figure BDA0002482886770000192
And
Figure BDA0002482886770000193
respectively, the current gradient value of the t time. The expression is as follows:
Figure BDA0002482886770000194
selecting the calculated gradient value with the fastest convergence according to the direction of the rapid convergence, and obtaining a parameter updating rule of an old diet characteristic extraction model based on rapid random gradient decline:
about parameter ux,kThe update rule of (2) is as follows:
Figure BDA0002482886770000195
with respect to parameter iy,kThe update rule of (2) is as follows:
Figure BDA0002482886770000201
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e., the kth implicit feature of geriatric x and diet y.
Figure BDA0002482886770000202
And
Figure BDA0002482886770000203
respectively, the current gradient value of the t time.
And S53, judging whether the training of all the data blocks is finished, if the data is not completely processed in the current iteration, returning to S51, and acquiring the unassigned data blocks to continue the training. If the training is completed, the process proceeds to S54.
S54, judging whether the iteration end condition is satisfied, wherein the convergence judgment condition is that the iteration round number T reaches the maximum training iteration round number T or RLY obtained by the calculation of the current iteration and the last iterationQThe absolute value of the difference in values is less than the convergence termination threshold a.
While embodiments of the invention have been shown and described, it will be understood by those of ordinary skill in the art that: various changes, modifications, substitutions and alterations can be made to the embodiments without departing from the principles and spirit of the invention, the scope of which is defined by the claims and their equivalents.

Claims (7)

1. The utility model provides an old man's diet feature extraction element based on quick random gradient descends, its characterized in that includes:
the data receiving module is used for acquiring old people-diet scoring data from the old people health management platform and transmitting the acquired data into the data preprocessing module;
the data preprocessing module is used for processing the received data into a data form which can be directly used and transmitting the data into the data storage module;
the data storage module is used for storing the acquired known grading data and storing a hidden feature matrix of the old and the daily diet types;
the characteristic rapid extraction module is used for initializing data, constructing an old people diet characteristic extraction model with rapid random gradient descent and performing data training;
and the characteristic output module outputs the prediction data of the unknown old people-diet scores and the implicit characteristics of the old people and diet.
2. The elderly diet feature extraction device based on fast stochastic gradient descent according to claim 1, wherein the data storage module comprises a diet score data storage unit and an elderly diet feature storage unit,
the diet scoring data storage unit is configured to: and establishing a matrix of | X | rows and | Y | columns as an old people-diet scoring matrix R for storing known scoring data, wherein | X | is a set of old people and | Y | is a set of diet.
The elderly diet feature storage unit is configured to: storing the diet characteristic data of the old. Storing feature data of old people and diet in a matrix form, and storing the feature data by using U, I two hidden feature matrices, wherein the feature dimensions of the two hidden feature matrices are the same as d.
3. The old people diet feature extraction device based on rapid stochastic gradient descent according to claim 2, wherein the feature rapid extraction module comprises a data initialization unit, a model construction unit and a data rapid training unit,
the data initialization unit is configured to: initializing two implicit feature matrixes U and I of the old and diet; initializing a hidden feature space dimension d; initializing a regularization factor λ2The method comprises the steps of initializing the number N of nodes calculated in parallel, initializing the number b of data division sections, initializing the maximum iteration round number T, initializing an iteration round number control variable T in the training process, initializing a momentum acceleration factor gamma, initializing a model training learning step η, initializing a convergence termination threshold value a, wherein d determines the feature space dimension of each hidden feature matrix, in the hidden feature matrices U and I, U is a hidden feature matrix with L and Y rows and d columns, I is a hidden feature matrix with Y and d columns, the maximum training iteration round number T is a variable for controlling the upper limit of the training times, the iteration round number control variable T and a regularization factor lambda2The method comprises the steps of measuring the limiting effect of a regularization item on a model, determining the number of nodes of the parallel calculation of the model by the number N of nodes of the parallel calculation of the regularization item, determining the number b of data division sections, determining that a data set is divided into b data sections, wherein each data section is provided with b data blocks which are independent of each other and do not share a row or a column, determining the acceleration effect of momentum on gradient reduction to accelerate the convergence speed of the model, determining the learning step size of model training by the learning step size η of the model training, and judging whether the training is finished or not by the convergence termination threshold value a.
The model building unit is configured to: constructing an old people diet characteristic extraction model with rapid random gradient descent, and constructing a target function RLY of the model according to a known data set Q in an old people-diet scoring matrix RQThe formula is as follows:
Figure FDA0002482886760000021
wherein Q represents a known scoring data set of daily diet for the elderly in the elderly-daily diet scoring matrix R; r isx,yThe meaning is that the x-to-daily diet y preference degree of the old is the x-to-daily diet y preference degree of the oldPreference score for diet y; u. ofx′Representing a hidden eigenvalue corresponding to the xth old man in the old man hidden eigenvalue matrix U; i.e. iy′And (4) representing the implicit characteristic value corresponding to the y-th diet in the daily diet implicit characteristic matrix I.
To prevent overfitting of the model, the generalization performance of the model is enhanced, usually in the objective function RLYQAdding L2The regularization term constrains the model, which can be given by the following equation:
Figure 3
wherein λ2Is a regularization term control parameter, λ2Represents the measure L2The limiting effect of the regularization term on the model.
The data fast training unit is configured to: and combining the target function constructed by the model construction unit, quickly training data and extracting characteristic values. In order to satisfy U and I for accumulating error RLY on setQAt a minimum, the cumulative error RLY is computed using a fast converging stochastic gradient descent optimization algorithmQAnd training the matrixes U and I to obtain the global optimal solution of the matrixes U and I. In the training process, the selection of the hidden characteristics of the old and the daily diet types is adopted to quickly update the direction, the linear combination of the current gradient and the last update gradient of the parameters is realized according to the learning step length and the momentum acceleration factor, and the current update u is calculatedx′And iy′The formula expression of the gradient value corresponding to the fastest convergence speed is as follows:
Figure FDA0002482886760000032
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e., the kth implicit feature of geriatric x and diet y.
Figure FDA0002482886760000033
And
Figure FDA0002482886760000034
respectively, the current gradient value of the t time. The expression is as follows:
Figure FDA0002482886760000041
selecting the calculated gradient value with the fastest convergence according to the fast convergence direction, and obtaining a parameter updating rule of an old people diet characteristic extraction model based on fast random gradient decline:
about parameter ux,kThe update rule of (2) is as follows:
Figure FDA0002482886760000042
with respect to parameter iy,kThe update rule of (2) is as follows:
Figure FDA0002482886760000043
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e., the kth implicit feature of geriatric x and diet y.
Figure FDA0002482886760000044
And
Figure FDA0002482886760000045
respectively, the current gradient value of the t time.
The following new rule can accelerate the convergence speed of the model, in order to further accelerate the calculation efficiency of the model, the data is divided into b data segments before the model is updated, b blocks are arranged in each data segment, the blocks are irrelevant, and meanwhile, b calculation nodes are used for respectively and simultaneously training b independent blocks, so that the high-efficiency calculation speed is obtained. And executing a parameter updating rule in each block until all data block training is finished, which represents that the iteration is finished.
Repeating the training process on the known data set until the model is converged, wherein the convergence judgment condition is that the iteration number T reaches the maximum training iteration number T or RLY obtained by calculation of the current iteration and the last iteration isQThe absolute value of the difference in values is less than the convergence termination threshold a.
4. A method for extracting dietary features of old people with fast random gradient descent according to any one of claims 1 to 3, comprising the following steps:
s1: receiving old people diet scoring data of an information acquisition terminal, and establishing an old people-diet preference matrix;
s2: storing known data in a form of a triple, namely P ═ x, y, r, wherein l represents geriatric, y represents diet, and r represents geriatric x scores diet y;
s3: initializing relevant parameters in the process of extracting the diet characteristics of the old;
s4: constructing a target loss function by combining known acquired data and initialized related parameters;
s5: updating hidden characteristics of old people and diet according to the constructed target loss function, and obtaining two characteristic matrixes when the model reaches convergence;
s6: two feature matrices are output.
5. The method for elder diet feature extraction with fast stochastic gradient descent according to claim 4, wherein the step S3 comprises:
initializing hidden feature matrixes U and I; initializing a hidden feature space dimension d; initializing a regularization factor λ2The method comprises the steps of initializing the number N of nodes in parallel calculation, initializing the number b of data division sections, initializing the maximum iteration round number T, initializing an iteration round number control variable T in the training process, initializing a momentum acceleration factor gamma, initializing a model training learning step η and initializing a convergence termination threshold a.
6. The method for elder diet feature extraction with fast stochastic gradient descent according to claim 5, wherein the step S4 comprises:
constructing an objective function RLY of the model according to a known data set Q in the old people-diet scoring matrix RQThe formula is as follows:
Figure FDA0002482886760000061
wherein Q represents a known scoring data set of daily diet for the elderly in the elderly-daily diet scoring matrix R; r isx,yThe meaning is that the preference degree of the old people x to the daily diet y is the preference score of the old people x to the daily diet y; u. ofx′Representing a hidden eigenvalue corresponding to the xth old man in the old man hidden eigenvalue matrix U; i.e. iy′And (4) representing the implicit characteristic value corresponding to the y-th diet in the daily diet implicit characteristic matrix I.
To prevent overfitting of the model, the generalization performance of the model is enhanced, usually in the objective function RLYQAdding L2The regularization term constrains the model, which can be given by the following equation:
Figure 2
wherein λ2Is a regularization term control parameter, λ2Represents the measure L2The limiting effect of the regularization term on the model.
7. The method for elder diet feature extraction with fast stochastic gradient descent according to claim 6, wherein the step S5 comprises:
and S51, dividing the data into b data segments, wherein each data segment contains b data blocks and the b data blocks are not related to each other. And selecting one data segment every time, distributing the b data blocks to different computing nodes until all data are distributed and trained, and finishing one iteration.
And S52, different nodes acquire data blocks which are not related to each other, and simultaneously train and optimize the target loss function. In order to satisfy U and I for accumulating error RLY on setQAt a minimum, the cumulative error RLY is computed using a fast converging stochastic gradient descent optimization algorithmQAnd training the matrixes U and I to obtain the global optimal solution of the matrixes U and I. In the training process, the selection of the hidden characteristics of the old and the daily diet types is adopted to quickly update the direction, the linear combination of the current gradient and the last update gradient of the parameters is realized according to the learning step length and the momentum acceleration factor, and the current update u is calculatedx′And iy′The formula expression of the gradient value corresponding to the fastest convergence speed is as follows:
Figure FDA0002482886760000071
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e., the kth implicit feature of geriatric x and diet y.
Figure FDA0002482886760000072
And
Figure FDA0002482886760000073
respectively, the current gradient value of the t time. The expression is as follows:
Figure FDA0002482886760000074
selecting the calculated gradient value with the fastest convergence according to the fast convergence direction, and obtaining a parameter updating rule of an old people diet characteristic extraction model based on fast random gradient decline:
about parameter ux,kThe update rule of (2) is as follows:
Figure FDA0002482886760000075
with respect to parameter iy,kThe update rule of (2) is as follows:
Figure FDA0002482886760000081
where gamma is the momentum acceleration factor, η is the parameter training step length, parameter w(t-1)And k(t-1)Are respectively a parameter ux,kAnd iy,kThe update gradient size at the t-th update is respectively w(t-1)And k(t-1)And the corresponding decision parameter consists of two parts of an updated gradient and a current gradient accumulated value at t-1 times. Wherein u isx,kAnd iy,kAre each ux′And iy′The kth element of the vector, i.e., the kth implicit feature of geriatric x and diet y.
Figure FDA0002482886760000082
And
Figure FDA0002482886760000083
respectively, the current gradient value of the t time.
And S53, judging whether the training of all the data blocks is finished, if the data is not completely processed in the current iteration, returning to S51, and acquiring the unassigned data blocks to continue the training. If the training is completed, the process proceeds to S54.
S54, judging whether the iteration end condition is satisfied, wherein the convergence judgment condition is that the iteration round number T reaches the maximum training iteration round number T or RLY obtained by the calculation of the current iteration and the last iterationQThe absolute value of the difference in values is less than the convergence termination threshold a.
CN202010383353.XA 2020-05-08 2020-05-08 Old people diet feature extraction device and method based on rapid random gradient descent Pending CN111564200A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010383353.XA CN111564200A (en) 2020-05-08 2020-05-08 Old people diet feature extraction device and method based on rapid random gradient descent

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010383353.XA CN111564200A (en) 2020-05-08 2020-05-08 Old people diet feature extraction device and method based on rapid random gradient descent

Publications (1)

Publication Number Publication Date
CN111564200A true CN111564200A (en) 2020-08-21

Family

ID=72071833

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010383353.XA Pending CN111564200A (en) 2020-05-08 2020-05-08 Old people diet feature extraction device and method based on rapid random gradient descent

Country Status (1)

Country Link
CN (1) CN111564200A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117133407A (en) * 2023-10-26 2023-11-28 北京四海汇智科技有限公司 Nutritional balance assessment method and system for multi-label neural network for children

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107563114A (en) * 2017-08-15 2018-01-09 国家粮食局科学研究院 A kind of matching method for the specified food materials for realizing specific nutrition target
CN108597582A (en) * 2018-04-18 2018-09-28 中国科学院计算技术研究所 A kind of method and apparatus for executing Faster R-CNN neural network computings
CN110390561A (en) * 2019-07-04 2019-10-29 四川金赞科技有限公司 User-financial product of stochastic gradient descent is accelerated to select tendency ultra rapid predictions method and apparatus based on momentum

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107563114A (en) * 2017-08-15 2018-01-09 国家粮食局科学研究院 A kind of matching method for the specified food materials for realizing specific nutrition target
CN108597582A (en) * 2018-04-18 2018-09-28 中国科学院计算技术研究所 A kind of method and apparatus for executing Faster R-CNN neural network computings
CN110390561A (en) * 2019-07-04 2019-10-29 四川金赞科技有限公司 User-financial product of stochastic gradient descent is accelerated to select tendency ultra rapid predictions method and apparatus based on momentum

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117133407A (en) * 2023-10-26 2023-11-28 北京四海汇智科技有限公司 Nutritional balance assessment method and system for multi-label neural network for children
CN117133407B (en) * 2023-10-26 2024-02-13 北京四海汇智科技有限公司 Nutritional balance assessment method and system for multi-label neural network for children

Similar Documents

Publication Publication Date Title
JP6962263B2 (en) 3D point cloud label learning device, 3D point cloud label estimation device, 3D point cloud label learning method, 3D point cloud label estimation method, and program
CN111144329B (en) Multi-label-based lightweight rapid crowd counting method
CN108304440B (en) Game pushing method and device, computer equipment and storage medium
CN107145977A (en) A kind of method that structured attributes deduction is carried out to online social network user
CN106503106A (en) A kind of image hash index construction method based on deep learning
CN113221694B (en) Action recognition method
CN110390561B (en) User-financial product selection tendency high-speed prediction method and device based on momentum acceleration random gradient decline
CN106614273B (en) Pet feeding method and system based on Internet of Things big data analysis
CN112001887B (en) Full convolution genetic neural network method for infant brain medical record image segmentation
CN113688765A (en) Attention mechanism-based action recognition method for adaptive graph convolution network
CN110322509B (en) Target positioning method, system and computer equipment based on hierarchical class activation graph
CN107092812A (en) A kind of method based on genetic algorithm in identification key protein matter in PPI networks
CN108596264A (en) A kind of community discovery method based on deep learning
CN110659420A (en) Personalized catering method based on deep neural network Monte Carlo search tree
CN114998525A (en) Action identification method based on dynamic local-global graph convolutional neural network
Tang et al. A phylogenetic scan test on a Dirichlet-tree multinomial model for microbiome data
CN114912673A (en) Water level prediction method based on whale optimization algorithm and long-term and short-term memory network
CN114490618B (en) Ant-lion algorithm-based data filling method, device, equipment and storage medium
CN117999562A (en) Method and system for quantifying client contribution in federal learning
CN115018039A (en) Neural network distillation method, target detection method and device
CN111564200A (en) Old people diet feature extraction device and method based on rapid random gradient descent
CN112164474A (en) Drug sensitivity prediction method based on self-expression model
CN108304844A (en) Agricultural pest recognition methods based on deep learning binaryzation convolutional neural networks
Shobana et al. Plant disease detection using convolution neural network
CN113407820A (en) Model training method, related system and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20200821

WD01 Invention patent application deemed withdrawn after publication