CN110633433B - Page caching method and device, electronic equipment and storage medium - Google Patents

Page caching method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN110633433B
CN110633433B CN201910656376.0A CN201910656376A CN110633433B CN 110633433 B CN110633433 B CN 110633433B CN 201910656376 A CN201910656376 A CN 201910656376A CN 110633433 B CN110633433 B CN 110633433B
Authority
CN
China
Prior art keywords
page
target
current
caching
target page
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910656376.0A
Other languages
Chinese (zh)
Other versions
CN110633433A (en
Inventor
请求不公布姓名
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Youzhuju Network Technology Co Ltd
Original Assignee
Beijing Youzhuju Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Youzhuju Network Technology Co Ltd filed Critical Beijing Youzhuju Network Technology Co Ltd
Priority to CN201910656376.0A priority Critical patent/CN110633433B/en
Publication of CN110633433A publication Critical patent/CN110633433A/en
Application granted granted Critical
Publication of CN110633433B publication Critical patent/CN110633433B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0893Caches characterised by their organisation or structure
    • G06F12/0895Caches characterised by their organisation or structure of parts of caches, e.g. directory or tag array
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/957Browsing optimisation, e.g. caching or content distillation
    • G06F16/9574Browsing optimisation, e.g. caching or content distillation of access to content, e.g. by caching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2212/00Indexing scheme relating to accessing, addressing or allocation within memory systems or architectures
    • G06F2212/10Providing a specific technical effect
    • G06F2212/1041Resource optimization
    • G06F2212/1044Space efficiency improvement
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The embodiment of the disclosure provides a page caching method, a page caching device, electronic equipment and a storage medium, wherein the method comprises the following steps: when the operation of triggering the current page to enter the next page is detected, determining the current storage information in the first cache region, and when the current storage information meets a first preset condition, respectively caching the target page data in the target page in the first cache region and the target page frame. According to the technical scheme, the problems that in the prior art, if more pages are opened step by step, the pages are required to be stored in the system memory to ensure that the function of returning to the previous level of page can be executed, but the pages occupy the larger memory of the system, the conditions of insufficient system memory, stuck operation and the like exist, the user experience is affected, the fact that partial page data are stored in advance is achieved, the storage strategy is optimized, the memory capacity of the system memory is reduced, and the technical effect of the user experience is improved are solved.

Description

Page caching method and device, electronic equipment and storage medium
Technical Field
The embodiment of the disclosure relates to the technical field of computers, in particular to a page caching method, a page caching device, electronic equipment and a storage medium.
Background
With the development of electronic products, users can read various articles or watch videos through application programs. For example, after the user triggers the application program, a certain page is opened for browsing, if the user triggers a certain control on the current page, the user can enter the next page of the current page for browsing, that is to say, a new page can be entered infinitely through links or controls in the page. A return control is also typically provided in the page, and after clicking, the previous page of the current page can be returned.
Therefore, if there are more pages opened step by step in the interface, these pages need to be stored in the memory to ensure that the function of returning to the previous page can be executed. Multiple pages are stored in the memory, so that the technical problem that the user experience is affected due to the fact that the memory occupation is large and the system memory is insufficient and is blocked is solved.
Disclosure of Invention
The embodiment of the disclosure provides a page caching method, a page caching device, electronic equipment and a page caching storage medium, so that a storage strategy is optimized, the memory capacity of a system memory is reduced, and the technical effect of user experience is improved.
In a first aspect, an embodiment of the present disclosure further provides a page caching method, where the method includes:
when detecting an operation triggering the entry from the current page to the next page, determining current storage information in a first cache area;
when the current storage information meets a first preset condition, respectively caching target page data in a target page in the first cache area and a target page frame;
the first cache area is a first memory of the system.
In a second aspect, an embodiment of the present disclosure further provides a page caching apparatus, where the apparatus includes:
the current storage information detection module is used for determining the current storage information in the first cache area when detecting the operation of triggering the current page to enter the next page;
the storage module is used for respectively caching the target page data in the target page in the first cache area and the target page frame when the current storage information meets a first preset condition;
the first cache area is a first memory of the system.
In a third aspect, embodiments of the present disclosure further provide an electronic device, including:
One or more processors;
storage means for storing one or more programs,
the one or more programs, when executed by the one or more processors, cause the one or more processors to implement the page caching method as described in any one of the embodiments of the present disclosure.
In a fourth aspect, the disclosed embodiments also provide a storage medium containing computer-executable instructions, which when executed by a computer processor, are for performing a page caching method as described in any of the disclosed embodiments.
According to the technical scheme, when the operation of triggering the next page from the current page is detected, the current storage information in the first cache area is determined, and when the current storage information meets the first preset condition, target page data in target pages in the first cache area are stored, so that the technical effects that if more pages are opened step by step in the prior art, the pages need to be stored in a system memory to ensure that the function of returning to the previous page can be executed, but the pages need to occupy a larger memory of the system, the technical problems that the system memory is insufficient, the situation of blocking occurs during operation and the like and the user experience is affected are solved, when the memory storage amount of the system memory meets the preset condition, the storage strategy is optimized, the memory storage amount of the system memory is reduced, and the technical effect of the user experience is improved.
Drawings
The above and other features, advantages, and aspects of embodiments of the present disclosure will become more apparent by reference to the following detailed description when taken in conjunction with the accompanying drawings. The same or similar reference numbers will be used throughout the drawings to refer to the same or like elements. It should be understood that the figures are schematic and that elements and components are not necessarily drawn to scale.
Fig. 1 is a schematic flow chart of a page buffering method according to a first embodiment of the disclosure;
fig. 2 is a schematic flow chart of a page buffering method according to a second embodiment of the disclosure;
fig. 3 is a schematic flow chart of a page rollback method according to a third embodiment of the disclosure;
FIG. 4 is a schematic flow chart of a fourth preferred embodiment according to the present disclosure;
fig. 5 is a schematic structural diagram of a page buffer device according to a fifth embodiment of the present disclosure;
fig. 6 is a schematic structural diagram of an electronic device according to a sixth embodiment of the disclosure.
Detailed Description
Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While certain embodiments of the present disclosure have been shown in the accompanying drawings, it is to be understood that the present disclosure may be embodied in various forms and should not be construed as limited to the embodiments set forth herein, but are provided to provide a more thorough and complete understanding of the present disclosure. It should be understood that the drawings and embodiments of the present disclosure are for illustration purposes only and are not intended to limit the scope of the present disclosure.
It should be understood that the various steps recited in the method embodiments of the present disclosure may be performed in a different order and/or performed in parallel. Furthermore, method embodiments may include additional steps and/or omit performing the illustrated steps. The scope of the present disclosure is not limited in this respect.
The term "including" and variations thereof as used herein are intended to be open-ended, i.e., including, but not limited to. The term "based on" is based at least in part on. The term "one embodiment" means "at least one embodiment"; the term "another embodiment" means "at least one additional embodiment"; the term "some embodiments" means "at least some embodiments. Related definitions of other terms will be given in the description below.
It should be noted that the terms "first," "second," and the like in this disclosure are merely used to distinguish between different devices, modules, or units and are not used to define an order or interdependence of functions performed by the devices, modules, or units.
It should be noted that references to "one", "a plurality" and "a plurality" in this disclosure are intended to be illustrative rather than limiting, and those of ordinary skill in the art will appreciate that "one or more" is intended to be understood as "one or more" unless the context clearly indicates otherwise.
Example 1
Fig. 1 is a schematic flow chart of a page caching method provided in an embodiment of the present disclosure, where the method of the present embodiment may be suitable for a case of caching each page stored in a system memory when the system memory storage amount meets a preset condition, and the method may be performed by a page caching device, where the device may be implemented in a software and/or hardware form, and optionally, the electronic device may be a mobile terminal, a PC end, or the like.
As shown in fig. 1, the method of this embodiment includes:
s110, when the operation of triggering the current page to enter the next page is detected, determining the current storage information in the first cache area.
The page currently browsed by the user is used as the current page. At least two controls may be included on the current page for the user to trigger to go to the next page. When a user triggers one of the controls, the user can jump from the current page to the next page for browsing, namely the page jump step by step is realized. That is, by triggering a link or control in the current page, the next page to the current page can be entered. The first cache area can be understood as a system memory of an application program or a browser, and is used as a first memory area of the system for storing each page opened by a user step by step. The current storage information can be understood as the total memory occupied by each page stored in the first memory of the system and/or the number of pages stored in the first cache area.
Specifically, when detecting that a user triggers a link or a control on a current page and jumps to a next page, the number of pages stored in the first cache area and/or the memory occupied by each page stored in the first cache area may be obtained.
It should be noted that, when the user triggers a jump from the current page to the next page, the current page may be jumped to the first buffer area.
And S120, when the current storage information meets a first preset condition, respectively caching the target page data in the target page in the first cache region and the target page frame.
And taking a preset threshold value met by the storage amount in the current stored information as a first preset condition. The first preset condition may be: the number of the pages stored in the current storage information is larger than the first preset number of the pages, and/or the memory occupied by each page stored in the current storage information exceeds the preset storage memory. That is, the currently stored information may acquire the target page from the first cache area as long as one of the above conditions is satisfied. The target page is understood to be a page to be retrieved from the first cache area and stored to the target location. The page includes page data, and a page frame. And taking the page data of the target page as target page data, and taking the page frame in the target page as target page frame.
In this embodiment, the target page data in the target page and the target page frame are respectively cached, which may be: storing target page data in the target page into a local disk, and storing a target page frame in the target page into a cache pool, wherein the cache pool is a second memory of the system.
The cache pool is a second memory of the application program. The first memory and the second memory are both part of the system memory.
Specifically, when the number of pages stored in the first cache area is greater than the first preset number of pages, and/or the total memory occupied by storing each page exceeds the preset memory, the target page to be stored to the target position may be determined from the first cache area. After the target page is determined, page data on the target page can be respectively cached to a local disk, and a page frame of the target page is cached to a cache pool.
The first preset condition is that the threshold number of the pages stored in the first cache area is 10, and/or the memory occupied by storing the number of each page is 512M. When a user triggers a control on a current page to enter a next page, the current page can be stored in a first cache area, the number of pages stored in the first cache area is 11, and the number of pages exceeds a threshold value of 10, then a target page can be obtained from the 11 pages, page data of the target page is stored in a local disk, and a target page frame is stored in a cache pool.
It should be noted that, if the current stored information does not satisfy the first preset condition, the above operation may not be performed.
In this embodiment, determining the target page from the first cache area may be: when the current stored information meets a first preset condition, acquiring the cache moment of caching each page in a first cache area; according to the caching moment, determining a time difference value from the current moment, taking a page corresponding to the time difference value from the current moment as a target page, caching target page data in the target page into a local disk, and caching a target page frame into a caching pool; the current time is the time when the current stored information is detected to meet the first preset condition.
The time when the page is stored in the first cache area is taken as the cache time of the page. That is, the pages stored in the first cache area are different, and the corresponding cache times are also different. The current time is: and after detecting that the user triggers to enter the next page from the current page and store the current page in the first cache area, acquiring the moment when the stored information in the first cache area meets the preset condition.
It may be understood that, if the current stored information meets the first preset condition, the buffering time of storing each page in the first buffering area may be obtained. And calculating the time difference between the cache time of each page and the current time. When the time difference is larger, the earlier the time of page buffer storage to the first buffer storage area is, namely the page browsed earlier by the user is indicated. The page corresponding to the largest time difference from the current moment can be used as the target page, namely, the page which is browsed earliest by the user in the first cache area is used as the target page. And caching page data on the target page into a local disk, and caching a page frame on the target page into a cache pool.
For example, the time when the current stored information meets the first preset condition is a, and the cache time when each page is stored in the first cache area is A1, A2, A3, A4, … a10 respectively. Where A1 represents the time when the page was first stored in the first cache region, and similarly a10 represents the time when it was last stored in the first cache region. Then, the page corresponding to the cache time A1 is the target page, and the data and the page frame in the target page are cached respectively.
Of course, each page stored in the first cache area may be numbered, the first page stored in the first cache area is numbered 1, the second page stored in the cache area is numbered 2, and so on, each page stored in the first cache area is numbered, and each page is stored in the order of the numbers. When the page number stored in the first cache area is detected to be 11 and exceeds the page number threshold value 10, the page with the number of 1 can be used as a target page, and page data and a page frame of the target page are respectively cached; and if the page number stored in the first cache area again is 12, caching the page with the number of 2 in the first cache area as a target page to a target position, and determining the target page from the first cache area by analogy. That is, in the process of determining the target page and storing the target page to the target position, the number of the stored pages in the first cache area is always less than or equal to the preset page number threshold.
According to the technical scheme, when the operation of triggering the next page from the current page is detected, the current storage information in the first cache area is determined, and when the current storage information meets the first preset condition, target page data in target pages in the first cache area are stored, so that the technical effects that if more pages are opened step by step in the prior art, the pages need to be stored in a system memory to ensure that the function of returning to the previous page can be executed, but the pages need to occupy a larger memory of the system, the technical problems that the system memory is insufficient, the situation of blocking occurs during operation and the like and the user experience is affected are solved, the storage strategy is optimized when the memory storage amount of the system memory meets the preset condition, the memory storage amount of the system memory is reduced, and the technical effect of user experience is improved.
Example two
On the basis of the first embodiment, after caching the page data in the target page to the local disk and caching the target page frame to the cache pool, if the operation performed when the current page is triggered to enter the next page can be referred to as fig. 2. Fig. 2 is a flowchart of a page buffering method according to a second embodiment of the present disclosure.
As shown in fig. 2, the method includes:
s210, when the operation of triggering the current page to enter the next page is detected, determining the current storage information in the first cache area.
For example, if a user starts an application program, when browsing a page in the application program, a control in the current page may be triggered, so as to enter a page next to the current page. When the triggering operation of the user is detected, the current page can be stored in the first cache area, and the current storage information in the first cache area, namely the current storage information in the first memory of the system of the application program, is acquired, and optionally, 6 pages are stored in the first memory of the system. Wherein the page includes page data and a page frame.
When the user triggers the operation of entering the next page from the current page, the current page may be preferentially stored in the first cache area, and the current storage information of the first cache area is acquired. For example, when a user triggers a control on a first page to enter a second page, the first page can be cached in a first cache area, and the current storage information in the first cache area is obtained at this time, and the number of the stored pages is 1; when the user triggers the control on the second page and jumps to the third page, the second page can be cached in the first cache area, and at the moment, the current storage information in the first cache area is obtained, and the number of the stored pages is 2. After each page is stored in the above manner, the current storage information in the first cache area is obtained.
And S220, when the current storage information meets a first preset condition, respectively caching the target page data in the target page in the first cache region and the target page frame.
The first preset condition is, for example: in the first cache area, the page number storage threshold is 5 pages, namely the page number stored in the first cache area needs to be less than or equal to five pages. Because the number of the pages stored in the current storage information is 6 pages, and the page corresponding to the earliest storage time in the first cache area is taken as the target page when the number exceeds the preset page number threshold. For example, the time for storing each page in the first cache area is sequentially: twenty-five seconds at eight points, twenty-thirty seconds at eight points, twenty-five seconds at eight points, twenty-two forty seconds at eight points, twenty-three thirty seconds at eight points, twenty-five forty seconds at eight points. From this, the page stored in the first cache area at the earliest time is stored for twenty-twenty minutes and fifteen seconds, and the page corresponding to this time can be used as the target page.
And caching the target page data in the target page into a local disk, and caching the target page frame into a cache pool, namely a second memory of the system.
In this embodiment, the target page may be determined according to a buffering time of buffering each page in the first buffering area; when the first cache area meets the first preset condition, the page with the smallest number can be used as the target page.
And S230, loading the data to be processed corresponding to the next page when the touch operation triggering the next page is detected, and loading the target page frame from the cache pool.
The image to be processed refers to page data corresponding to the next page. The target page frame refers to the page frame cached to the cache pool determined in S210 to S220.
On the basis of S210 to S220, when it is detected that the control on the current page is triggered to enter the next page, page data corresponding to the next page may be loaded, and the page frame may be loaded from the cache pool. The advantage of this arrangement is that the page frame can be obtained from the system memory quickly without reloading the page frame, thereby achieving the technical effect of saving resources.
It should be noted that this step is applicable to the case where the page frame is stored in the cache pool. If the cache pool does not have the page frame, when touch operation triggering the current page to enter the next page is detected, the data to be processed and the page frame corresponding to the next page can be loaded, and rendering is carried out on the data to be processed and the page frame to obtain the next page.
And S240, rendering the data to be processed and the target page frame to obtain the next page.
Specifically, the next page can be obtained after rendering the acquired page data corresponding to the next page and the target frame acquired from the cache pool.
According to the technical scheme, when the operation of triggering the next page from the current page is detected, the current storage information in the first cache area is determined, and when the current storage information meets the first preset condition, target page data in target pages in the first cache area are stored, so that the technical effects that if more pages are opened step by step in the prior art, the pages need to be stored in a system memory to ensure that the function of returning to the previous page can be executed, but the pages need to occupy a larger memory of the system, the technical problems that the system memory is insufficient, the situation of blocking occurs during operation and the like and the user experience is affected are solved, the storage strategy is optimized when the memory storage amount of the system memory meets the preset condition, the memory storage amount of the system memory is reduced, and the technical effect of user experience is improved.
Example III
On the basis of the technical scheme, when the operation of triggering page rollback by the user is detected, each page data stored in the target position and the page frames stored in the cache pool can be processed according to the information stored in the first cache region. Fig. 3 is a flowchart of a page rollback method according to a third embodiment of the present disclosure.
As shown in fig. 3, the method includes:
s310, when detecting the operation triggering the rollback from the current page to the previous page, determining the current storage information in the first cache region.
The previous page is understood to be the page with the smallest time interval from the current page among all the pages browsed before the current page.
Specifically, when a trigger operation triggering the rollback from the current page to the previous page is detected, the storage information in the first cache area is read, and the storage information read at the moment is used as the current storage information.
Illustratively, the current storage information of the first cache area is obtained and stored for 4 pages.
S320, when the current storage information meets a second preset condition, acquiring target page data to be called back from the local disk.
It should be noted that, when a trigger operation from a current page to a next page is detected, whether the page stored in the first cache area needs to be cached to the target position or not needs to satisfy a first preset condition. Correspondingly, when detecting that the user triggers the rollback operation, whether page data to be called back is acquired from the target position or not needs to meet a second preset condition.
Optionally, the second preset condition may be: the number of pages stored in the current stored information is smaller than the first preset number of pages; and/or the total memory occupied by each page stored in the current storage information is smaller than the preset storage memory.
Specifically, when the number of pages stored in the current storage information is smaller than the first preset number of pages, and/or the memory occupied by each page stored in the current storage information is smaller than the preset storage memory, the target page data to be recalled can be obtained from the target position.
Optionally, determining the target callback data may be: acquiring storage time of storing each page data in a local disk, and acquiring a time difference value corresponding to each page data according to the storage time and the return time; the return time is the time when the user triggers the operation of backing to the previous page from the current page; and taking the page data corresponding to the minimum time difference value as target page data to be called back.
When the page data is stored from the first cache area to the local disk, the corresponding time is taken as the storage time. The return time refers to the time when the user triggers the operation of backing from the current page to the previous page. The target page data to be called back is page data to be called back from the target position to the first cache area.
According to the storage time and the return time, the time difference of each page in the local disk from the return time can be determined. The larger the time difference value is, the earliest the page user corresponding to the storage time browses; correspondingly, the smaller the time difference value is, the page corresponding to the storage time is the most recently browsed page by the user.
It should be noted that, when the page is rolled back, the page that is rolled back first is the page that the user browses recently, for example, the user browses the page a, after triggering the content on the page a to enter the page B and triggering the control on the page B to enter the page C, the user triggers the roll-back operation, and then the page that is rolled back first is the page B, and after triggering the roll-back operation, the user can roll back to the page a. That is, when detecting the operation that the user triggers the page rollback, the page data corresponding to the smallest time difference from the current moment can be used as the target page data to be recalled.
Specifically, the storage time of each page data stored in the local disk is obtained, and the time difference of each page from the return time is determined according to the return time. And taking the page data corresponding to the small time difference value as target page data to be called back.
It should be noted that when the page data corresponding to each page is stored in the local disk, each page data may be numbered, the page data stored first in the local disk is marked as 1, the page data stored second in the local disk is marked as 2, and so on, and the page data stored in the local disk are numbered sequentially. When the page data is called back, the page data with the largest number can be preferentially used as target page data to be called back.
S330, obtaining a target callback page based on the target page data to be callback and the page frame, and callback the target callback page to a preset position in the first cache area.
The preset position may be a first position of each stored page in the first cache area.
Specifically, rendering is performed on the target page data to be called back and the page frame to obtain the target callback page. And storing the target callback page to the first position of each page in the first cache area.
In order to obtain the target callback page, after obtaining the target callback page data, a page frame is required to be obtained. Optionally, detecting whether a page frame is stored in the cache pool; if yes, acquiring a page frame, and performing rendering processing with the target page data to be called back to obtain a target page to be called back; if not, a page frame is newly established, rendering processing is carried out on the page frame and the target page data to be called back, and the target page to be called back is obtained.
When the page frames are cached in the cache pool, the page frames can be obtained from the cache pool, and rendering processing is carried out on the page frames and target callback data to obtain target callback pages. If the page frame is not stored in the cache pool, a page frame can be newly established and is processed with the acquired target page data to be called back to obtain the target callback page.
For example, the second preset condition threshold is 5, when it is detected that the number of pages stored in the first cache area is 4 and is smaller than the second preset condition, and the page numbers stored in the first cache area are 6, 5, 4 and 3, respectively, then there should be two page data stored in the local disk, the corresponding numbers may be 2 and 1, then the page data corresponding to the number 2 is regarded as the page data to be called back, and is inserted into the back of the number 3, at this time, the page numbers stored in the first cache area are 6, 5, 4, 3 and 2, respectively, and when page rollback is triggered, the page is rolled back to the page with the number 6, then to the page with the number 5, and so on.
It should be noted that, if the number of pages stored in the first cache area meets the second preset condition and the local disk does not store page data, even if the page rollback operation is detected, the page data in the local disk may not be read any more.
According to the technical scheme, when the operation of triggering the next page from the current page is detected, the current storage information in the first cache area is determined, and when the current storage information meets the first preset condition, target page data in target pages in the first cache area are stored, so that the technical effects that if more pages are opened step by step in the prior art, the pages need to be stored in a system memory to ensure that the function of returning to the previous page can be executed, but the pages need to occupy a larger memory of the system, the technical problems that the system memory is insufficient, the situation of blocking occurs during operation and the like and the user experience is affected are solved, the storage strategy is optimized when the memory storage amount of the system memory meets the preset condition, the memory storage amount of the system memory is reduced, and the technical effect of user experience is improved.
Example IV
As a preferred embodiment of the foregoing embodiments, fig. 4 is a schematic flow chart of a preferred embodiment provided in a fourth embodiment of the disclosure.
As shown in fig. 4, the method includes:
s410, when the control on the current page is triggered to enter the next page, the number of the current storage pages in the first cache area is obtained.
For example, when a user triggers a control on a current page and enters a next page, the current page may be stored in the first cache area, and the number of the currently stored pages in the first cache area is 6. That is, when the user browses the current page, the number of pages stored in the first buffer area is 5, and when the user goes from the current page to the next page, the number of currently stored pages in the first buffer area is 6 after the current page is stored in the first buffer area.
S420, judging whether the number of the current storage pages is larger than a preset page number threshold value, if so, executing S430; if not, return to execution S410.
The threshold value of the number of the preset pages is 5.
That is, when the number of the current memory pages is greater than the preset page number threshold, the page in the first buffer area may be processed, that is, S430 may be executed. If the number of the current memory pages is smaller than the preset page number threshold, whether the user triggers the operation of entering the next page from the current page is continuously detected, that is, S410 and S420 are repeatedly executed.
S430, acquiring the target page in the first cache area, and caching target page data in the target page and the target page frame to a target position.
In this embodiment, the target page may be determined according to the cache time at which each page is stored in the first cache area, or the number of each page is stored. The above two modes can be respectively taken as examples.
The method includes the steps of obtaining a cache time of each page stored in a first cache area, and taking a page with the largest time difference from the current time as a target page. And caching page data in the target page to a local disk, and caching a target page frame in the target page to a cache pool.
For example, when each page is cached in the first cache area, the page with the number 6 may be numbered according to the sequence of caching each page, that is, the page with the number 1 cached in the first cache area is first, and the page with the number 6 is obtained by pushing. And taking the page with the smallest number, namely the page stored in the first cache area firstly, as a target page. And caching page data in the target page to a local disk, and caching the page frame to a cache pool.
S440, when detecting that the user triggers to fall back from the current page to the previous page, acquiring the number of the current storage pages in the first cache region.
When detecting that the user touches to send a page back to the previous page from the current page, the number of the pages currently stored in the first cache area needs to be acquired.
For example, when detecting the rollback operation, that is, when the current page is the previous page or jumps to the previous page, the number of the pages currently stored in the first cache area is obtained to be 4.
S450, judging whether the number of the current storage pages is smaller than a preset page number threshold value, if so, executing S460; if not, execution returns to S440.
The preset page number threshold is 5 pages, for example. When the number of the current stored pages is 4 and is smaller than the preset page number threshold by 5, page data to be recalled from the local disk can be executed S460. If the number of the current memory pages is not less than the preset page number threshold, it may be continuously detected whether the user triggers the page rollback operation, that is, S440 is executed.
S460, acquiring target callback page data from a local disk, acquiring a page frame from a cache pool, and processing the target callback page data and the page frame to obtain a target callback page.
Specifically, when each page data is stored in the local disk, the page data corresponding to the smallest time difference value from the current moment is used as the target page data to be called back. And obtaining a page frame from the cache pool, and rendering the page data to be called back and the page frame to obtain the target callback page.
And storing the target callback data to the first bits of all the storage pages in the first cache area.
According to the technical scheme, when the operation of triggering the next page from the current page is detected, the current storage information in the first cache area is determined, and when the current storage information meets the first preset condition, target page data in target pages in the first cache area are stored, so that the technical effects that if more pages are opened step by step in the prior art, the pages need to be stored in a system memory to ensure that the function of returning to the previous page can be executed, but the pages need to occupy a larger memory of the system, the technical problems that the system memory is insufficient, the situation of blocking occurs during operation and the like and the user experience is affected are solved, the storage strategy is optimized when the memory storage amount of the system memory meets the preset condition, the memory storage amount of the system memory is reduced, and the technical effect of user experience is improved.
Example five
Fig. 5 is a schematic structural diagram of a page caching apparatus according to a fifth embodiment of the present disclosure, where the apparatus includes: a currently stored information detection module 510, and a storage module 520.
The current storage information detection module 510 is configured to determine, when an operation triggering the entry from the current page to the next page is detected, current storage information in the first cache area; the storage module 520 is configured to cache, when the current storage information meets a first preset condition, target page data in a target page in the first cache area and a target page frame respectively; the first cache area is a first memory of the system.
According to the technical scheme, when the operation of triggering the next page from the current page is detected, the current storage information in the first cache area is determined, and when the current storage information meets the first preset condition, target page data in target pages in the first cache area are stored, so that the technical effects that if more pages are opened step by step in the prior art, the pages need to be stored in a system memory to ensure that the function of returning to the previous page can be executed, but the pages need to occupy a larger memory of the system, the technical problems that the system memory is insufficient, the situation of blocking occurs during operation and the like and the user experience is affected are solved, the storage strategy is optimized when the memory storage amount of the system memory meets the preset condition, the memory storage amount of the system memory is reduced, and the technical effect of user experience is improved.
On the basis of the above technical solutions, the storage module is further configured to:
caching page data in the target page to a local disk, and caching a target page frame to a cache pool; the cache pool is a second memory of the system.
On the basis of the above technical solutions, the storage module further includes: a preset condition judgment unit;
the preset condition judging unit is configured to cache, when it is detected that at least one of the following conditions is met, target page data and a target page frame in a target page in the first cache area respectively: the number of pages stored in the current storage information is larger than the first preset page number; the total memory occupied by each page stored in the current storage information exceeds a preset storage memory; the page comprises page data and a page frame.
On the basis of the technical schemes, the storage module is further used for:
when the current storage information meets the first preset condition, acquiring the cache moment of caching each page in the first cache area; according to the caching moment, determining a time difference value from the current moment, taking a page corresponding to the time difference value from the current moment as a target page, caching target page data in the target page into a local disk, and caching a target page frame into the caching pool; the current time is the time when the current stored information is detected to meet a first preset condition.
On the basis of the above technical solutions, when the current storage information meets a first preset condition, the storage module is further configured to, after buffering target page data in the target page and the target page frame in the first buffer area, respectively:
when touch operation triggering to enter a next page is detected, loading data to be processed corresponding to the next page, and loading the target page frame from the cache pool; and rendering the data to be processed and the target page frame to obtain the next page.
On the basis of the technical schemes, the device further comprises: a rollback detection module; the rollback detection module includes:
a rollback operation detection unit, configured to determine current storage information in the first cache area when an operation triggering rollback from a current page to a previous page is detected;
the callback page obtaining unit is used for obtaining target page data to be callback from the local disk when the current storage information meets a second preset condition;
and the callback page caching unit is used for obtaining a target callback page based on the target page data to be callback and the page frame, and callback the target callback page to a preset position in the first caching area.
Based on the above technical solutions, the callback page obtaining unit further includes: callback page judging unit for:
when the current stored information is detected to meet at least one of the following conditions, acquiring target page data to be called back from the local disk: the number of pages stored in the current storage information is smaller than the first preset number of pages; and the total occupied memory stored in the current storage information is smaller than the preset storage memory.
On the basis of the above technical solutions, the callback page obtaining unit is further configured to:
acquiring storage time of storing each page data in a local disk, and acquiring a time difference value corresponding to each page data according to the storage time and the return time; the return time is the time when the user triggers the operation of backing to the previous page from the current page; and taking the page data corresponding to the minimum time difference value as target page data to be called back.
On the basis of the above technical solutions, the callback page cache unit includes a detection page storage frame subunit, configured to:
detecting whether a page frame is stored in the cache pool; if yes, acquiring the page frame, and performing rendering processing with the target page data to be called back to obtain a target page to be called back; if not, a page frame is newly established, rendering processing is carried out on the page frame and the target page data to be called back, and the target page to be called back is obtained.
The page caching device provided by the embodiment of the disclosure can execute the page caching method provided by any embodiment of the disclosure, and has the corresponding functional modules and beneficial effects of the execution method.
It should be noted that each unit and module included in the above apparatus are only divided according to the functional logic, but not limited to the above division, so long as the corresponding functions can be implemented; in addition, the specific names of the functional units are also only for convenience of distinguishing from each other, and are not used to limit the protection scope of the embodiments of the present disclosure.
Example six
Referring now to fig. 6, a schematic diagram of an electronic device (e.g., a terminal device or server in fig. 6) 600 suitable for use in implementing embodiments of the present disclosure is shown. The terminal devices in the embodiments of the present disclosure may include, but are not limited to, mobile terminals such as mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablet computers), PMPs (portable multimedia players), in-vehicle terminals (e.g., in-vehicle navigation terminals), and the like, and stationary terminals such as digital TVs, desktop computers, and the like. The electronic device shown in fig. 6 is merely an example and should not be construed to limit the functionality and scope of use of the disclosed embodiments.
As shown in fig. 6, the electronic device 600 may include a processing means (e.g., a central processing unit, a graphics processor, etc.) 601, which may perform various appropriate actions and processes according to a program stored in a Read Only Memory (ROM) 602 or a program loaded from a storage means 606 into a Random Access Memory (RAM) 603. In the RAM603, various programs and data required for the operation of the electronic apparatus 600 are also stored. The processing device 601, the ROM602, and the RAM603 are connected to each other through a bus 604. An input/output (I/O) interface 605 is also connected to bus 604.
In general, the following devices may be connected to the I/O interface 605: input devices 606 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, and the like; an output device 607 including, for example, a Liquid Crystal Display (LCD), a speaker, a vibrator, and the like; storage 606 including, for example, magnetic tape, hard disk, etc.; and a communication device 609. The communication means 609 may allow the electronic device 600 to communicate with other devices wirelessly or by wire to exchange data. While fig. 6 shows an electronic device 600 having various means, it is to be understood that not all of the illustrated means are required to be implemented or provided. More or fewer devices may be implemented or provided instead.
In particular, according to embodiments of the present disclosure, the processes described above with reference to flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a non-transitory computer readable medium, the computer program comprising program code for performing the method shown in the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network via communication means 609, or from storage means 606, or from ROM 602. The above-described functions defined in the methods of the embodiments of the present disclosure are performed when the computer program is executed by the processing device 601.
The terminal provided by the embodiment of the present disclosure and the page caching method provided by the foregoing embodiment belong to the same inventive concept, and technical details not described in detail in the embodiment of the present disclosure may refer to the foregoing embodiment, and the embodiment of the present disclosure has the same beneficial effects as the foregoing embodiment.
Example seven
A seventh embodiment of the present disclosure provides a computer storage medium having stored thereon a computer program which, when executed by a processor, implements the page caching method provided by the above embodiment.
It should be noted that the computer readable medium described in the present disclosure may be a computer readable signal medium or a computer readable storage medium, or any combination of the two. The computer readable storage medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the foregoing. More specific examples of the computer-readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this disclosure, a computer-readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In the present disclosure, however, the computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, with the computer-readable program code embodied therein. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, fiber optic cables, RF (radio frequency), and the like, or any suitable combination of the foregoing.
In some implementations, the clients, servers may communicate using any currently known or future developed network protocol, such as HTTP (HyperText Transfer Protocol ), and may be interconnected with any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include a local area network ("LAN"), a wide area network ("WAN"), the internet (e.g., the internet), and peer-to-peer networks (e.g., ad hoc peer-to-peer networks), as well as any currently known or future developed networks.
The computer readable medium may be contained in the electronic device; or may exist alone without being incorporated into the electronic device.
The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to:
when detecting an operation triggering the entry from the current page to the next page, determining current storage information in a first cache area;
when the current storage information meets a first preset condition, respectively caching target page data in a target page in the first cache area and a target page frame;
The first cache area is a first memory of the system.
Computer program code for carrying out operations of the present disclosure may be written in one or more programming languages, including, but not limited to, an object oriented programming language such as Java, smalltalk, C ++ and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computer (for example, through the Internet using an Internet service provider).
The flowcharts and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units involved in the embodiments of the present disclosure may be implemented by means of software, or may be implemented by means of hardware. Wherein the name of the unit does not constitute a limitation of the unit itself in some cases, for example, the rollback operation detection unit may also be described as a "rollback detection unit".
The functions described above herein may be performed, at least in part, by one or more hardware logic components. For example, without limitation, exemplary types of hardware logic components that may be used include: a Field Programmable Gate Array (FPGA), an Application Specific Integrated Circuit (ASIC), an Application Specific Standard Product (ASSP), a system on a chip (SOC), a Complex Programmable Logic Device (CPLD), and the like.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. The machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
According to one or more embodiments of the present disclosure, there is provided a page caching method, the method comprising:
when detecting an operation triggering the entry from the current page to the next page, determining current storage information in a first cache area;
when the current storage information meets a first preset condition, respectively caching target page data in a target page in the first cache area and a target page frame;
the first cache area is a first memory of the system.
According to one or more embodiments of the present disclosure, there is provided a page caching method, further including:
optionally, the caching the target page data and the page frame in the target page in the first cache area includes:
caching page data in the target page to a local disk, and caching a target page frame to a cache pool;
the cache pool is a second memory of the system.
According to one or more embodiments of the present disclosure, there is provided a page caching method, further comprising:
optionally, when the current stored information meets a first preset condition, respectively caching target page data and a target page frame in a target page in the first cache area, including:
When at least one condition is detected to be met, respectively caching target page data and target page frames in the target pages in the first cache region:
the number of pages stored in the current storage information is larger than the first preset page number;
the total memory occupied by each page stored in the current storage information exceeds a preset storage memory;
the page comprises page data and a page frame.
According to one or more embodiments of the present disclosure, there is provided a page caching method, further including:
optionally, when the current stored information meets a first preset condition, respectively caching target page data and a target page frame in a target page in the first cache area, including:
when the current storage information meets the first preset condition, acquiring the cache moment of caching each page in the first cache area;
according to the caching moment, determining a time difference value from the current moment, taking a page corresponding to the time difference value from the current moment as a target page, caching target page data in the target page into a local disk, and caching a target page frame into the caching pool;
The current time is the time when the current stored information is detected to meet a first preset condition.
According to one or more embodiments of the present disclosure, there is provided a page caching method, further comprising:
optionally, after the target page data and the target page frame in the target page in the first cache area are cached respectively when the current stored information meets a first preset condition, the method further includes:
when touch operation triggering to enter a next page is detected, loading data to be processed corresponding to the next page, and loading the target page frame from the cache pool;
and rendering the data to be processed and the target page frame to obtain the next page.
According to one or more embodiments of the present disclosure, there is provided a page caching method, further including:
optionally, when detecting an operation triggering to roll back from a current page to a previous page, determining current stored information in the first cache region;
when the current storage information meets a second preset condition, acquiring target page data to be called back from the local disk;
And obtaining a target callback page based on the target page data to be callback and the page frame, and callback the target callback page to a preset position in the first cache area.
According to one or more embodiments of the present disclosure, there is provided a page caching method [ example seventh ], the method further comprising:
optionally, when the current stored information meets a second preset condition, acquiring target page data to be called back from the local disk includes:
when the current stored information is detected to meet at least one of the following conditions, acquiring target page data to be called back from the local disk:
the number of pages stored in the current storage information is smaller than the first preset number of pages;
and the total memory occupied by each page stored in the current storage information is smaller than the preset storage memory.
According to one or more embodiments of the present disclosure, there is provided a page caching method, further comprising:
optionally, the obtaining target page data to be called back from the local disk includes:
acquiring storage time of storing each page data in a local disk, and acquiring a time difference value corresponding to each page data according to the storage time and the return time; the return time is the time when the user triggers the operation of backing to the previous page from the current page;
And taking the page data corresponding to the minimum time difference value as target page data to be called back.
According to one or more embodiments of the present disclosure, there is provided a page caching method, further including:
optionally, the obtaining the target callback page based on the target callback page data and the page frame includes:
detecting whether a page frame is stored in the cache pool;
if yes, acquiring the page frame, and performing rendering processing with the target page data to be called back to obtain a target page to be called back;
if not, a page frame is newly established, rendering processing is carried out on the page frame and the target page data to be called back, and the target page to be called back is obtained.
According to one or more embodiments of the present disclosure, there is provided a page caching apparatus, including:
the current storage information detection module is used for determining the current storage information in the first cache area when detecting the operation of triggering the current page to enter the next page;
the storage module is used for respectively caching the target page data in the target page in the first cache area and the target page frame when the current storage information meets a first preset condition;
The first cache area is a first memory of the system.
The foregoing description is only of the preferred embodiments of the present disclosure and description of the principles of the technology being employed. It will be appreciated by persons skilled in the art that the scope of the disclosure referred to in this disclosure is not limited to the specific combinations of features described above, but also covers other embodiments which may be formed by any combination of features described above or equivalents thereof without departing from the spirit of the disclosure. Such as those described above, are mutually substituted with the technical features having similar functions disclosed in the present disclosure (but not limited thereto).
Moreover, although operations are depicted in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order. In certain circumstances, multitasking and parallel processing may be advantageous. Likewise, while several specific implementation details are included in the above discussion, these should not be construed as limiting the scope of the present disclosure. Certain features that are described in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination.
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are example forms of implementing the claims.

Claims (9)

1. The page caching method is characterized by comprising the following steps of:
when detecting an operation of triggering the current page to enter the next page, firstly storing the current page into a first cache area, and then determining current storage information in the first cache area;
when the current storage information meets a first preset condition, respectively caching target page data in a target page in the first cache area and a target page frame;
the first cache area is a first memory of the system;
the caching the target page data in the target page in the first cache area and the target page frame respectively includes:
caching target page data in the target page to a local disk, and caching a target page frame to a cache pool;
The cache pool is a second memory of the system;
the target page is a page to be acquired from the first cache area and stored to a target position, and is not a current page;
the step of entering the next page from the current page is to enter the next page of the current page after triggering a link or a control in the current page;
when detecting an operation triggering to fall back from a current page to a previous page, determining current storage information in the first cache region;
when the current storage information meets a second preset condition, acquiring target page data to be called back from the local disk;
based on the target page data to be called back and the page frame, obtaining a target callback page, and calling back the target callback page to a preset position in the first cache area, wherein the preset position is the first position of each stored page in the first cache area;
when the current stored information meets a first preset condition, respectively caching target page data in a target page in the first cache region and a target page frame, wherein the method comprises the following steps:
when the current storage information meets the first preset condition, acquiring the cache moment of caching each page in the first cache area;
According to the caching moment, determining a time difference value from the current moment, taking a page corresponding to the time difference value from the current moment as a target page, caching target page data in the target page into a local disk, and caching a target page frame into a cache pool;
the current time is the time when the current stored information is detected to meet a first preset condition.
2. The method of claim 1, wherein when the current stored information meets a first preset condition, respectively caching target page data in a target page in the first cache area and a target page frame, including:
when at least one condition is detected to be met, respectively caching target page data in the target pages in the first cache area and the target page frame:
the number of pages stored in the current storage information is larger than the first preset page number;
the total memory occupied by each page stored in the current storage information exceeds a preset storage memory;
the page comprises page data and a page frame.
3. The method according to claim 1, wherein after the target page data in the target page in the first cache area and the target page frame are cached respectively when the current stored information meets a first preset condition, further comprising:
When touch operation triggering to enter a next page is detected, loading data to be processed corresponding to the next page, and loading the target page frame from the cache pool;
and rendering the data to be processed and the target page frame to obtain the next page.
4. The method of claim 1, wherein when the current stored information meets a second preset condition, the obtaining target page data to be recalled from the local disk includes:
when the current stored information is detected to meet at least one of the following conditions, acquiring target page data to be called back from the local disk:
the number of pages stored in the current storage information is smaller than the first preset page number;
and the total memory occupied by each page stored in the current storage information is smaller than the preset storage memory.
5. The method of claim 1, wherein the obtaining target page data to be recalled from the local disk comprises:
acquiring storage time of storing each page data in a local disk, and acquiring a time difference value corresponding to each page data according to the storage time and the return time; the return time is the time when the user triggers the operation of backing to the previous page from the current page;
And taking the page data corresponding to the minimum time difference value as target page data to be called back.
6. The method of claim 1, wherein the obtaining the target callback page based on the target callback page data and the page frame comprises:
detecting whether a page frame is stored in the cache pool;
if yes, acquiring the page frame, and performing rendering processing with the target page data to be called back to obtain a target callback page;
if not, a page frame is newly established, rendering processing is carried out on the page frame and the target page data to be called back, and a target call back page is obtained.
7. A page caching apparatus, comprising:
the current storage information detection module is used for firstly storing the current page into a first cache area when detecting the operation of triggering the current page to enter the next page, and then determining the current storage information in the first cache area;
the storage module is used for respectively caching the target page data in the target page in the first cache area and the target page frame when the current storage information meets a first preset condition;
the first cache area is a first memory of the system;
The storage module is further configured to:
caching page data in the target page to a local disk, and caching a target page frame to a cache pool;
the cache pool is a second memory of the system;
the target page is a page to be acquired from the first cache area and stored to a target position, and is not a current page;
the step of entering the next page from the current page is to enter the next page of the current page after triggering a link or a control in the current page;
a rollback operation detection unit, configured to determine current storage information in the first cache area when an operation triggering rollback from a current page to a previous page is detected;
the callback page obtaining unit is used for obtaining target page data to be callback from the local disk when the current storage information meets a second preset condition;
the callback page caching unit is used for obtaining a target callback page based on the target page data to be callback and a page frame, and callback the target callback page to a preset position in the first caching area, wherein the preset position is the first position of each stored page in the first caching area;
The storage module is further configured to:
when the current storage information meets the first preset condition, acquiring the cache moment of caching each page in the first cache area;
according to the caching moment, determining a time difference value from the current moment, taking a page corresponding to the time difference value from the current moment as a target page, caching target page data in the target page into a local disk, and caching a target page frame into the caching pool;
the current time is the time when the current stored information is detected to meet a first preset condition.
8. An electronic device, the electronic device comprising:
one or more processors;
storage means for storing one or more programs,
the one or more programs, when executed by the one or more processors, cause the one or more processors to implement the page caching method of any one of claims 1-6.
9. A storage medium containing computer executable instructions which, when executed by a computer processor, are for performing the page caching method of any one of claims 1-6.
CN201910656376.0A 2019-07-19 2019-07-19 Page caching method and device, electronic equipment and storage medium Active CN110633433B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910656376.0A CN110633433B (en) 2019-07-19 2019-07-19 Page caching method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910656376.0A CN110633433B (en) 2019-07-19 2019-07-19 Page caching method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN110633433A CN110633433A (en) 2019-12-31
CN110633433B true CN110633433B (en) 2024-02-27

Family

ID=68968916

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910656376.0A Active CN110633433B (en) 2019-07-19 2019-07-19 Page caching method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN110633433B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113378085B (en) * 2021-05-31 2023-11-14 游艺星际(北京)科技有限公司 Page loading method, device, terminal and storage medium
CN113742625A (en) * 2021-09-07 2021-12-03 北京百度网讯科技有限公司 Page data processing method, device, equipment and medium
CN115858987B (en) * 2023-02-21 2023-04-25 成都萌想科技有限责任公司 Method and device for rendering H5 question bank webpage, electronic equipment and storage medium

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101729590A (en) * 2008-10-15 2010-06-09 北大方正集团有限公司 Method, system and device for providing web page
CN102368258A (en) * 2011-09-30 2012-03-07 广州市动景计算机科技有限公司 Webpage page caching management method and system
CN103336812A (en) * 2013-06-27 2013-10-02 优视科技有限公司 Webpage resource caching method and device for improving secondary loading efficiency
CN104050253A (en) * 2014-06-12 2014-09-17 北京金山网络科技有限公司 Webpage display method and browser
CN104199684A (en) * 2014-08-13 2014-12-10 百度在线网络技术(北京)有限公司 Browser cold-booting method and device
CN104462455A (en) * 2014-12-16 2015-03-25 北京京东尚科信息技术有限公司 Method and device for displaying and processing network data
CN104573025A (en) * 2015-01-12 2015-04-29 北京京东尚科信息技术有限公司 Method and system for increasing page loading rate
CN107943825A (en) * 2017-10-19 2018-04-20 阿里巴巴集团控股有限公司 Data processing method, device and the electronic equipment of page access
CN108153588A (en) * 2016-12-06 2018-06-12 阿里巴巴集团控股有限公司 A kind of page navigation method and device, a kind of memory allocation method and device
CN109558251A (en) * 2017-09-26 2019-04-02 北京京东尚科信息技术有限公司 The method and terminal of page structure information modification
CN109740085A (en) * 2019-01-10 2019-05-10 北京字节跳动网络技术有限公司 A kind of methods of exhibiting of content of pages, device, equipment and storage medium

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6915384B2 (en) * 2002-03-22 2005-07-05 Microsoft Corporation Multiple-level persisted template caching
GB2546239A (en) * 2015-11-23 2017-07-19 Acadiant Ltd A Method and system for providing and executing web applications

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101729590A (en) * 2008-10-15 2010-06-09 北大方正集团有限公司 Method, system and device for providing web page
CN102368258A (en) * 2011-09-30 2012-03-07 广州市动景计算机科技有限公司 Webpage page caching management method and system
CN103336812A (en) * 2013-06-27 2013-10-02 优视科技有限公司 Webpage resource caching method and device for improving secondary loading efficiency
CN104050253A (en) * 2014-06-12 2014-09-17 北京金山网络科技有限公司 Webpage display method and browser
CN104199684A (en) * 2014-08-13 2014-12-10 百度在线网络技术(北京)有限公司 Browser cold-booting method and device
CN104462455A (en) * 2014-12-16 2015-03-25 北京京东尚科信息技术有限公司 Method and device for displaying and processing network data
CN104573025A (en) * 2015-01-12 2015-04-29 北京京东尚科信息技术有限公司 Method and system for increasing page loading rate
CN108153588A (en) * 2016-12-06 2018-06-12 阿里巴巴集团控股有限公司 A kind of page navigation method and device, a kind of memory allocation method and device
CN109558251A (en) * 2017-09-26 2019-04-02 北京京东尚科信息技术有限公司 The method and terminal of page structure information modification
CN107943825A (en) * 2017-10-19 2018-04-20 阿里巴巴集团控股有限公司 Data processing method, device and the electronic equipment of page access
CN109740085A (en) * 2019-01-10 2019-05-10 北京字节跳动网络技术有限公司 A kind of methods of exhibiting of content of pages, device, equipment and storage medium

Also Published As

Publication number Publication date
CN110633433A (en) 2019-12-31

Similar Documents

Publication Publication Date Title
CN110633434B (en) Page caching method and device, electronic equipment and storage medium
CN110633433B (en) Page caching method and device, electronic equipment and storage medium
CN113395572B (en) Video processing method and device, storage medium and electronic equipment
CN111488185B (en) Page data processing method, device, electronic equipment and readable medium
CN112650790B (en) Target point cloud plane determining method and device, electronic equipment and storage medium
CN110647702B (en) Picture preloading method and device, electronic equipment and readable medium
CN111258736B (en) Information processing method and device and electronic equipment
CN109614195B (en) Method, device and equipment for preloading list items and storage medium
CN112379982B (en) Task processing method, device, electronic equipment and computer readable storage medium
CN113934958A (en) Page loading method and device, electronic equipment and computer readable medium
CN110765386A (en) Webpage image monitoring method and device, electronic equipment and readable storage medium
CN111723309B (en) Method and device for web page searching
CN111353296B (en) Article processing method, apparatus, electronic device and computer readable storage medium
CN110717126A (en) Page browsing method and device, electronic equipment and computer readable storage medium
WO2022095932A1 (en) Livestreaming room page jumping method and apparatus, livestreaming room page returning method and apparatus, and electronic device
CN115357361A (en) Task processing method, device, equipment and medium
CN112867119B (en) Terminal control method, device, terminal and storage medium
CN110377362B (en) Method, device, terminal and storage medium for cleaning application program
CN111240758A (en) Material display method and device, electronic equipment and storage medium
CN114915849B (en) Video preloading method, device, equipment and medium
CN111209042A (en) Method, device, medium and electronic equipment for establishing function stack
CN116932910A (en) Content pushing method, device, equipment and storage medium
CN117009168A (en) Application index acquisition method, device, equipment and medium
CN116257307A (en) Popup window control method, device, equipment and medium
CN117082116A (en) Session management method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20230506

Address after: Room 802, Information Building, 13 Linyin North Street, Pinggu District, Beijing, 101299

Applicant after: Beijing youzhuju Network Technology Co.,Ltd.

Address before: No. 715, 7th floor, building 3, 52 Zhongguancun South Street, Haidian District, Beijing 100081

Applicant before: Beijing infinite light field technology Co.,Ltd.

TA01 Transfer of patent application right
GR01 Patent grant
GR01 Patent grant