CN117851347A - File management method and device - Google Patents

File management method and device Download PDF

Info

Publication number
CN117851347A
CN117851347A CN202410051462.XA CN202410051462A CN117851347A CN 117851347 A CN117851347 A CN 117851347A CN 202410051462 A CN202410051462 A CN 202410051462A CN 117851347 A CN117851347 A CN 117851347A
Authority
CN
China
Prior art keywords
information
target file
target
file
interface
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202410051462.XA
Other languages
Chinese (zh)
Inventor
马刚
周芯宇
卢章玮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Lenovo Beijing Ltd
Original Assignee
Lenovo Beijing Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lenovo Beijing Ltd filed Critical Lenovo Beijing Ltd
Priority to CN202410051462.XA priority Critical patent/CN117851347A/en
Publication of CN117851347A publication Critical patent/CN117851347A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/16File or folder operations, e.g. details of user interfaces specifically adapted to file systems
    • G06F16/164File meta data generation

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The application discloses a file management method and device, wherein the method comprises the following steps: under the condition of determining an operation target file, acquiring corresponding environment information, wherein the environment information at least comprises at least one of the following: audio information, video information, interface information and text information; analyzing the environment information to obtain description information corresponding to the target file; and setting the description information as the identification of the target file.

Description

File management method and device
Technical Field
The embodiment of the application relates to the field of file management, and relates to a file management method and device.
Background
In modern society, people often need to process and view a large number of documents on different electronic devices such as mobile phones, personal computers, tablet computers, which causes a series of document management and retrieval problems. In the conventional file management system, only the content information of the file itself is stored when the file is stored, and the user is required to manually add meta information such as a tag to record the file context, but the efficiency of manually adding the meta information is low. When the file is stored, the operation log of the file is directly recorded, but a large amount of irrelevant information is contained, which is not beneficial to file management and retrieval.
Disclosure of Invention
In view of this, embodiments of the present application provide a file management method, apparatus, device, and storage medium.
The technical scheme of the embodiment of the application is realized as follows:
in a first aspect, an embodiment of the present application provides a file management method, where the method includes:
under the condition of determining an operation target file, acquiring corresponding environment information, wherein the environment information at least comprises at least one of the following: audio information, video information, interface information and text information;
analyzing the environment information to obtain description information corresponding to the target file;
and setting the description information as the identification of the target file.
In a second aspect, an embodiment of the present application provides a file management apparatus, including:
the first acquisition module is used for acquiring corresponding environment information under the condition of determining the operation target file, wherein the environment information at least comprises at least one of the following: audio information, video information, interface information and text information;
the analysis module is used for analyzing the environment information to obtain the description information corresponding to the target file;
and the setting module is used for setting the description information as the identification of the target file.
In a third aspect, an embodiment of the present application provides an electronic device, including a memory and a processor, where the memory stores a computer program executable on the processor, and where the processor implements the method described above when executing the program.
In a fourth aspect, embodiments of the present application provide a storage medium storing executable instructions for implementing the above method when executed by a processor.
In the embodiment of the application, under the condition that an operation target file is determined, corresponding environment information is acquired; then analyzing the environment information to obtain the description information corresponding to the target file; and finally, setting the description information as the identification of the target file. In this way, the environment information can be used as the target file description information to be set as the identification of the target file, so that the description information in the identification is more vivid, the vivid description information is longer in the memory of the user, and the file can be more easily memorized and managed by the user.
Drawings
Fig. 1 is a schematic implementation flow chart of a file management method according to an embodiment of the present application;
fig. 2A is a schematic implementation flow chart of obtaining environmental information according to an embodiment of the present application;
FIG. 2B is a screen capture image provided in an embodiment of the present application;
FIG. 2C is a dual screen capture image provided in an embodiment of the present application;
fig. 3A is a schematic implementation flow chart of obtaining description information according to an embodiment of the present application;
FIG. 3B is a dual screen shot provided in an embodiment of the present application;
fig. 4 is a schematic implementation flow chart of a method for retrieving a file according to an embodiment of the present application;
FIG. 5 is a schematic diagram of a composition structure of a document management apparatus according to an embodiment of the present disclosure;
fig. 6 is a schematic diagram of a hardware entity of an electronic device according to an embodiment of the present application.
Detailed Description
For the purposes, technical solutions and advantages of the embodiments of the present application to be more apparent, the specific technical solutions of the embodiments of the present application will be further described in detail below with reference to the accompanying drawings in the embodiments of the present application. The following examples are illustrative of the present application, but are not intended to limit the scope of the present application.
In the following description, reference is made to "some embodiments" which describe a subset of all possible embodiments, but it is to be understood that "some embodiments" can be the same subset or different subsets of all possible embodiments and can be combined with one another without conflict.
In the following description, the terms "first", "second", "third" and the like are merely used to distinguish similar objects and do not represent a specific ordering of the objects, it being understood that the "first", "second", "third" may be interchanged with a specific order or sequence, as permitted, to enable embodiments of the application described herein to be practiced otherwise than as illustrated or described herein.
Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs. The terminology used herein is for the purpose of describing embodiments of the present application only and is not intended to be limiting of the present application.
An embodiment of the present application provides a file management method, as shown in fig. 1, including:
step S110, under the condition of determining an operation target file, obtaining corresponding environment information, wherein the environment information at least comprises at least one of the following: audio information, video information, interface information and text information;
in some embodiments, the operation target file includes at least one of: storing the target file; moving the target file; and sharing the target file.
For example, receiving a target file at a user using a chat tool and storing the target file; the user moves the file from the D disk to the E disk stored in the electronic device; the target file is shared on the remote sharing screen of the user, or the target file is set to be in a sharing mode by the user.
In the implementation process, under the condition that the action of any operation target file is determined to occur, the environment information corresponding to the operation target file can be obtained. Here, the environment information includes at least one of: audio information, video information, interface information, and text information.
For example, when the user receives the target file and stores the target file, the audio information of the environment where the user is located is obtained, where the audio information may be the audio information of the user describing the file collected by using the microphone, the audio information may be music playing of the surrounding environment, and the audio information may also be the audio information of the audio and video being played by the electronic device.
The video information of the environment where the user is located is obtained, and the video information can be video playing information of the surrounding environment acquired by using the camera device, and can also be video information being played by the electronic equipment.
And acquiring interface information of the environment where the user is located, wherein the interface information can be the interface information of the electronic equipment.
The text information of the environment where the user is located is acquired, and the text information can be the text information of the surrounding environment acquired by the camera device or the text information of the playing interface of the electronic equipment.
Step S120, analyzing the environment information to obtain description information corresponding to the target file;
in some embodiments, the description information corresponding to the target file may be in audio form, or may be in video form, or may be in image form, or may be in text form.
Here, since the environment information may be in the form of audio, video, interface, etc., in the case that the description information is text form information, the text form description information corresponding to the target file may be obtained by analyzing the above environment information.
Step S130, setting the description information as the identification of the target file.
In the implementation process, the description information can be stored corresponding to the identification of the target file. For example, the description information may be stored in attribute information of the target file, or may be stored in hint information of the target file.
In the case of storing the description information to the hint information of the target file, the user may view the description information by clicking on the attribute information of the target file.
In the case of storing the description information to the hint information of the target file, the user can view the description information while moving the cursor to the target file.
In the embodiment of the application, under the condition that an operation target file is determined, corresponding environment information is acquired; then analyzing the environment information to obtain the description information corresponding to the target file; and finally, setting the description information as the identification of the target file. In this way, the environment information can be used as the target file description information to be set as the identification of the target file, so that the description information in the identification is more vivid, the vivid description information is longer in the memory of the user, and the file can be more easily memorized and managed by the user.
In some embodiments, in the case of determining the operation target file in step S110, the obtaining of the corresponding environment information is implemented by at least one of the following steps as shown in fig. 2A:
step S210, under the condition that the operation of the target file is determined, acquiring at least one frame of corresponding target interface information on an interface for operating the target file;
in some embodiments, at least one frame of target interface information corresponding to the interface for operating the target file may be image information of one display screen, that is, complete screen content of the electronic device. For example, fig. 2B is a screen capturing image provided in an embodiment of the present application.
In some embodiments, at least one frame of target interface information corresponding to the interface for operating the target file may also be displayed in two or more frames. For example, in the case where the user operates using a dual screen, screen shots of two screens may be acquired simultaneously. Fig. 2C is a dual screen capture image provided in an embodiment of the present application.
Step S220, collecting and operating environment audio information corresponding to the target file;
in the implementation process, the audio information of the file can be described by a user collected by a sound receiving device such as a microphone arranged on the electronic equipment, the audio information of music playing in the surrounding environment can be collected, and the audio information of the audio and video being played by the electronic equipment can be collected.
For example, the user may initiate a sound recording and begin the language description "the file is a file for providing reference assistance to the A customer". In this way, the target file can be efficiently located when the user retrieves the A-client help document. The audio information of the song being played by the user or the audio information of the audio and video played by the electronic equipment can also be collected. Thus, the user can search the song information or video background music in the memory to efficiently find the target file.
Step S230, acquiring and operating environment video information corresponding to the target file;
in the implementation process, the video playing information of the collected surrounding environment can be collected by using the camera device arranged on the electronic equipment, and the video information being played by the electronic equipment can be obtained.
For example, in the case of downloading a target file, the user is watching a movie, and then movie playback information may be collected. Thus, the user can search the movie playing information to find the target file. And under the condition of sharing the target file, the user plays the short video by utilizing the electronic equipment. In this way, the user can search the short video information for the target file.
Step S240, obtaining an operation mode and/or operation time for operating the target file.
In practice, the time provided by the operating system of the electronic device may be used to determine the time of operation of the target file.
For example, the manner of operation and/or the time of operation of the target file may be obtained. Thus, the user can find the target file based on the operation mode of the file, can find the target file based on the operation time of the file, and can find the target file based on the operation mode and the operation time of the file.
Here, the methods of acquiring the environmental information provided in the above steps S210 to S240 may be optionally used in combination to acquire more abundant environmental information.
In this embodiment of the present application, when determining the operation target file, obtaining the corresponding environmental information includes at least one step of: under the condition of determining to operate the target file, acquiring at least one frame of corresponding target interface information on an interface for operating the target file; collecting and operating environment audio information corresponding to the target file; collecting and operating environment video information corresponding to the target file; and acquiring an operation mode and/or operation time for operating the target file. Thus, the above provided method for acquiring the environmental information can be optionally combined for use to acquire more abundant environmental information.
In some embodiments, in the case that the above step S210 "determines to operate the target file, the obtaining the corresponding at least one frame of target interface information on the interface for operating the target file" may be implemented by the following steps:
step 211, determining a target time for operating the target file;
here, when the user performs file saving, moving, or sharing, the user can determine the corresponding save time, moving time, or sharing time.
In an implementation, a target time for operating the target file may be determined using a time of an operating system of the electronic device.
212, acquiring at least one frame of interface information in a time period corresponding to the target time on an interface for operating the target file based on a preset time period;
here, the user may set the preset time period according to the requirement and the storage capability of the electronic device system. For example, the preset duration may be set to 10 minutes. In the implementation process, at least one frame of interface information within 10 minutes before and after the target time can be acquired on the interface of the operation target file.
For example, the screen capture system may be run in the background of the electronic device, with screen captures taken every two seconds. The complete screen content of the current electronic device is recorded. This screen capture record remains for one hour. The background system monitors the file system at the same time, and when the file is found to be saved, moved or shared, 10 minutes before and after the file is called for screen capture records.
Step 213, comparing any two interface information in the at least one frame of interface information, and determining at least one frame of target interface information with the interface information more than the information threshold.
Here, since there may be a large number of similar duplicate interface information in the acquired at least one frame of interface information. In the implementation process, the information threshold value can be determined based on the actual requirement, then any two interface information in at least one frame of interface information are compared, and at least one frame of target interface information with the interface information more than the information threshold value is determined through comparison.
For example, multiple shots that differ by more than a certain threshold may be selected from among numerous shots of the screenshot record. Thus, the screen shots containing the key information are filtered out, and the number of the screen shots is reduced.
In the embodiment of the application, first, determining a target time for operating the target file; then, at least one frame of interface information in the time length corresponding to the target time is acquired on an interface for operating the target file based on a preset time length; and finally, comparing any two interface information in the at least one frame of interface information, and determining at least one frame of target interface information with the interface information more than an information threshold. In this way, key interface information before and after the time point of operating the target file can be acquired, so that effective description information of the target file can be extracted based on the key interface information.
In some embodiments, the above step S120 "parse the environmental information to obtain the description information corresponding to the target file" as shown in fig. 3A may be implemented by at least one of the following steps:
step S310, analyzing the at least one frame of target interface information to obtain the description information;
here, the description information may be presented in a text form.
In some embodiments, the target interface information may be converted into text form using a visual large model for describing the current operation of the target interface. Here, since there is software of operation in the target interface, and the content of the software, the visual large model can obtain the current operation of the user at the target interface.
In some embodiments, the target interface information may be parsed using optical character recognition (Optical Character Recognition, OCR) plus a text model.
In some embodiments, a "visual model (image understanding model)" and an "OCR plus text model" may be used in combination.
For example, the target interface may be a movie that the user is watching, where the movie has a national landscape, and the WeChat window is open, and a person who is called sends a message by WeChat: fast, presentation is sent to me. The target file (presentation) can be obtained by utilizing the visual large model analysis based on the target interface, and the user receives a certain WeChat when the user is in a national landscape on the computer and requests to send the WeChat to the user.
Thus, the stored object file not only has the information of the file itself, but also knows the front and back stories of the operation of the object file. The user can search the file according to the scene conveniently. For example, when the user looks at a scenic spot on a computer, the user receives a certain WeChat and requests to send the WeChat to the WeChat. After a long time, the user may remember to give a certain file when looking at the scenic spot, and the file name is not recorded. The user can be assisted in retrieving this file with the current scene description, as a result of the user's story being stored. The story is more vivid, the memory of human beings is longer, and the user is easier to memorize and manage files.
Step S320, analyzing the environment audio information corresponding to the target file to obtain the description information;
in some embodiments, the environmental audio information may be converted by using audio written text software to obtain text information corresponding to the environmental audio information, and the text information may be added to the description information.
In some embodiments, noise filtering may be performed on the environmental audio information to obtain relatively pure audio information as the description information.
Step S330, analyzing the environment video information corresponding to the target file to obtain the description information;
in some embodiments, the environmental video information may be split into image information and audio information, and then the image information and the audio information may be processed to obtain the description information. For example, the visual large model image information can be utilized to obtain corresponding text information, the audio information is converted into text information, and finally the text information is summarized to obtain the description information.
Step S340, adding the operation mode and/or the operation time to the description information.
In some embodiments, the operational mode may be added to the descriptive information. For example, a file moving operation of moving a target file from a C disc to a D disc may be added to the description information.
In some embodiments, the operation time may also be added to the descriptive information. For example, the time to download and save the target file may be added to the description information.
In some embodiments, the operation mode and the operation time may also be added to the description information at the same time. For example, the sharing operation and the sharing time of the sharing target file may be added to the description information at the same time.
In the implementation process, the methods for obtaining the description information provided in the above steps S310 to S340 may be used in any combination.
In this embodiment of the present application, analyzing the environmental information to obtain the description information corresponding to the target file includes at least one method of: analyzing the at least one frame of target interface information to obtain the description information; analyzing the environment audio information corresponding to the target file to obtain the description information; analyzing the environment video information corresponding to the target file to obtain the description information; and adding the operation mode and/or the operation time to the description information. Thus, different information corresponding to the target file can be analyzed to obtain the description information of the target file.
In some embodiments, the step S310 "parse the at least one frame of target interface information to obtain the description information" may be implemented by the following steps:
Step 311, converting the at least one frame of target interface information into text information by using a visual large model;
here, the vision macro model may be to construct a vision system having high recognition, analysis, and expression capabilities using a large amount of data and algorithms. The working principle of the human visual system can be imitated, the image is processed and analyzed, and the algorithm of the human visual system is continuously optimized through machine learning. For example, GPT-4V is a visually functional GPT-4, a large model of vision.
In an implementation process, the visual large model can be utilized to convert at least one frame of target interface information into text information.
For example, as shown in FIG. 2C, after submitting the double screen shot to GPT-4V, GPT-4V outputs the following text information:
this is a screenshot of a computer screen showing a number of open application windows.
The main window on the left side of the screen is a code editor (possibly Visual Studio Code) and the displayed file structure indicates that the Python programming language is being used. The code window shows what appears to be a program that merges multiple data files, comprising the steps of:
1. files are grouped based on file name parameters.
2. The files of each group are sequenced and combined, and the index, file name, file location, source file, usage format, related personnel, scene location, file content and summary information are combined into a data structure.
3. And finally, returning the combined data.
On the right side of the screen is a browser window, which displays a new tab page containing a search bar and some web shortcuts. There are several unread Gmail mails in the browser tab, as well as some bookmarks and other extensions.
At the bottom of the screen is a taskbar displaying some fixed program icons and running applications.
And 312, processing the text information by using the text large model to generate the description information.
In the implementation process, the trained text large model can be utilized to process the text information so as to extract key information from a large amount of text information.
For example, the text large model may be used to extract key information of "interface display user is processing some tasks of data processing or data integration, and at the same time, the browser window is opened" from the text information obtained as above.
In the embodiment of the application, firstly, the visual large model is utilized to convert the at least one frame of target interface information into text information; and then processing the text information by using a text large model to generate the description information. In this way, the conversion of target interface information into descriptive information using the visual and text large models may be achieved.
In some embodiments, the above step 312 "process the text information using a text big model, generating the descriptive information" may be implemented by:
step 3121, obtaining time information corresponding to each piece of target interface information;
in the implementation process, the acquired sequence of the acquired multiple target interfaces is smooth, and can be used as information to be added to the description information, so that time information corresponding to each target interface information needs to be acquired.
Step 3122, adding the time information to the text information to obtain text information including a time cue;
in the implementation process, the time information can be added to the corresponding text information to obtain the text information comprising the time clues. For example, corresponding time information may be added for text information corresponding to each interface information, and corresponding time information may also be added for each operation in one interface information.
And 3123, processing the text information including the time clues by using the text large model, and generating description information including the time clues corresponding to the target file.
For example, after all the shots are converted to text, the timeline information of the shots may be added to the text and submitted together to a text large model, which may generate an operating story for the file based on the descriptions with the timelines. And then placing the operation story of the file and the target file into a file management system together to be used as the identification of the target file.
For example, fig. 3B is a dual screen shot provided in an embodiment of the present application. In the implementation process, after the double screen capture as shown in FIG. 3B is submitted to GPT-4V, the output of GPT-4V is described in the following text:
this screen shot illustrates a working environment in which the dual display is used. The left screen shows a code editor window, which should be the PyCharm used, which is a popular Python Integrated Development Environment (IDE). The window is opened with a Python file named "ecs _ecosystem_tmp.py", and the code contains logic to process the data and generate chat conversation content. On the left side of the screen, the project structure and file list can be seen, indicating that the user may be doing software development or data processing work.
The right screen shows the desktop of Windows operating system, and the window shows chat software, which is commonly used for instant messaging in China. A record of file transfer occurs in the chat window, and there are two files named "chromeextensions_make_page2pdf.zip", which may be compressed packages of the browser extension program, whose file type is ZIP compressed files. This indicates that the user may be discussing or exchanging files about the browser extension with someone (in the contact list).
In summary, the employee may perform software development or data processing, and communicate with other people, possibly including file transmission. The file may be a browser extension or related file made by the developer, while the file displayed in the code editor on the left screen may be part of the development or a script related to the data processing.
The output content of the visual model is submitted to GPT4 with the timestamp of the screenshot, which requires the generation of a coherent story, resulting in the final "file story" as follows:
during a common work day of 2023, 11 and 29, a software developer is busy on his work computer. This is not just a simple daily routine of tapping the keyboard, but rather the day is focused on important moments of data processing and real-time communication.
11 a.m. for 14 minutes 07 seconds, staff is doing programming work using Visual Studio Code. His screen displays the unique symbols of the Python programming language, his goal being to merge multiple data files into one structured message. Through a series of steps including grouping, sorting, and merging of files, a database containing indexes, file information, and content summaries is ultimately formed. The browser on the right displays the new tab page open waiting for a possible online search or for the just received mail. This screenshot, in addition to showing the content of the file he is processing, also implies that he may be looking up information or waiting for important communications in order to solve the problem.
Immediately thereafter, at 11, 15 minutes and 31 seconds we see how much more tasks are taken by the employee. He starts to use the dual display to increase efficiency. The left screen shows "ecs _ecosystemtmp. Py" by PyCharm, a Python file that exposes software development and data processing logic. It appears that he is working to improve the data management of the system, perhaps in generating algorithmically driven chat conversation content. The screen on the right shows a more personalized side-instant messaging software is open and a file transfer record named "chromaex extensions_make_page2pdf.
In the embodiment of the application, time information corresponding to each piece of target interface information is firstly acquired; then adding the time information to the text information to obtain text information comprising a time cue; and finally, processing the text information comprising the time clues by using the text large model, and generating the description information comprising the time clues corresponding to the target file. In this way, descriptive information including a time cue can be obtained.
In some embodiments, the above step S130 "set the description information to the identifier of the target file" may be implemented by the following steps:
Step 131, extracting at least one scene tag from the description information;
in the implementation process, the corresponding scene tag can be extracted through the keyword in the description information. For example, "staff is doing programming work using Visual Studio Code" in the description information. His screen displays the unique symbols of the Python programming language, his goal being to merge multiple data files into one structured message. The "can extract farm Jing Biaoqian as" code edit ". For example, the screen on the right side of the description shows a more personalized side-instant messaging software is open, and a file transfer record named "chromeextensions_make_page2pdf. "Can extract scene label" communication software ".
In practice, this may be accomplished using large model extraction, or other natural language models, based on the final generated "file stories", from which the brief "scene tags" are extracted.
Step 132, setting the at least one scene tag as the identification of the target file.
For example, scene tags such as "code edit" and "communication software" may be set as the identification of the target file. In this way, the user can search through these tags when the user needs to search for the target file.
In the embodiment of the application, at least one scene tag is firstly extracted from the description information; the at least one scene tag is then set as an identification of the target file. Therefore, the scene tag is used as the identification of the target file, and the retrieval efficiency of the target file can be effectively improved.
In some embodiments, after the description information is set to the identifier of the target file in step S130", the embodiments of the present application provide a method for retrieving a file, as shown in fig. 4, where the method includes:
step S410, obtaining a retrieval instruction comprising target information;
in the implementation process, the user can search the target file based on any memory of the operation scene, operation time, operation mode and the like of the target file as target information.
For example, the target information may be set to "write code in the morning, pass to my browser plug-in package through chat tool", and retrieval may be performed without the file name of the target file.
Step S420, determining that the target information is matched with any one of the description information based on the retrieval instruction;
in the implementation process, the target information and the description information can be compared to determine the target file.
For example, the corresponding target file may be retrieved as "chromeextensions_make_page2pdf.zip" by comparing the target information "when writing code in certain morning, passed to my browser plug-in package through chat tool" with the stored description information of each file.
Step S430, determining the target file as a retrieved file.
In the embodiment of the application, firstly, a retrieval instruction comprising target information is obtained; then determining that the target information is matched with any one of the descriptive information based on the retrieval instruction; and finally, determining the target file as the retrieved file. Therefore, the file can be searched by the user based on the operation and scene information of the file in the memory, and the file searching efficiency is effectively improved under the condition that the user does not know the file name.
Based on the foregoing embodiments, the embodiments of the present application provide a file management apparatus, where the apparatus includes each module including each sub-module, and each sub-module includes a unit, and may be implemented by a processor in an electronic device; of course, the method can also be realized by a specific logic circuit; in practice, the processor may be a central processing unit (Central Processing Unit, CPU), microprocessor (Microprocessor Unit, MPU), digital signal processor (Digital Signal Process, DSP) or field programmable gate array (Field Programmable Gate Array, FPGA), etc.
Fig. 5 is a schematic structural diagram of a file management apparatus according to an embodiment of the present application, as shown in fig. 5, where the apparatus 500 includes:
the first obtaining module 510 obtains corresponding environmental information when determining the operation target file, where the environmental information includes at least one of the following: audio information, video information, interface information and text information;
the parsing module 520 is configured to parse the environmental information to obtain description information corresponding to the target file;
a setting module 530, configured to set the description information as an identifier of the target file.
In some embodiments, the operation target file includes at least one of: storing the target file; moving the target file; and sharing the target file.
In some embodiments, the first obtaining module 510 includes a first obtaining sub-module, a first collecting sub-module, a second collecting sub-module, and a second obtaining sub-module, where the first obtaining sub-module is configured to obtain, when determining to operate the target file, at least one frame of corresponding target interface information on an interface that operates the target file; the first acquisition sub-module is used for acquiring and operating the environment audio information corresponding to the target file; the second acquisition sub-module is used for acquiring and operating the environment video information corresponding to the target file; the second obtaining sub-module is used for obtaining the operation mode and/or the operation time for operating the target file.
In some embodiments, the first acquisition submodule includes a determining unit, an acquiring unit and an comparing unit, where the determining unit is configured to determine a target time for operating the target file; the acquiring unit is used for acquiring at least one frame of interface information in the time length corresponding to the target time on the interface for operating the target file based on the preset time length; the comparison unit is used for comparing any two interface information in the at least one frame of interface information and determining at least one frame of target interface information with the interface information more than an information threshold.
In some embodiments, the parsing module 520 includes sub-modules of at least one of: the system comprises a first analysis sub-module, a second analysis sub-module, a third analysis sub-module and an adding sub-module, wherein the first analysis sub-module is used for analyzing the at least one frame of target interface information to obtain the description information; the second analysis submodule is used for analyzing the environment audio information corresponding to the target file to obtain the description information; the third analysis submodule is used for analyzing the environment video information corresponding to the target file to obtain the description information; the adding submodule is used for adding the operation mode and/or the operation time to the description information.
In some embodiments, the first parsing sub-module includes a conversion unit and a processing unit, where the conversion unit is configured to convert the at least one frame of target interface information into text information using a visual large model; and the processing unit is used for processing the text information by using the text large model and generating the description information.
In some embodiments, the processing unit includes an obtaining subunit, an adding subunit, and a processing subunit, where the obtaining subunit is configured to obtain time information corresponding to each piece of target interface information; the adding subunit is configured to add the time information to the text information to obtain text information including a time cue; and the processing subunit is used for processing the text information comprising the time clues by using the text large model and generating the description information comprising the time clues corresponding to the target file.
In some embodiments, the setting module 530 includes an extracting sub-module and a setting sub-module, where the extracting sub-module is configured to extract at least one scene tag from the description information; the setting submodule is used for setting the at least one scene tag as the identification of the target file.
In some more embodiments, the apparatus further includes a second obtaining module, a first determining module, and a second determining module, where the second obtaining module is configured to obtain a retrieval instruction including the target information; the first determining module is used for determining that the target information is matched with any one of the description information based on the retrieval instruction; the second determining module is configured to determine the target file as a retrieved file.
The description of the apparatus embodiments above is similar to that of the method embodiments above, with similar advantageous effects as the method embodiments. For technical details not disclosed in the device embodiments of the present application, please refer to the description of the method embodiments of the present application for understanding.
It should be noted that, in the embodiment of the present application, if the method is implemented in the form of a software functional module, and sold or used as a separate product, the method may also be stored in a computer readable storage medium. Based on such understanding, the technical solutions of the embodiments of the present application may be embodied essentially or in a part contributing to the related art in the form of a software product stored in a storage medium, including several instructions for causing an electronic device (which may be a mobile phone, a tablet computer, a notebook computer, a desktop computer, etc.) to perform all or part of the methods described in the embodiments of the present application. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read Only Memory (ROM), a magnetic disk, an optical disk, or other various media capable of storing program codes. Thus, embodiments of the present application are not limited to any specific combination of hardware and software.
Accordingly, embodiments of the present application provide a storage medium having stored thereon a computer program which, when executed by a processor, implements the steps of the file management method provided in the above embodiments.
Correspondingly, an electronic device is provided in the embodiment of the present application, fig. 6 is a schematic diagram of a hardware entity of the electronic device provided in the embodiment of the present application, and as shown in fig. 6, the hardware entity of the device 600 includes: comprising a memory 601 and a processor 602, said memory 601 storing a computer program executable on the processor 602, said processor 602 implementing the steps of the file management method provided in the above embodiments when said program is executed.
The memory 601 is configured to store instructions and applications executable by the processor 602, and may also cache data (e.g., image data, audio data, voice communication data, and video communication data) to be processed or processed by the processor 602 and the modules in the electronic device 600, which may be implemented by a FLASH memory (FLASH) or a random access memory (Random Access Memory, RAM).
It should be noted here that: the description of the storage medium and apparatus embodiments above is similar to that of the method embodiments described above, with similar benefits as the method embodiments. For technical details not disclosed in the embodiments of the storage medium and the apparatus of the present application, please refer to the description of the method embodiments of the present application for understanding.
It should be appreciated that reference throughout this specification to "one embodiment" or "an embodiment" means that a particular feature, structure or characteristic described in connection with the embodiment is included in at least one embodiment of the present application. Thus, the appearances of the phrases "in one embodiment" or "in an embodiment" in various places throughout this specification are not necessarily all referring to the same embodiment. Furthermore, the particular features, structures, or characteristics may be combined in any suitable manner in one or more embodiments. It should be understood that, in various embodiments of the present application, the sequence numbers of the foregoing processes do not mean the order of execution, and the order of execution of the processes should be determined by the functions and internal logic thereof, and should not constitute any limitation on the implementation process of the embodiments of the present application. The foregoing embodiment numbers of the present application are merely for describing, and do not represent advantages or disadvantages of the embodiments.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
In the several embodiments provided in this application, it should be understood that the disclosed apparatus and method may be implemented in other ways. The above described device embodiments are only illustrative, e.g. the division of the units is only one logical function division, and there may be other divisions in practice, such as: multiple units or components may be combined or may be integrated into another system, or some features may be omitted, or not performed. In addition, the various components shown or discussed may be coupled or directly coupled or communicatively coupled to each other via some interface, whether indirectly coupled or communicatively coupled to devices or units, whether electrically, mechanically, or otherwise.
The units described above as separate components may or may not be physically separate, and components shown as units may or may not be physical units; can be located in one place or distributed to a plurality of network units; some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional unit in each embodiment of the present application may be integrated in one processing unit, or each unit may be separately used as one unit, or two or more units may be integrated in one unit; the integrated units may be implemented in hardware or in hardware plus software functional units.
Those of ordinary skill in the art will appreciate that: all or part of the steps for implementing the above method embodiments may be implemented by hardware related to program instructions, and the foregoing program may be stored in a computer readable storage medium, where the program, when executed, performs steps including the above method embodiments; and the aforementioned storage medium includes: a mobile storage device, a Read Only Memory (ROM), a magnetic disk or an optical disk, or the like, which can store program codes.
Alternatively, the integrated units described above may be stored in a computer readable storage medium if implemented in the form of software functional modules and sold or used as a stand-alone product. Based on such understanding, the technical solutions of the embodiments of the present application may be embodied essentially or in a part contributing to the related art in the form of a software product stored in a storage medium, including several instructions for causing an electronic device (which may be a mobile phone, a tablet computer, a notebook computer, a desktop computer, etc.) to perform all or part of the methods described in the embodiments of the present application. And the aforementioned storage medium includes: various media capable of storing program codes, such as a removable storage device, a ROM, a magnetic disk, or an optical disk.
The methods disclosed in the several method embodiments provided in the present application may be arbitrarily combined without collision to obtain a new method embodiment.
The features disclosed in the several product embodiments provided in the present application may be combined arbitrarily without conflict to obtain new product embodiments.
The features disclosed in the several method or apparatus embodiments provided in the present application may be arbitrarily combined without conflict to obtain new method embodiments or apparatus embodiments.
The foregoing is merely an embodiment of the present application, but the protection scope of the present application is not limited thereto, and any person skilled in the art can easily think about changes or substitutions within the technical scope of the present application, and the changes and substitutions are intended to be covered in the protection scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.

Claims (10)

1. A method of file management, the method comprising:
under the condition of determining an operation target file, acquiring corresponding environment information, wherein the environment information at least comprises at least one of the following: audio information, video information, interface information and text information;
Analyzing the environment information to obtain description information corresponding to the target file;
and setting the description information as the identification of the target file.
2. The method of claim 1, the operation target file comprising at least one of:
storing the target file;
moving the target file;
and sharing the target file.
3. The method according to claim 1, wherein in the case of determining the operation target file, the corresponding environment information is acquired, comprising at least one of the following steps:
under the condition of determining to operate the target file, acquiring at least one frame of corresponding target interface information on an interface for operating the target file;
collecting and operating environment audio information corresponding to the target file;
collecting and operating environment video information corresponding to the target file;
and acquiring an operation mode and/or operation time for operating the target file.
4. The method of claim 3, wherein the obtaining, on the interface that operates the target file, the corresponding at least one frame of target interface information includes:
determining a target time for operating the target file;
acquiring at least one frame of interface information in the time length corresponding to the target time on an interface for operating the target file based on a preset time length;
And comparing any two interface information in the at least one frame of interface information, and determining at least one frame of target interface information with the interface information more than an information threshold.
5. The method of claim 3, wherein the parsing the environmental information to obtain the description information corresponding to the target file includes at least one of the following methods:
analyzing the at least one frame of target interface information to obtain the description information;
analyzing the environment audio information corresponding to the target file to obtain the description information;
analyzing the environment video information corresponding to the target file to obtain the description information;
and adding the operation mode and/or the operation time to the description information.
6. The method of claim 5, wherein said parsing the at least one frame of target interface information to obtain the description information comprises:
converting the at least one frame of target interface information into text information by utilizing a visual large model;
and processing the text information by using a text large model to generate the description information.
7. The method of claim 6, wherein the processing the text information using a text large model to generate the descriptive information comprises:
acquiring time information corresponding to each piece of target interface information;
Adding the time information to the text information to obtain text information comprising a time cue;
and processing the text information comprising the time clues by using the text large model, and generating the description information comprising the time clues corresponding to the target file.
8. The method of claim 1, the setting the description information as the identification of the target file, comprising:
extracting at least one scene tag from the description information;
and setting the at least one scene tag as the identification of the target file.
9. The method of any one of claims 1 to 8, further comprising:
acquiring a retrieval instruction comprising target information;
determining that the target information is matched with any one of the descriptive information based on the retrieval instruction;
and determining the target file as a retrieved file.
10. A file management apparatus, the apparatus comprising:
the first acquisition module is used for acquiring corresponding environment information under the condition of determining the operation target file, wherein the environment information at least comprises at least one of the following: audio information, video information, interface information and text information;
the analysis module is used for analyzing the environment information to obtain the description information corresponding to the target file;
And the setting module is used for setting the description information as the identification of the target file.
CN202410051462.XA 2024-01-12 2024-01-12 File management method and device Pending CN117851347A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202410051462.XA CN117851347A (en) 2024-01-12 2024-01-12 File management method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202410051462.XA CN117851347A (en) 2024-01-12 2024-01-12 File management method and device

Publications (1)

Publication Number Publication Date
CN117851347A true CN117851347A (en) 2024-04-09

Family

ID=90534267

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202410051462.XA Pending CN117851347A (en) 2024-01-12 2024-01-12 File management method and device

Country Status (1)

Country Link
CN (1) CN117851347A (en)

Similar Documents

Publication Publication Date Title
US11055342B2 (en) System and method for rich media annotation
CN111143610B (en) Content recommendation method and device, electronic equipment and storage medium
CN113115099B (en) Video recording method and device, electronic equipment and storage medium
US8126294B2 (en) Video structuring device
US20150370808A1 (en) Intelligent video summaries in information access
US10366154B2 (en) Information processing device, information processing method, and computer program product
CN110377908B (en) Semantic understanding method, semantic understanding device, semantic understanding equipment and readable storage medium
CN110414404A (en) Image processing method, device and storage medium based on instant messaging
CN110781328A (en) Video generation method, system, device and storage medium based on voice recognition
CN110602516A (en) Information interaction method and device based on live video and electronic equipment
CN113569037A (en) Message processing method and device and readable storage medium
KR20100095633A (en) Method and apparatus for playing pictures
CN113536172B (en) Encyclopedia information display method and device and computer storage medium
CN111279333B (en) Language-based search of digital content in a network
CN109710799B (en) Voice interaction method, medium, device and computing equipment
CN112040339A (en) Method and device for making video data, computer equipment and storage medium
WO2024140432A1 (en) Ingredient recommendation method based on knowledge graph, and device and storage medium
JP2013054417A (en) Program, server and terminal for tagging content
CN113542797A (en) Interaction method and device in video playing and computer readable storage medium
CN100388671C (en) Method for providing local user operation inlet in instant communication client
CN117851347A (en) File management method and device
JP2022181319A (en) Video search apparatus, video search system, and program
CN113269855A (en) Method, equipment and storage medium for converting text semantics into scene animation
CN114245174B (en) Video preview method and related equipment
Popescu-Belis et al. TQB: Accessing Multimodal Data Using a Transcript-based Query and Browsing Interface.

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination