[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

WO2023019510A1 - Data indexing method, apparatus and device, and storage medium - Google Patents

Data indexing method, apparatus and device, and storage medium Download PDF

Info

Publication number
WO2023019510A1
WO2023019510A1 PCT/CN2021/113518 CN2021113518W WO2023019510A1 WO 2023019510 A1 WO2023019510 A1 WO 2023019510A1 CN 2021113518 W CN2021113518 W CN 2021113518W WO 2023019510 A1 WO2023019510 A1 WO 2023019510A1
Authority
WO
WIPO (PCT)
Prior art keywords
scene
data
segment
video data
information
Prior art date
Application number
PCT/CN2021/113518
Other languages
French (fr)
Chinese (zh)
Inventor
金晨
卢红喜
夏欢
周俊杰
李国庆
Original Assignee
浙江吉利控股集团有限公司
宁波吉利汽车研究开发有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 浙江吉利控股集团有限公司, 宁波吉利汽车研究开发有限公司 filed Critical 浙江吉利控股集团有限公司
Priority to PCT/CN2021/113518 priority Critical patent/WO2023019510A1/en
Priority to CN202180099925.4A priority patent/CN117597680A/en
Publication of WO2023019510A1 publication Critical patent/WO2023019510A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/71Indexing; Data structures therefor; Storage structures

Definitions

  • the present application relates to the technical field of data processing, and in particular to a data indexing method, device, equipment and storage medium.
  • the main purpose of this application is to propose a data indexing method, device, device, and storage medium, aiming at solving the problem of separately intercepting, naming, and storing scenes in video data in the prior art, which takes a long time and consumes a lot of manpower and material resources technical issues.
  • the present application provides a data indexing method, the data indexing method includes the following steps:
  • a video data index corresponding to each scene segment in the video data to be processed is constructed according to the data inter-frame tag and the scene classification information.
  • the generation of data inter-frame labels and scene classification information corresponding to each scene segment according to the segment information includes:
  • the generating the data inter-frame tags corresponding to each scene segment according to the data frame information includes:
  • Model training is performed according to the target scene segment.
  • the extracting the target scene segment from the video data to be processed according to the target scene information and the video data index includes:
  • the extracting target scene segments from the video data to be processed according to the target scene classification information and the target data inter-frame tags includes:
  • the performing model training according to the target scene segment includes:
  • the target scene segments are sorted according to the corresponding start time and end time of each target scene segment, and the sorted target scene segments are obtained;
  • Model training is performed sequentially according to the sorted target scene segments.
  • the present application also proposes a data indexing device, the data indexing device includes:
  • the scene classification module is used to classify the scene of the video data to be processed, and obtain a plurality of scene fragments
  • An information acquisition module configured to acquire segment information corresponding to each scene segment
  • a data generation module configured to generate inter-frame labels and scene classification information corresponding to each scene segment according to the segment information
  • a data index module configured to construct a video data index corresponding to each scene segment in the video data to be processed according to the data inter-frame tags and the scene classification information.
  • the present application also proposes a data indexing device, the data indexing device includes: a memory, a processor, and a data indexing program stored in the memory and operable on the processor, the When the data indexing program is executed by the processor, the above-mentioned data indexing method is realized.
  • the present application also proposes a storage medium, on which a data index program is stored, and when the data index program is executed by a processor, the above data index method is implemented.
  • the data indexing method proposed in this application performs scene classification on the video data to be processed to obtain multiple scene segments; obtains segment information corresponding to each scene segment; generates data inter-frame labels and scene classification information corresponding to each scene segment according to the segment information ; Constructing a video data index corresponding to each scene segment in the video data to be processed according to the data inter-frame label and the scene classification information.
  • FIG. 1 is a schematic structural diagram of a data indexing device in a hardware operating environment involved in the solution of the embodiment of the present application;
  • FIG. 2 is a schematic flow chart of the first embodiment of the data indexing method of the present application
  • FIG. 3 is a schematic diagram of a driving scene segment of an embodiment of the data indexing method of the present application.
  • FIG. 4 is a schematic diagram of weather scene fragments in an embodiment of the data indexing method of the present application.
  • FIG. 5 is a schematic diagram of a combination of a driving scene and a weather scene according to an embodiment of the data indexing method of the present application;
  • FIG. 6 is a schematic flow chart of the second embodiment of the data indexing method of the present application.
  • FIG. 7 is a schematic diagram of a data frame of a scene segment in an embodiment of the data indexing method of the present application.
  • FIG. 8 is a schematic flowchart of a third embodiment of the data indexing method of the present application.
  • FIG. 9 is a schematic diagram of functional modules of the first embodiment of the data indexing device of the present application.
  • FIG. 1 is a schematic structural diagram of a data indexing device in a hardware operating environment involved in an embodiment of the present application.
  • the data indexing device may include: a processor 1001 , such as a central processing unit (Central Processing Unit, CPU), a communication bus 1002 , a user interface 1003 , a network interface 1004 , and a memory 1005 .
  • the communication bus 1002 is used to realize connection and communication between these components.
  • the user interface 1003 may include a display screen (Display) and an input unit such as a button, and the optional user interface 1003 may also include a standard wired interface and a wireless interface.
  • the network interface 1004 may include a standard wired interface and a wireless interface (such as a Wi-Fi interface).
  • the memory 1005 may be a high-speed random access memory (Random Access Memory, RAM), or a stable memory (non-volatile memory), such as a disk memory.
  • RAM Random Access Memory
  • non-volatile memory such as a disk memory.
  • the memory 1005 may also be a storage device independent of the aforementioned processor 1001 .
  • FIG. 1 does not constitute a limitation on the data indexing device, and may include more or less components than those shown in the figure, or combine some components, or arrange different components.
  • memory 1005 as a storage medium may include an operating system, a network communication module, a user interface module, and a data index program.
  • the network interface 1004 is mainly used to connect to the external network and perform data communication with other network devices;
  • the user interface 1003 is mainly used to connect to user equipment and perform data communication with the user equipment; this application
  • the device invokes the data indexing program stored in the memory 1005 through the processor 1001, and executes the data indexing method provided in the embodiment of the present application.
  • FIG. 2 is a schematic flowchart of the first embodiment of the data indexing method of the present application.
  • the data indexing method includes the following steps:
  • Step S10 performing scene classification on the video data to be processed to obtain a plurality of scene fragments.
  • the executor of this embodiment may be a data indexing device, such as a computer device with data processing functions, or other devices capable of realizing the same or similar functions, which is not limited in this embodiment.
  • the video data to be processed in this embodiment can be the video data obtained by collecting the environmental data during the driving process of the test vehicle, which can be automatically collected by the vehicle-mounted camera equipment on the test vehicle, or can be collected by setting the The external camera equipment on the test vehicle can automatically collect the environmental data, and other methods can also be used to automatically collect the environmental data to obtain the video data to be processed, which is not limited in this embodiment.
  • the test vehicle has been driving for 6 hours, and the environmental data has been collected automatically during the driving process of the test vehicle to obtain 6 hours of video data. After the data collection, the collected video data needs to be processed. processing, the video data can be used as video data to be processed.
  • scene classification may be performed on the video data to be processed to obtain multiple scene segments. It should be noted that, in this step of the solution, it is not necessary to intercept these scene fragments from the video data to be processed, but only need to identify these scene fragments in the video data to be processed. Specifically, scene classification is performed on the video data to be processed according to a preset classification type, and scene classification may also be performed in other ways, which is not limited in this embodiment.
  • the preset classification type can be set in advance according to the actual use requirements. For example, if the weather scene is used to distinguish, it can be divided into sunny days, rainy days, foggy days, etc., and further, sunny days can be subdivided into sunny days and cloudy days. , few clouds, etc., rainy days can be subdivided into heavy rain, moderate rain, light rain, etc., and foggy days can be subdivided into equal dense fog, medium fog, light fog, etc. If the driving scene is distinguished, it can be divided into cruising, lane changing, braking, etc. Further, cruising can be subdivided into high-speed cruising, medium-speed cruising, low-speed cruising, etc.
  • Lane changing can be subdivided into left lane changing, Right lane change, continuous lane change, etc.
  • Braking can be subdivided into emergency braking, slow braking, point braking, etc.
  • other more scene classification types may also be included, such as time scenes, terrain scenes, etc., which are not limited in this embodiment.
  • the preset classification types include The weather scene and the driving scene are taken as examples for illustration.
  • image analysis can be performed on the video data to be processed, and multiple scene segments in the video data to be processed can be obtained according to the image analysis results.
  • Figure 3 is a schematic diagram of a driving scene segment
  • the horizontal line in Figure 3 is the time axis of the video data to be processed
  • the video data to be processed can be identified
  • the scene segments in the video data include a lane change scene segment 1 , a cruising scene segment 1 , and a lane change scene segment 2 .
  • Figure 4 is a schematic diagram of a weather scene segment
  • the horizontal line in Figure 4 is the time axis of the video data to be processed, and the video data to be processed can be identified
  • the scene segments include a sunny scene segment 1 and a rainy day scene segment 1 .
  • Figure 5 is a schematic diagram of the combination of the driving scene and the weather scene
  • the horizontal line in Figure 5 is the time axis of the video data to be processed.
  • the driving scene and the weather scene can also be combined to obtain four scene fragments A1, A2, A3, and A4 in Figure 5, where A1 is the lane change scene fragment 1 on a sunny day, A2 is the cruising scene segment 1 in sunny days, A3 is the cruising scene segment 1 in rainy days, and A4 is the lane changing scene segment 1 in rainy days.
  • Step S20 acquiring segment information corresponding to each scene segment.
  • scene segments also contain many types of information. Therefore, after determining a plurality of scene fragments contained in the video data to be processed, the fragment information corresponding to these scene fragments can be obtained, wherein the fragment information can include data frame information and fragment attribute information, besides, it can also include other Type information, which is not limited in this embodiment.
  • Step S30 generating data inter-frame tags and scene classification information corresponding to each scene segment according to the segment information.
  • inter-frame labeling can be performed on each scene segment according to the segment information, so as to determine the start frame and the end frame corresponding to each scene segment, and then generate data inter-frame tags.
  • Scene classification information may also be generated according to attribute information such as scene category, duration, and storage location included in the fragment information.
  • Step S40 constructing a video data index corresponding to each scene segment in the video data to be processed according to the inter-frame tag of the data and the scene classification information.
  • a video data index corresponding to each scene segment in the video data to be processed may be constructed according to the data inter-frame tag and scene classification information.
  • These video data indexes can be stored, and the complete video data to be processed can be stored directly, without time-consuming operations such as separate interception, naming and storage of the video data to be processed, and scene fragments in the video data to be processed can be used , the relevant information of the scene segments to be used can be determined directly according to the video data index, and then these scene segments are extracted from the video data to be processed for use.
  • the video data index can be stored in the form of a database or a table, or can be stored in other programming languages.
  • these video data indexes can be stored in an EXCLE table.
  • the video data index may also be stored in other ways, which is not limited in this embodiment.
  • storage of the video data index in a table is taken as an example for illustration. Compared with segmented video data, the video data index requires less storage space, is easy to store, and does not require complicated management. The information can be clearly recorded through the table, which can reduce the cost of data processing. time and avoid wasting a lot of manpower and material resources.
  • the video data indexes corresponding to all scene segments in the video data to be processed can be recorded in the same table for storage, and the video data indexes corresponding to different types of scene segments can also be recorded in different tables for storage.
  • the video data index corresponding to the scene segment corresponding to the weather scene is recorded in a table for storage
  • the video data index corresponding to the scene segment corresponding to the driving scene is recorded in another table for storage.
  • it can also be further refined and the video data index corresponding to the scene segment corresponding to the sunny scene is recorded in a table for storage, and the video data index corresponding to the corresponding scene segment corresponding to the rainy scene is recorded in another table. storage.
  • other storage methods may also be used for storage according to actual usage requirements, which is not limited in this embodiment.
  • scene classification is performed on the video data to be processed to obtain a plurality of scene segments; segment information corresponding to each scene segment is obtained; data inter-frame tags and scene classification information corresponding to each scene segment are generated according to the segment information; The data inter-frame tags and the scene classification information construct a video data index corresponding to each scene segment in the video data to be processed.
  • the step S30 includes:
  • Step S301 Determine the data frame information and segment attribute information corresponding to each scene segment according to the segment information.
  • video data may consist of multiple frames of image data
  • the data frame information in this embodiment refers to video frame information related to each scene segment.
  • the frame header information and frame tail information corresponding to each scene segment can be determined according to the data frame information, wherein the frame header information refers to the video frame information at the beginning of the scene segment, and the frame tail information refers to the video frame at the end of the scene segment information.
  • data frame information corresponding to the scene segment can be generated according to the frame header information and frame trailer information.
  • Figure 7 is a schematic diagram of the data frame of the scene segment
  • the frame header of the lane change scene segment 1 is O1
  • the frame The tail is O2, that is, the video data between the beginning of the O1 frame and the end of the O2 frame in the video to be processed is the video data corresponding to the lane change scene segment 1
  • the frame header of the cruising scene segment 1 is O2
  • the frame tail is O3, that is, the video data between the beginning of the O2 frame and the end of the O3 frame in the video to be processed is the video data corresponding to the cruise scene segment 1
  • the frame header of the lane change scene segment 2 is O3, and the frame tail is O4, That is, the video data between the beginning of the O3th frame and the end of the O4th frame in the video to be processed is the video data corresponding to the lane change scene segment 2.
  • Step S302 generating data inter-frame tags corresponding to each scene segment according to the data frame information, and generating scene classification information corresponding to each scene segment according to the segment attribute information.
  • inter-frame labeling can be performed on each scene segment in the video data to be processed according to the data frame information to generate a data inter-frame tag corresponding to each scene segment, and the scene segment can be determined according to the data inter-frame tag corresponding to the scene segment The position in the video data to be processed.
  • the segment attribute information includes attributes such as scene category, duration, and storage location
  • the scene classification information corresponding to each scene segment can be generated according to the segment attribute information, and the scene classification information corresponding to the scene segment can be determined. Attributes such as classification information corresponding to the scene segment.
  • both the data inter-frame label and the scene classification information contain partial information corresponding to each scene segment in the video data to be processed, the former is used for positioning and the latter is used for classification. Therefore, when it is necessary to use the scene segment corresponding to Data can be indexed and located through inter-frame labels and scene classification information to facilitate data extraction.
  • the data frame information and segment attribute information corresponding to each scene segment are determined according to the segment information; the data inter-frame tags corresponding to each scene segment are generated according to the data frame information, and generated according to the segment attribute information Scene classification information corresponding to each scene segment.
  • positioning and classification can be carried out respectively through the data inter-frame tags and scene classification information corresponding to each scene segment in the video data to be processed, and when the scene segment in the video data to be processed needs to be used, it can be conveniently obtained from the video data to be processed Data corresponding to the scene clip is extracted.
  • the third embodiment of the data indexing method of this application is proposed based on the first embodiment or the second embodiment.
  • the description is made based on the first embodiment.
  • the steps After S40 it also includes:
  • Step S50 when receiving a model training instruction, determine target scene information according to the model training instruction.
  • the deep learning model can be divided into multiple purposes and corresponding training types. For example, if it is a deep learning model used for weather scenes, the corresponding training type is weather scene training, which needs to be paired with some weather scene related data. The deep learning model is trained; if it is a deep learning model for driving scenes, the corresponding training type is driving scene training, and some data related to driving scenes is needed to train the deep learning model.
  • the computer device when it receives the model training instruction, it can determine the target scene information corresponding to the currently required data according to the model training instruction. For example, when data related to a weather scene is currently needed, the corresponding target scene information is a weather scene; when data related to a driving scene is currently needed, the corresponding target scene information is a driving scene.
  • Step S60 extracting target scene segments from the video data to be processed according to the target scene information and the video data index.
  • the target scene segment corresponding to the target scene information can be easily found according to the stored video data index, and The data corresponding to the target scene segment is extracted from the video data to be processed for model training.
  • the target scene information can be matched with the scene classification information in the video data index, and the target scene classification information can be determined according to the matching result. Furthermore, according to the video data index, the scene segment corresponding to the target scene classification information can be used as the target scene segment, and the data inter-frame label corresponding to the target scene segment can be used as the target data inter-frame label, and then according to the target scene classification information and the target data frame The inter tag extracts the data corresponding to the target scene segment from the video data to be processed.
  • the target scene classification information related to the lane-changing scene can be matched in the video data index, and then Determine the corresponding target scene segments as lane change scene segment 1 and lane change scene segment 2, and further determine that the data inter-frame label corresponding to lane change scene segment 1 is frame O1 ⁇ O2, and the data corresponding to lane change scene segment 2
  • the inter-frame labels are frame O3 ⁇ O4. Therefore, the data corresponding to these two lane-changing scene segments can be extracted from the video data to be processed for model training.
  • the storage location of the video data to be processed may be first determined according to the target scene classification information.
  • the data location can be positioned according to the target scene classification information and the inter-frame labels of the target data to determine the start time and end time of the target scene segment in the video data to be processed, wherein the start time and the target scene segment's The frame header corresponds, and the end time corresponds to the frame end of the target scene segment. Then, the data corresponding to the target scene segment may be extracted from the video data to be processed according to the start time and the end time corresponding to the target scene segment.
  • Step S70 perform model training according to the target scene segment.
  • the data corresponding to the target scene segment is directly extracted from the video data to be processed for model training. If there are multiple target scene segments, the target scene segments can be sorted according to the corresponding start time and end time of each target scene segment, and the sorted target scene segments are obtained, and the model training is performed according to the sorted target scene segments in turn.
  • the lane change scene segment 1 and the lane change scene segment 2 according to the sorting result, it can be known that the lane change scene segment 1 is before the lane change scene segment 2, therefore, it can be passed through
  • the data corresponding to the lane change scene segment 1 is used for model training, and then after the training is completed, the jump is automatically performed, and then the model training is performed with the data corresponding to the lane change scene segment 2. It can be understood that if there are more target scene fragments, the above method can continue to jump according to the sorting results, and carry out model training until all the data corresponding to the target scene fragments have been trained for the model.
  • the data corresponding to the target scene segment can be automatically indexed through the video data index during model training, and automatically and quickly linked into the scene data channel. After the training of the data corresponding to a target scene segment When , it can automatically switch to the data corresponding to the next target scene segment, and continue training, so as to achieve a better effect of automatic training of the model.
  • target scene information is determined according to the model training instruction; target scene segments are extracted from the video data to be processed according to the target scene information and the video data index; Model training is performed according to the target scene segment.
  • target scene information is used to perform scene matching, and the current needs are located according to the video data index.
  • the target scene segment and extract the data corresponding to the target scene segment from the original video data to be processed for model training. Through indexing and fast positioning, the efficiency of model training is improved, and labor costs are also saved.
  • the embodiment of the present application also proposes a storage medium, on which a data index program is stored, and when the data index program is executed by a processor, the steps of the above-mentioned data index method are implemented.
  • the storage medium adopts all the technical solutions of all the above-mentioned embodiments, it at least has all the beneficial effects brought by the technical solutions of the above-mentioned embodiments, which will not be repeated here.
  • the embodiment of the present application also proposes a data indexing device, the data indexing device includes:
  • the scene classification module 10 is configured to perform scene classification on the video data to be processed to obtain a plurality of scene fragments.
  • the video data to be processed in this embodiment can be the video data obtained by collecting the environmental data during the driving process of the test vehicle, which can be automatically collected by the vehicle-mounted camera equipment on the test vehicle, or can be collected by setting the The external camera equipment on the test vehicle can automatically collect the environmental data, and other methods can also be used to automatically collect the environmental data to obtain the video data to be processed, which is not limited in this embodiment.
  • the test vehicle has been driving for 6 hours, and the environmental data has been collected automatically during the driving process of the test vehicle to obtain 6 hours of video data. After the data collection, the collected video data needs to be processed. processing, the video data can be used as video data to be processed.
  • scene classification may be performed on the video data to be processed to obtain multiple scene segments. It should be noted that, in this step of the solution, it is not necessary to intercept these scene fragments from the video data to be processed, but only need to identify these scene fragments in the video data to be processed. Specifically, the scene classification is performed on the video data to be processed according to the preset classification type, and the scene classification may also be performed in other ways, which is not limited in this embodiment.
  • the preset classification type can be set in advance according to the actual use requirements. For example, if the weather scene is used to distinguish, it can be divided into sunny days, rainy days, foggy days, etc., and further, sunny days can be subdivided into sunny days and cloudy days. , few clouds, etc., rainy days can be subdivided into heavy rain, moderate rain, light rain, etc., and foggy days can be subdivided into equal dense fog, medium fog, light fog, etc. If the driving scene is distinguished, it can be divided into cruising, lane changing, braking, etc. Further, cruising can be subdivided into high-speed cruising, medium-speed cruising, low-speed cruising, etc.
  • Lane changing can be subdivided into left lane changing, Right lane change, continuous lane change, etc.
  • Braking can be subdivided into emergency braking, slow braking, point braking, etc.
  • other more scene classification types may also be included, such as time scenes, terrain scenes, etc., which are not limited in this embodiment.
  • the preset classification types include The weather scene and the driving scene are taken as examples for illustration.
  • image analysis can be performed on the video data to be processed, and multiple scene segments in the video data to be processed can be obtained according to the image analysis results.
  • Figure 3 is a schematic diagram of a driving scene segment
  • the horizontal line in Figure 3 is the time axis of the video data to be processed
  • the video data to be processed can be identified
  • the scene segments in the video data include a lane change scene segment 1 , a cruising scene segment 1 , and a lane change scene segment 2 .
  • Figure 4 is a schematic diagram of a weather scene segment
  • the horizontal line in Figure 4 is the time axis of the video data to be processed, and the video data to be processed can be identified
  • the scene segments include a sunny scene segment 1 and a rainy day scene segment 1 .
  • Figure 5 is a schematic diagram of the combination of the driving scene and the weather scene
  • the horizontal line in Figure 5 is the time axis of the video data to be processed.
  • the driving scene and the weather scene can also be combined to obtain four scene fragments A1, A2, A3, and A4 in Figure 5, where A1 is the lane change scene fragment 1 on a sunny day, A2 is the cruising scene segment 1 in sunny days, A3 is the cruising scene segment 1 in rainy days, and A4 is the lane changing scene segment 1 in rainy days.
  • the information acquisition module 20 is configured to acquire segment information corresponding to each scene segment.
  • scene segments also contain many types of information. Therefore, after determining a plurality of scene fragments contained in the video data to be processed, the fragment information corresponding to these scene fragments can be obtained, wherein the fragment information can include data frame information and fragment attribute information, besides, it can also include other Type information, which is not limited in this embodiment.
  • the data generation module 30 is configured to generate data inter-frame tags and scene classification information corresponding to each scene segment according to the segment information.
  • inter-frame labeling can be performed on each scene segment according to the segment information, so as to determine the start frame and the end frame corresponding to each scene segment, and then generate data inter-frame tags.
  • Scene classification information may also be generated according to attribute information such as scene category, duration, and storage location included in the fragment information.
  • the data index module 40 is configured to construct a video data index corresponding to each scene segment in the video data to be processed according to the inter-frame tag of the data and the scene classification information.
  • a video data index corresponding to each scene segment in the video data to be processed may be constructed according to the data inter-frame tag and scene classification information.
  • These video data indexes can be stored, and the complete video data to be processed can be stored directly, without time-consuming operations such as separate interception, naming and storage of the video data to be processed, and scene fragments in the video data to be processed can be used , the relevant information of the scene segments to be used can be determined directly according to the video data index, and then these scene segments are extracted from the video data to be processed for use.
  • the video data index can be stored in the form of a database or a table, or can be stored in other programming languages.
  • these video data indexes can be stored in an EXCLE table.
  • the video data index may also be stored in other ways, which is not limited in this embodiment.
  • storage of the video data index in a table is taken as an example for illustration. Compared with segmented video data, the video data index requires less storage space, is easy to store, and does not require complicated management. The information can be clearly recorded through the table, which can reduce the cost of data processing. time and avoid wasting a lot of manpower and material resources.
  • the video data indexes corresponding to all scene segments in the video data to be processed can be recorded in the same table for storage, and the video data indexes corresponding to different types of scene segments can also be recorded in different tables for storage.
  • the video data index corresponding to the scene segment corresponding to the weather scene is recorded in a table for storage
  • the video data index corresponding to the scene segment corresponding to the driving scene is recorded in another table for storage.
  • it can also be further refined and the video data index corresponding to the scene segment corresponding to the sunny scene is recorded in a table for storage, and the video data index corresponding to the corresponding scene segment corresponding to the rainy scene is recorded in another table. storage.
  • other storage methods may also be used for storage according to actual usage requirements, which is not limited in this embodiment.
  • scene classification is performed on the video data to be processed to obtain a plurality of scene segments; segment information corresponding to each scene segment is obtained; data inter-frame tags and scene classification information corresponding to each scene segment are generated according to the segment information; The data inter-frame tags and the scene classification information construct a video data index corresponding to each scene segment in the video data to be processed.
  • the data generation module 30 is further configured to determine the data frame information and segment attribute information corresponding to each scene segment according to the segment information; generate the data frame information corresponding to each scene segment according to the data frame information tags, and generate scene classification information corresponding to each scene segment according to the segment attribute information.
  • the data generating module 30 is further configured to determine frame header information and frame trailer information corresponding to each scene segment according to the data frame information; generate each frame header information and frame trailer information according to the frame header information and the frame trailer information The inter-frame label of the data corresponding to the scene fragment.
  • the data indexing device further includes a model training module, configured to determine target scene information according to the model training command when receiving a model training command; Extracting a target scene segment from the video data to be processed; performing model training according to the target scene segment.
  • a model training module configured to determine target scene information according to the model training command when receiving a model training command; Extracting a target scene segment from the video data to be processed; performing model training according to the target scene segment.
  • the model training module is further configured to match the target scene information with the scene classification information in the video data index to determine the target scene classification information; the scene corresponding to the target scene classification information The segment is used as the target scene segment, and the data inter-frame tag corresponding to the target scene segment is used as the target data inter-frame tag; extract from the video data to be processed according to the target scene classification information and the target data inter-frame tag Target scene fragment.
  • the model training module is further configured to perform data location positioning according to the target scene classification information and the inter-frame label of the target data, so as to determine the start time and end time of the target scene segment; according to the The start time and the end time extract target scene segments from the video data to be processed.
  • the model training module is further configured to sort the target scene segments according to the start time and end time corresponding to each target scene segment when there are multiple target scene segments, to obtain the sorted target scene segments. Scene fragments; perform model training according to the sorted target scene fragments in turn.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The present application relates to the technical field of data processing. Disclosed are a data indexing method, apparatus and device, and a storage medium. The method comprises: performing scene classification on video data to be processed, so as to obtain a plurality of scene segments; acquiring segment information corresponding to the scene segments; and according to the segment information, generating data inter-frame labels and scene classification information, which correspond to the scene segments, and constructing video data indexes corresponding to the scene segments in said video data. Therefore, there is no need to separately capture scene segments from video data to be processed and to name and store same; it is only necessary to generate corresponding video data indexes according to data inter-frame labels and scene classification information, which correspond to the scene segments, and to store the video data indexes; and the position, in a video to be processed, of a scene segment needing to be used can be found according to the video data indexes, such that the time of video data processing is reduced, and a large amount of manpower and material resources are prevented from being consumed, thereby improving the efficiency of video data processing.

Description

数据索引方法、装置、设备及存储介质Data indexing method, device, equipment and storage medium 技术领域technical field
本申请涉及数据处理技术领域,尤其涉及一种数据索引方法、装置、设备及存储介质。The present application relates to the technical field of data processing, and in particular to a data indexing method, device, equipment and storage medium.
背景技术Background technique
随着自动驾驶的发展及技术迭代,高级别自动驾驶车辆的量产越来越近,深度学习的引入大大加速了自动驾驶技术的落地和应用。但深度学习模型训练需要大量的数据进行训练,当前公开的自动驾驶数据集如coco、kitti等,但公开数据集由一定的局限性,所以当前绝大数的主机厂或者自动驾驶系统解决商都在基于自己的算法模型等建立自己的数据库或者搭建数据采集系统进行数据采集更新,进而进行算法训练。With the development and technical iteration of autonomous driving, the mass production of high-level autonomous vehicles is getting closer and closer, and the introduction of deep learning has greatly accelerated the implementation and application of autonomous driving technology. However, deep learning model training requires a large amount of data for training. The current public autonomous driving data sets such as coco, kitti, etc., but the public data sets have certain limitations, so most of the current OEMs or automatic driving system solution providers are in Establish your own database based on your own algorithm model, or build a data acquisition system for data acquisition and update, and then perform algorithm training.
目前,当采集数据后、算法训练之前一般都需要进行数据处理,基于场景的数据库传统的处理方式是将一段场景单独截取,命名并存储,这种方式需要的时间冗长,耗费大量的人力物力。At present, after data collection and before algorithm training, data processing is generally required. The traditional processing method of scene-based databases is to intercept a scene separately, name it and store it. This method takes a long time and consumes a lot of manpower and material resources.
上述内容仅用于辅助理解本申请的技术方案,并不代表承认上述内容是现有技术。The above content is only used to assist in understanding the technical solution of the present application, and does not mean that the above content is admitted as prior art.
技术问题technical problem
本申请的主要目的在于提出一种数据索引方法、装置、设备及存储介质,旨在解决现有技术中将视频数据中的场景单独截取,命名并存储,需要的时间冗长,耗费大量的人力物力的技术问题。The main purpose of this application is to propose a data indexing method, device, device, and storage medium, aiming at solving the problem of separately intercepting, naming, and storing scenes in video data in the prior art, which takes a long time and consumes a lot of manpower and material resources technical issues.
技术解决方案technical solution
为实现上述目的,本申请提供一种数据索引方法,所述数据索引方法包括以下步骤:In order to achieve the above purpose, the present application provides a data indexing method, the data indexing method includes the following steps:
对待处理视频数据进行场景分类,获得多个场景片段;Perform scene classification on the video data to be processed to obtain multiple scene fragments;
获取各场景片段对应的片段信息;Obtain the fragment information corresponding to each scene fragment;
根据所述片段信息生成各场景片段对应的数据帧间标签和场景分类信息;Generate inter-frame data tags and scene classification information corresponding to each scene segment according to the segment information;
根据所述数据帧间标签和所述场景分类信息构建所述待处理视频数据中各场景片段对应的视频数据索引。A video data index corresponding to each scene segment in the video data to be processed is constructed according to the data inter-frame tag and the scene classification information.
在一实施例中,所述根据所述片段信息生成各场景片段对应的数据帧间标签和场景分类信息,包括:In one embodiment, the generation of data inter-frame labels and scene classification information corresponding to each scene segment according to the segment information includes:
根据所述片段信息确定各场景片段对应的数据帧信息和片段属性信息;Determining data frame information and segment attribute information corresponding to each scene segment according to the segment information;
根据所述数据帧信息生成各场景片段对应的数据帧间标签,并根据所述片段属性信息生成各场景片段对应的场景分类信息。Generate data inter-frame tags corresponding to each scene segment according to the data frame information, and generate scene classification information corresponding to each scene segment according to the segment attribute information.
在一实施例中,所述根据所述数据帧信息生成各场景片段对应的数据帧间标签,包括:In an embodiment, the generating the data inter-frame tags corresponding to each scene segment according to the data frame information includes:
根据所述数据帧信息确定各场景片段对应的帧头信息和帧尾信息;determining frame header information and frame tail information corresponding to each scene segment according to the data frame information;
根据所述帧头信息和所述帧尾信息生成各场景片段对应的数据帧间标签。Generate data inter-frame tags corresponding to each scene segment according to the frame header information and the frame trailer information.
在一实施例中,所述根据所述数据帧间标签和所述场景分类信息构建所述待处理视频数据中各场景片段对应的视频数据索引之后,还包括:In an embodiment, after constructing the video data index corresponding to each scene segment in the video data to be processed according to the data inter-frame tag and the scene classification information, further comprising:
在接收到模型训练指令时,根据所述模型训练指令确定目标场景信息;When receiving a model training instruction, determine target scene information according to the model training instruction;
根据所述目标场景信息和所述视频数据索引从所述待处理视频数据中提取目标场景片段;extracting target scene segments from the video data to be processed according to the target scene information and the video data index;
根据所述目标场景片段进行模型训练。Model training is performed according to the target scene segment.
在一实施例中,所述根据所述目标场景信息和所述视频数据索引从所述待处理视频数据中提取目标场景片段,包括:In an embodiment, the extracting the target scene segment from the video data to be processed according to the target scene information and the video data index includes:
将所述目标场景信息与视频数据索引中的场景分类信息进行匹配,以确定目标场景分类信息;Matching the target scene information with the scene classification information in the video data index to determine the target scene classification information;
将所述目标场景分类信息对应的场景片段作为目标场景片段,并将所述目标场景片段对应的数据帧间标签作为目标数据帧间标签;Using the scene segment corresponding to the target scene classification information as the target scene segment, and using the data inter-frame tag corresponding to the target scene segment as the target data inter-frame tag;
根据所述目标场景分类信息和所述目标数据帧间标签从所述待处理视频数据中提取目标场景片段。Extracting target scene segments from the video data to be processed according to the target scene classification information and the target data inter-frame tags.
在一实施例中,所述根据所述目标场景分类信息和所述目标数据帧间标签从所述待处理视频数据中提取目标场景片段,包括:In an embodiment, the extracting target scene segments from the video data to be processed according to the target scene classification information and the target data inter-frame tags includes:
根据所述目标场景分类信息和所述目标数据帧间标签进行数据位置定位,以确定目标场景片段的开始时间和结束时间;performing data location positioning according to the target scene classification information and the target data inter-frame label to determine the start time and end time of the target scene segment;
根据所述开始时间和所述结束时间从所述待处理视频数据中提取目标场景片段。Extracting target scene segments from the video data to be processed according to the start time and the end time.
在一实施例中,所述根据所述目标场景片段进行模型训练,包括:In one embodiment, the performing model training according to the target scene segment includes:
在所述目标场景片段为多个时,根据各目标场景片段对应的开始时间和结束时间对目标场景片段进行排序,得到排序后的目标场景片段;When there are multiple target scene segments, the target scene segments are sorted according to the corresponding start time and end time of each target scene segment, and the sorted target scene segments are obtained;
依次根据排序后的目标场景片段进行模型训练。Model training is performed sequentially according to the sorted target scene segments.
此外,为实现上述目的,本申请还提出一种数据索引装置,所述数据索引装置包括:In addition, in order to achieve the above purpose, the present application also proposes a data indexing device, the data indexing device includes:
场景分类模块,用于对待处理视频数据进行场景分类,获得多个场景片段;The scene classification module is used to classify the scene of the video data to be processed, and obtain a plurality of scene fragments;
信息获取模块,用于获取各场景片段对应的片段信息;An information acquisition module, configured to acquire segment information corresponding to each scene segment;
数据生成模块,用于根据所述片段信息生成各场景片段对应的数据帧间标签和场景分类信息;A data generation module, configured to generate inter-frame labels and scene classification information corresponding to each scene segment according to the segment information;
数据索引模块,用于根据所述数据帧间标签和所述场景分类信息构建所述待处理视频数据中各场景片段对应的视频数据索引。A data index module, configured to construct a video data index corresponding to each scene segment in the video data to be processed according to the data inter-frame tags and the scene classification information.
此外,为实现上述目的,本申请还提出一种数据索引设备,所述数据索引设备包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的数据索引程序,所述数据索引程序被处理器执行时实现如上所述的数据索引方法。In addition, in order to achieve the above purpose, the present application also proposes a data indexing device, the data indexing device includes: a memory, a processor, and a data indexing program stored in the memory and operable on the processor, the When the data indexing program is executed by the processor, the above-mentioned data indexing method is realized.
此外,为实现上述目的,本申请还提出一种存储介质,所述存储介质上存储有数据索引程序,所述数据索引程序被处理器执行时实现如上所述的数据索引方法。In addition, in order to achieve the above purpose, the present application also proposes a storage medium, on which a data index program is stored, and when the data index program is executed by a processor, the above data index method is implemented.
有益效果Beneficial effect
本申请提出的数据索引方法,对待处理视频数据进行场景分类,获得多个场景片段;获取各场景片段对应的片段信息;根据所述片段信息生成各场景片段对应的数据帧间标签和场景分类信息;根据所述数据帧间标签和所述场景分类信息构建所述待处理视频数据中各场景片段对应的视频数据索引。从而不需要将待处理视频数据中的场景片段单独截取出来、命名并存储,而是只需要根据各场景片段对应的数据帧间标签和场景分类信息生成对应的视频数据索引,对视频数据索引进行存储即可,根据视频数据索引便可查找到需要使用的场景片段在待处理视频中的位置,节约了视频数据处理的时间,也避免了耗费大量的人力物力,进而提高了视频数据处理的效率。The data indexing method proposed in this application performs scene classification on the video data to be processed to obtain multiple scene segments; obtains segment information corresponding to each scene segment; generates data inter-frame labels and scene classification information corresponding to each scene segment according to the segment information ; Constructing a video data index corresponding to each scene segment in the video data to be processed according to the data inter-frame label and the scene classification information. Therefore, it is not necessary to separately intercept, name and store the scene fragments in the video data to be processed, but only need to generate the corresponding video data index according to the data inter-frame labels and scene classification information corresponding to each scene fragment, and carry out the video data index It only needs to be stored, and according to the video data index, the position of the scene segment to be used in the video to be processed can be found, which saves the time of video data processing and avoids the consumption of a lot of manpower and material resources, thereby improving the efficiency of video data processing .
附图说明Description of drawings
图1是本申请实施例方案涉及的硬件运行环境的数据索引设备结构示意图;FIG. 1 is a schematic structural diagram of a data indexing device in a hardware operating environment involved in the solution of the embodiment of the present application;
图2为本申请数据索引方法第一实施例的流程示意图;FIG. 2 is a schematic flow chart of the first embodiment of the data indexing method of the present application;
图3为本申请数据索引方法一实施例的行驶场景片段示意图;FIG. 3 is a schematic diagram of a driving scene segment of an embodiment of the data indexing method of the present application;
图4为本申请数据索引方法一实施例的天气场景片段示意图;FIG. 4 is a schematic diagram of weather scene fragments in an embodiment of the data indexing method of the present application;
图5为本申请数据索引方法一实施例的行驶场景与天气场景结合示意图;FIG. 5 is a schematic diagram of a combination of a driving scene and a weather scene according to an embodiment of the data indexing method of the present application;
图6为本申请数据索引方法第二实施例的流程示意图;FIG. 6 is a schematic flow chart of the second embodiment of the data indexing method of the present application;
图7为本申请数据索引方法一实施例的场景片段的数据帧示意图;FIG. 7 is a schematic diagram of a data frame of a scene segment in an embodiment of the data indexing method of the present application;
图8为本申请数据索引方法第三实施例的流程示意图;FIG. 8 is a schematic flowchart of a third embodiment of the data indexing method of the present application;
图9为本申请数据索引装置第一实施例的功能模块示意图。FIG. 9 is a schematic diagram of functional modules of the first embodiment of the data indexing device of the present application.
本申请目的的实现、功能特点及优点将结合实施例,参照附图做进一步说明。The realization, functional features and advantages of the present application will be further described in conjunction with the embodiments and with reference to the accompanying drawings.
本申请的实施方式Embodiment of this application
应当理解,此处所描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。It should be understood that the specific embodiments described here are only used to explain the present application, and are not intended to limit the present application.
参照图1,图1为本申请实施例方案涉及的硬件运行环境的数据索引设备结构示意图。Referring to FIG. 1 , FIG. 1 is a schematic structural diagram of a data indexing device in a hardware operating environment involved in an embodiment of the present application.
如图1所示,该数据索引设备可以包括:处理器1001,例如中央处理器(Central Processing Unit,CPU),通信总线1002、用户接口1003,网络接口1004,存储器1005。其中,通信总线1002用于实现这些组件之间的连接通信。用户接口1003可以包括显示屏(Display)、输入单元比如按键,可选用户接口1003还可以包括标准的有线接口、无线接口。网络接口1004可选的可以包括标准的有线接口、无线接口(如Wi-Fi接口)。存储器1005可以是高速随机存取存储器(Random Access Memory,RAM),也可以是稳定的存储器(non-volatile memory),例如磁盘存储器。存储器1005可选的还可以是独立于前述处理器1001的存储装置。As shown in FIG. 1 , the data indexing device may include: a processor 1001 , such as a central processing unit (Central Processing Unit, CPU), a communication bus 1002 , a user interface 1003 , a network interface 1004 , and a memory 1005 . Wherein, the communication bus 1002 is used to realize connection and communication between these components. The user interface 1003 may include a display screen (Display) and an input unit such as a button, and the optional user interface 1003 may also include a standard wired interface and a wireless interface. Optionally, the network interface 1004 may include a standard wired interface and a wireless interface (such as a Wi-Fi interface). The memory 1005 may be a high-speed random access memory (Random Access Memory, RAM), or a stable memory (non-volatile memory), such as a disk memory. Optionally, the memory 1005 may also be a storage device independent of the aforementioned processor 1001 .
本领域技术人员可以理解,图1中示出的设备结构并不构成对数据索引设备的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置。Those skilled in the art can understand that the device structure shown in FIG. 1 does not constitute a limitation on the data indexing device, and may include more or less components than those shown in the figure, or combine some components, or arrange different components.
如图1所示,作为一种存储介质的存储器1005中可以包括操作系统、网络通信模块、用户接口模块以及数据索引程序。As shown in FIG. 1 , memory 1005 as a storage medium may include an operating system, a network communication module, a user interface module, and a data index program.
在图1所示的数据索引设备中,网络接口1004主要用于连接外网,与其他网络设备进行数据通信;用户接口1003主要用于连接用户设备,与所述用户设备进行数据通信;本申请设备通过处理器1001调用存储器1005中存储的数据索引程序,并执行本申请实施例提供的数据索引方法。In the data indexing device shown in Figure 1, the network interface 1004 is mainly used to connect to the external network and perform data communication with other network devices; the user interface 1003 is mainly used to connect to user equipment and perform data communication with the user equipment; this application The device invokes the data indexing program stored in the memory 1005 through the processor 1001, and executes the data indexing method provided in the embodiment of the present application.
基于上述硬件结构,提出本申请数据索引方法实施例。Based on the above hardware structure, an embodiment of the data indexing method of the present application is proposed.
参照图2,图2为本申请数据索引方法第一实施例的流程示意图。Referring to FIG. 2 , FIG. 2 is a schematic flowchart of the first embodiment of the data indexing method of the present application.
在第一实施例中,所述数据索引方法包括以下步骤:In the first embodiment, the data indexing method includes the following steps:
步骤S10,对待处理视频数据进行场景分类,获得多个场景片段。Step S10, performing scene classification on the video data to be processed to obtain a plurality of scene fragments.
需要说明的是,本实施例的执行主体可为数据索引设备,例如具有数据处理功能的计算机设备,还可为其他可实现相同或相似功能的设备,本实施例对此不作限制。It should be noted that the executor of this embodiment may be a data indexing device, such as a computer device with data processing functions, or other devices capable of realizing the same or similar functions, which is not limited in this embodiment.
需要说明的是,本实施例中的待处理视频数据可为测试车辆在行驶过程中对环境数据进行采集得到的视频数据,可通过测试车辆上的车载摄像设备进行自动采集,也可通过设置在测试车辆上的外部摄像设备进行自动采集,还可通过其他方式对环境数据进行自动采集,以得到待处理视频数据,本实施例对此不作限制。It should be noted that the video data to be processed in this embodiment can be the video data obtained by collecting the environmental data during the driving process of the test vehicle, which can be automatically collected by the vehicle-mounted camera equipment on the test vehicle, or can be collected by setting the The external camera equipment on the test vehicle can automatically collect the environmental data, and other methods can also be used to automatically collect the environmental data to obtain the video data to be processed, which is not limited in this embodiment.
在具体实现中,例如,测试车辆行驶了6个小时,在测试车辆行驶的过程中自动对环境数据进行了采集,得到6个小时的视频数据,在数据采集结束后需要对采集的视频数据进行处理,便可将该视频数据作为待处理视频数据。In the specific implementation, for example, the test vehicle has been driving for 6 hours, and the environmental data has been collected automatically during the driving process of the test vehicle to obtain 6 hours of video data. After the data collection, the collected video data needs to be processed. processing, the video data can be used as video data to be processed.
应当理解的是,在获取待处理视频数据之后,可对待处理视频数据进行场景分类,获得多个场景片段。需要说明的是,在本方案的该步骤中并不需要从待处理视频数据中截取出这些场景片段,而是只需要在待处理视频数据中识别出这些场景片段即可。具体地,根据预设分类类型对待处理视频数据进行场景分类,也可通过其他方式进行场景分类,本实施例对此不作限制。It should be understood that after the video data to be processed is acquired, scene classification may be performed on the video data to be processed to obtain multiple scene segments. It should be noted that, in this step of the solution, it is not necessary to intercept these scene fragments from the video data to be processed, but only need to identify these scene fragments in the video data to be processed. Specifically, scene classification is performed on the video data to be processed according to a preset classification type, and scene classification may also be performed in other ways, which is not limited in this embodiment.
在具体实现中,预设分类类型可根据实际使用需求预先进行设置,例如,如果以天气场景进行区分,可划分为晴天、雨天、雾天等,进一步地,可将晴天细分为艳阳、多云、少云等,可将雨天细分为大雨、中雨、小雨等,可将雾天细分为等浓雾、中雾、轻雾等。如果以行驶场景进行区分,可划分为巡航、变道、刹车等,进一步地,可将巡航细分为高速巡航、中速巡航、低速巡航等,可将变道细分为向左变道、向右变道、连续变道等,可将刹车细分为紧急刹车、缓慢刹车、点刹等。应当理解的是,除了上述分类类型外,还可包括其他更多的场景分类类型,例如时间场景、地形场景等,本实施例对此不作限制,在本实施例中,以预设分类类型包括天气场景和行驶场景为例进行说明。In the specific implementation, the preset classification type can be set in advance according to the actual use requirements. For example, if the weather scene is used to distinguish, it can be divided into sunny days, rainy days, foggy days, etc., and further, sunny days can be subdivided into sunny days and cloudy days. , few clouds, etc., rainy days can be subdivided into heavy rain, moderate rain, light rain, etc., and foggy days can be subdivided into equal dense fog, medium fog, light fog, etc. If the driving scene is distinguished, it can be divided into cruising, lane changing, braking, etc. Further, cruising can be subdivided into high-speed cruising, medium-speed cruising, low-speed cruising, etc. Lane changing can be subdivided into left lane changing, Right lane change, continuous lane change, etc. Braking can be subdivided into emergency braking, slow braking, point braking, etc. It should be understood that, in addition to the above classification types, other more scene classification types may also be included, such as time scenes, terrain scenes, etc., which are not limited in this embodiment. In this embodiment, the preset classification types include The weather scene and the driving scene are taken as examples for illustration.
可以理解的是,可对待处理视频数据进行图像分析,可根据图像分析结果获得待处理视频数据中的多个场景片段。It can be understood that image analysis can be performed on the video data to be processed, and multiple scene segments in the video data to be processed can be obtained according to the image analysis results.
在具体实现中,例如,如果以行驶场景进行分类,可如图3所示,图3为行驶场景片段示意图,图3中的横线为待处理视频数据的时间轴,可识别出该待处理视频数据中的场景片段有变道场景片段1、巡航场景片段1、变道场景片段2。In a specific implementation, for example, if the driving scene is used for classification, as shown in Figure 3, Figure 3 is a schematic diagram of a driving scene segment, the horizontal line in Figure 3 is the time axis of the video data to be processed, and the video data to be processed can be identified The scene segments in the video data include a lane change scene segment 1 , a cruising scene segment 1 , and a lane change scene segment 2 .
又例如,如果以天气场景进行分类,可如图4所示,图4为天气场景片段示意图,图4中的横线为待处理视频数据的时间轴,可识别出该待处理视频数据中的场景片段有晴天场景片段1、雨天场景片段1。For another example, if the weather scene is used to classify, as shown in Figure 4, Figure 4 is a schematic diagram of a weather scene segment, the horizontal line in Figure 4 is the time axis of the video data to be processed, and the video data to be processed can be identified The scene segments include a sunny scene segment 1 and a rainy day scene segment 1 .
又例如,如果将行驶场景与天气场景结合起来进行分类,可如图5所示,图5为行驶场景与天气场景结合示意图,图5中的横线为待处理视频数据的时间轴,除了可识别出上述行驶场景和天气场景外,还可将行驶场景与天气场景进行结合,可得到图5中的A1、A2、A3、A4四个场景片段,其中,A1为晴天变道场景片段1,A2为晴天巡航场景片段1,A3为雨天巡航场景片段1,A4为雨天变道场景片段1。For another example, if the driving scene and the weather scene are combined for classification, it can be shown in Figure 5, which is a schematic diagram of the combination of the driving scene and the weather scene, and the horizontal line in Figure 5 is the time axis of the video data to be processed. In addition to identifying the above-mentioned driving scene and weather scene, the driving scene and the weather scene can also be combined to obtain four scene fragments A1, A2, A3, and A4 in Figure 5, where A1 is the lane change scene fragment 1 on a sunny day, A2 is the cruising scene segment 1 in sunny days, A3 is the cruising scene segment 1 in rainy days, and A4 is the lane changing scene segment 1 in rainy days.
可以理解的是,除了上述举例说明的独立场景分类方式以及结合场景分类方式外,还可通过其他独立场景进行分类,或者结合更多种类的场景进行分类,可根据实际使用需求进行设置,本实施例对此不作限制。It can be understood that, in addition to the independent scene classification method and combined scene classification method illustrated above, other independent scene classification methods can also be used for classification, or more types of scene classification can be combined, which can be set according to actual use requirements. This implementation Examples are not limited to this.
步骤S20,获取各场景片段对应的片段信息。Step S20, acquiring segment information corresponding to each scene segment.
应当理解的是,由于视频数据中包含很多种类的信息,而场景片段是视频数据中的一部分数据,因此,场景片段中也同样包含很多种类的信息。所以,在确定待处理视频数据中包含的多个场景片段后,可获取这些场景片段对应的片段信息,其中,片段信息可包括数据帧信息和片段属性信息,除此之外,还可包括其他类型的信息,本实施例对此不作限制。It should be understood that since video data contains many types of information, and scene segments are part of the video data, scene segments also contain many types of information. Therefore, after determining a plurality of scene fragments contained in the video data to be processed, the fragment information corresponding to these scene fragments can be obtained, wherein the fragment information can include data frame information and fragment attribute information, besides, it can also include other Type information, which is not limited in this embodiment.
步骤S30,根据所述片段信息生成各场景片段对应的数据帧间标签和场景分类信息。Step S30, generating data inter-frame tags and scene classification information corresponding to each scene segment according to the segment information.
可以理解的是,可根据片段信息对各场景片段分别进行帧间标注,以确定各场景片段对应的开始帧和结束帧,进而生成数据帧间标签。还可根据片段信息中包含的场景类别、时长、存储位置等属性信息,根据这些属性信息生成场景分类信息。It can be understood that the inter-frame labeling can be performed on each scene segment according to the segment information, so as to determine the start frame and the end frame corresponding to each scene segment, and then generate data inter-frame tags. Scene classification information may also be generated according to attribute information such as scene category, duration, and storage location included in the fragment information.
步骤S40,根据所述数据帧间标签和所述场景分类信息构建所述待处理视频数据中各场景片段对应的视频数据索引。Step S40, constructing a video data index corresponding to each scene segment in the video data to be processed according to the inter-frame tag of the data and the scene classification information.
应当理解的是,在生成各场景片段对应的数据帧间标签和场景分类信息后,可根据数据帧间标签和场景分类信息构建待处理视频数据中各场景片段对应的视频数据索引。可将这些视频数据索引进行存储,并且直接将完整的待处理视频数据进行存储,不需要对待处理视频数据进行单独截取、命名存储等耗时的操作,在需要使用待处理视频数据中的场景片段时,直接根据视频数据索引便可确定需要使用的场景片段的相关信息,进而从待处理视频数据中提取出这些场景片段进行使用。It should be understood that, after the data inter-frame tags and scene classification information corresponding to each scene segment are generated, a video data index corresponding to each scene segment in the video data to be processed may be constructed according to the data inter-frame tag and scene classification information. These video data indexes can be stored, and the complete video data to be processed can be stored directly, without time-consuming operations such as separate interception, naming and storage of the video data to be processed, and scene fragments in the video data to be processed can be used , the relevant information of the scene segments to be used can be determined directly according to the video data index, and then these scene segments are extracted from the video data to be processed for use.
可以理解的是,可通过数据库或者表格的方式对视频数据索引进行存储,或以其它编程语言方式实现对视频数据索引进行存储,例如,可将这些视频数据索引存储在EXCLE表格中,除了上述方式外,还可通过其他方式对视频数据索引进行存储,本实施例对此不作限制,在本实施例中,以通过表格对视频数据索引进行存储为例进行说明。由于视频数据索引相较于分割的视频数据而言,需要的存储空间较小,便于存储,而且不需要进行复杂的管理,通过表格就可以很清晰的对这些信息进行记录,可减少数据处理的时间,并且避免耗费大量的人力物力。It can be understood that the video data index can be stored in the form of a database or a table, or can be stored in other programming languages. For example, these video data indexes can be stored in an EXCLE table. In addition, the video data index may also be stored in other ways, which is not limited in this embodiment. In this embodiment, storage of the video data index in a table is taken as an example for illustration. Compared with segmented video data, the video data index requires less storage space, is easy to store, and does not require complicated management. The information can be clearly recorded through the table, which can reduce the cost of data processing. time and avoid wasting a lot of manpower and material resources.
在具体实现中,可将待处理视频数据中的所有场景片段对应的视频数据索引记录在同一个表格中进行存储,也可将不同类型的场景片段对应的视频数据索引记录在不同的表格中进行存储,例如,将天气场景对应的场景片段对应的视频数据索引记录在一个表格中进行存储,将行驶场景对应的场景片段对应的视频数据索引记录在另一个表格中进行存储。又例如,还可更加细化,将晴天场景对应的场景片段对应的视频数据索引记录在一个表格中进行存储,将雨天场景对应的对应的场景片段对应的视频数据索引记录在另一个表格中进行存储。除了上述方式外,还可根据实际使用需求通过其他存储方式进行存储,本实施例对此不作限制。In a specific implementation, the video data indexes corresponding to all scene segments in the video data to be processed can be recorded in the same table for storage, and the video data indexes corresponding to different types of scene segments can also be recorded in different tables for storage. For storage, for example, the video data index corresponding to the scene segment corresponding to the weather scene is recorded in a table for storage, and the video data index corresponding to the scene segment corresponding to the driving scene is recorded in another table for storage. For another example, it can also be further refined, and the video data index corresponding to the scene segment corresponding to the sunny scene is recorded in a table for storage, and the video data index corresponding to the corresponding scene segment corresponding to the rainy scene is recorded in another table. storage. In addition to the above methods, other storage methods may also be used for storage according to actual usage requirements, which is not limited in this embodiment.
在本实施例中,对待处理视频数据进行场景分类,获得多个场景片段;获取各场景片段对应的片段信息;根据所述片段信息生成各场景片段对应的数据帧间标签和场景分类信息;根据所述数据帧间标签和所述场景分类信息构建所述待处理视频数据中各场景片段对应的视频数据索引。从而不需要将待处理视频数据中的场景片段单独截取出来、命名并存储,而是只需要根据各场景片段对应的数据帧间标签和场景分类信息生成对应的视频数据索引,对视频数据索引进行存储即可,根据视频数据索引便可查找到需要使用的场景片段在待处理视频中的位置,节约了视频数据处理的时间,也避免了耗费大量的人力物力,进而提高了视频数据处理的效率。In this embodiment, scene classification is performed on the video data to be processed to obtain a plurality of scene segments; segment information corresponding to each scene segment is obtained; data inter-frame tags and scene classification information corresponding to each scene segment are generated according to the segment information; The data inter-frame tags and the scene classification information construct a video data index corresponding to each scene segment in the video data to be processed. Therefore, it is not necessary to separately intercept, name and store the scene fragments in the video data to be processed, but only need to generate the corresponding video data index according to the data inter-frame labels and scene classification information corresponding to each scene fragment, and carry out the video data index It only needs to be stored, and according to the video data index, the position of the scene segment to be used in the video to be processed can be found, which saves the time of video data processing and avoids the consumption of a lot of manpower and material resources, thereby improving the efficiency of video data processing .
在一实施例中,如图6所示,基于第一实施例提出本申请数据索引方法第二实施例,所述步骤S30,包括:In one embodiment, as shown in FIG. 6 , based on the first embodiment, a second embodiment of the data indexing method of the present application is proposed. The step S30 includes:
步骤S301,根据所述片段信息确定各场景片段对应的数据帧信息和片段属性信息。Step S301: Determine the data frame information and segment attribute information corresponding to each scene segment according to the segment information.
应当理解的是,视频数据可以由多帧图像数据组成,本实施例中的数据帧信息指的是与各场景片段相关的视频帧信息。进而可根据数据帧信息确定各场景片段对应的帧头信息和帧尾信息,其中,帧头信息指的是场景片段开始时的视频帧信息,帧尾信息指的是场景片段结束时的视频帧信息。在确定各场景片段对应的帧头信息和帧尾信息后,便可根据帧头信息和帧尾信息生成场景片段对应的数据帧信息。It should be understood that video data may consist of multiple frames of image data, and the data frame information in this embodiment refers to video frame information related to each scene segment. Furthermore, the frame header information and frame tail information corresponding to each scene segment can be determined according to the data frame information, wherein the frame header information refers to the video frame information at the beginning of the scene segment, and the frame tail information refers to the video frame at the end of the scene segment information. After the frame header information and frame trailer information corresponding to each scene segment are determined, data frame information corresponding to the scene segment can be generated according to the frame header information and frame trailer information.
在具体实现中,可在天气场景分类的基础上进行说明,可如图7所示,图7为场景片段的数据帧示意图,在图7中,变道场景片段1的帧头为O1,帧尾为O2,即在待处理视频中的第O1帧开始至第O2帧结束之间的视频数据都是变道场景片段1对应的视频数据;巡航场景片段1的帧头为O2,帧尾为O3,即在待处理视频中的第O2帧开始至第O3帧结束之间的视频数据都是巡航场景片段1对应的视频数据;变道场景片段2的帧头为O3,帧尾为O4,即在待处理视频中的第O3帧开始至第O4帧结束之间的视频数据都是变道场景片段2对应的视频数据。In the specific implementation, it can be described on the basis of weather scene classification, as shown in Figure 7, which is a schematic diagram of the data frame of the scene segment, in Figure 7, the frame header of the lane change scene segment 1 is O1, and the frame The tail is O2, that is, the video data between the beginning of the O1 frame and the end of the O2 frame in the video to be processed is the video data corresponding to the lane change scene segment 1; the frame header of the cruising scene segment 1 is O2, and the frame tail is O3, that is, the video data between the beginning of the O2 frame and the end of the O3 frame in the video to be processed is the video data corresponding to the cruise scene segment 1; the frame header of the lane change scene segment 2 is O3, and the frame tail is O4, That is, the video data between the beginning of the O3th frame and the end of the O4th frame in the video to be processed is the video data corresponding to the lane change scene segment 2.
步骤S302,根据所述数据帧信息生成各场景片段对应的数据帧间标签,并根据所述片段属性信息生成各场景片段对应的场景分类信息。Step S302, generating data inter-frame tags corresponding to each scene segment according to the data frame information, and generating scene classification information corresponding to each scene segment according to the segment attribute information.
应当理解的是,可根据数据帧信息对待处理视频数据中的各场景片段进行帧间标注,生成各场景片段对应的数据帧间标签,根据场景片段对应的数据帧间标签便可确定该场景片段在待处理视频数据中所处的位置。It should be understood that inter-frame labeling can be performed on each scene segment in the video data to be processed according to the data frame information to generate a data inter-frame tag corresponding to each scene segment, and the scene segment can be determined according to the data inter-frame tag corresponding to the scene segment The position in the video data to be processed.
应当理解的是,由于片段属性信息中包含有场景类别、时长、存储位置等属性,因此,可根据片段属性信息生成各场景片段对应的场景分类信息,根据场景片段对应的场景分类信息便可确定该场景片段对应的分类信息等属性。It should be understood that since the segment attribute information includes attributes such as scene category, duration, and storage location, the scene classification information corresponding to each scene segment can be generated according to the segment attribute information, and the scene classification information corresponding to the scene segment can be determined. Attributes such as classification information corresponding to the scene segment.
可以理解的是,由于数据帧间标签和场景分类信息均包含有待处理视频数据中的各场景片段对应的部分信息,前者用于定位,后者用于分类,因此,在需要使用场景片段对应的数据时,可通过数据帧间标签和场景分类信息进行索引和定位,便于数据的提取。It can be understood that since both the data inter-frame label and the scene classification information contain partial information corresponding to each scene segment in the video data to be processed, the former is used for positioning and the latter is used for classification. Therefore, when it is necessary to use the scene segment corresponding to Data can be indexed and located through inter-frame labels and scene classification information to facilitate data extraction.
在本实施例中,根据所述片段信息确定各场景片段对应的数据帧信息和片段属性信息;根据所述数据帧信息生成各场景片段对应的数据帧间标签,并根据所述片段属性信息生成各场景片段对应的场景分类信息。从而可通过待处理视频数据中的各场景片段对应的数据帧间标签和场景分类信息分别进行定位和分类,可以在需要使用待处理视频数据中的场景片段时,方便地从待处理视频数据中提取场景片段对应的数据。In this embodiment, the data frame information and segment attribute information corresponding to each scene segment are determined according to the segment information; the data inter-frame tags corresponding to each scene segment are generated according to the data frame information, and generated according to the segment attribute information Scene classification information corresponding to each scene segment. Thereby, positioning and classification can be carried out respectively through the data inter-frame tags and scene classification information corresponding to each scene segment in the video data to be processed, and when the scene segment in the video data to be processed needs to be used, it can be conveniently obtained from the video data to be processed Data corresponding to the scene clip is extracted.
在一实施例中,如图8所示,基于第一实施例或第二实施例提出本申请数据索引方法第三实施例,在本实施例中,基于第一实施例进行说明,所述步骤S40之后,还包括:In one embodiment, as shown in FIG. 8 , the third embodiment of the data indexing method of this application is proposed based on the first embodiment or the second embodiment. In this embodiment, the description is made based on the first embodiment. The steps After S40, it also includes:
步骤S50,在接收到模型训练指令时,根据所述模型训练指令确定目标场景信息。Step S50, when receiving a model training instruction, determine target scene information according to the model training instruction.
应当理解的是,在需要使用待处理视频数据中的数据对深度学习模型进行训练时,可向计算机设备发送模型训练指令。其中,深度学习模型可分为多种用途以及对应的训练类型,例如,假如是用于天气场景的深度学习模型,其对应的训练类型为天气场景训练,需要通过一些与天气场景相关的数据对深度学习模型进行训练;假如是用于行驶场景的深度学习模型,其对应的训练类型为行驶场景训练,需要通过一些与行驶场景相关的数据对深度学习模型训练。It should be understood that, when the data in the video data to be processed needs to be used to train the deep learning model, a model training instruction can be sent to the computer device. Among them, the deep learning model can be divided into multiple purposes and corresponding training types. For example, if it is a deep learning model used for weather scenes, the corresponding training type is weather scene training, which needs to be paired with some weather scene related data. The deep learning model is trained; if it is a deep learning model for driving scenes, the corresponding training type is driving scene training, and some data related to driving scenes is needed to train the deep learning model.
可以理解的是,计算机设备在接收到模型训练指令时,可根据模型训练指令确定当前所需要的数据对应的目标场景信息。例如,在当前需要天气场景相关的数据时,对应的目标场景信息为天气场景;在当前需要行驶场景相关的数据时,对应的目标场景信息为行驶场景。It can be understood that, when the computer device receives the model training instruction, it can determine the target scene information corresponding to the currently required data according to the model training instruction. For example, when data related to a weather scene is currently needed, the corresponding target scene information is a weather scene; when data related to a driving scene is currently needed, the corresponding target scene information is a driving scene.
步骤S60,根据所述目标场景信息和所述视频数据索引从所述待处理视频数据中提取目标场景片段。Step S60, extracting target scene segments from the video data to be processed according to the target scene information and the video data index.
应当理解的是,由于已经对待处理视频数据中的各场景片段对应的视频数据索引进行了存储,因此,根据存储的视频数据索引便可方便地查找到与目标场景信息对应的目标场景片段,并从待处理视频数据中提取出目标场景片段对应的数据,以用于模型训练。It should be understood that since the video data index corresponding to each scene segment in the video data to be processed has been stored, the target scene segment corresponding to the target scene information can be easily found according to the stored video data index, and The data corresponding to the target scene segment is extracted from the video data to be processed for model training.
可以理解的是,可将目标场景信息与视频数据索引中的场景分类信息进行匹配,并根据匹配结果确定目标场景分类信息。进而可根据视频数据索引将目标场景分类信息对应的场景片段作为目标场景片段,并将目标场景片段对应的数据帧间标签作为目标数据帧间标签,然后便可根据目标场景分类信息和目标数据帧间标签从所述待处理视频数据中提取目标场景片段对应的数据。It can be understood that the target scene information can be matched with the scene classification information in the video data index, and the target scene classification information can be determined according to the matching result. Furthermore, according to the video data index, the scene segment corresponding to the target scene classification information can be used as the target scene segment, and the data inter-frame label corresponding to the target scene segment can be used as the target data inter-frame label, and then according to the target scene classification information and the target data frame The inter tag extracts the data corresponding to the target scene segment from the video data to be processed.
在具体实现中,例如,假如目标场景信息为变道场景,即当前需要使用变道场景的数据对模型进行训练,则可在视频数据索引中匹配与变道场景相关的目标场景分类信息,进而确定对应的目标场景片段为变道场景片段1和变道场景片段2,可进一步确定变道场景片段1对应的数据帧间标签为第O1帧~第O2帧,变道场景片段2对应的数据帧间标签为第O3帧~第O4帧,因此,可从待处理视频数据中提取出这两个变道场景片段对应的数据,用于模型训练。In a specific implementation, for example, if the target scene information is a lane-changing scene, that is, the data of the lane-changing scene needs to be used to train the model, the target scene classification information related to the lane-changing scene can be matched in the video data index, and then Determine the corresponding target scene segments as lane change scene segment 1 and lane change scene segment 2, and further determine that the data inter-frame label corresponding to lane change scene segment 1 is frame O1~O2, and the data corresponding to lane change scene segment 2 The inter-frame labels are frame O3~O4. Therefore, the data corresponding to these two lane-changing scene segments can be extracted from the video data to be processed for model training.
应当理解的是,由于场景分类信息中包含时长以及存储位置等属性,可先根据目标场景分类信息确定待处理视频数据的存储位置。为了便于数据的提取,可根据目标场景分类信息和目标数据帧间标签进行数据位置定位,以确定目标场景片段在待处理视频数据中的开始时间和结束时间,其中,开始时间与目标场景片段的帧头对应,结束时间与目标场景片段的帧尾对应。然后,便可根据目标场景片段对应的开始时间和结束时间从所述待处理视频数据中提取目标场景片段对应的数据。It should be understood that since the scene classification information includes attributes such as duration and storage location, the storage location of the video data to be processed may be first determined according to the target scene classification information. In order to facilitate data extraction, the data location can be positioned according to the target scene classification information and the inter-frame labels of the target data to determine the start time and end time of the target scene segment in the video data to be processed, wherein the start time and the target scene segment's The frame header corresponds, and the end time corresponds to the frame end of the target scene segment. Then, the data corresponding to the target scene segment may be extracted from the video data to be processed according to the start time and the end time corresponding to the target scene segment.
步骤S70,根据所述目标场景片段进行模型训练。Step S70, perform model training according to the target scene segment.
应当理解的是,如果只有一个目标场景片段,则直接从待处理视频数据中提取该目标场景片段对应的数据,以进行模型训练。如果有多个目标场景片段,则可根据各目标场景片段对应的开始时间和结束时间对目标场景片段进行排序,得到排序后的目标场景片段,依次根据排序后的目标场景片段进行模型训练。It should be understood that if there is only one target scene segment, the data corresponding to the target scene segment is directly extracted from the video data to be processed for model training. If there are multiple target scene segments, the target scene segments can be sorted according to the corresponding start time and end time of each target scene segment, and the sorted target scene segments are obtained, and the model training is performed according to the sorted target scene segments in turn.
在具体实现中,例如,假如有变道场景片段1和变道场景片段2这两个目标场景片段,根据排序结果可知,变道场景片段1在变道场景片段2之前,因此,可先通过变道场景片段1对应的数据进行模型训练,然后在训练完成之后,自动进行跳转,再通过变道场景片段2对应的数据进行模型训练。可以理解的是,如果存在更多的目标场景片段,则可通过上述方式继续根据排序结果进行跳转,并且进行模型训练,直至所有的目标场景片段对应的数据都已进行过模型训练为止。In a specific implementation, for example, if there are two target scene segments, the lane change scene segment 1 and the lane change scene segment 2, according to the sorting result, it can be known that the lane change scene segment 1 is before the lane change scene segment 2, therefore, it can be passed through The data corresponding to the lane change scene segment 1 is used for model training, and then after the training is completed, the jump is automatically performed, and then the model training is performed with the data corresponding to the lane change scene segment 2. It can be understood that if there are more target scene fragments, the above method can continue to jump according to the sorting results, and carry out model training until all the data corresponding to the target scene fragments have been trained for the model.
可以理解的是,通过本实施例的方案,可在模型训练时通过视频数据索引自动索引目标场景片段对应的数据,并自动快速链入场景数据通道,在对一个目标场景片段对应的数据训练结束时,可以自动切换至下一个目标场景片段对应的数据,持续进行训练,从而达到更好的对模型进行自动训练的效果。It can be understood that, through the solution of this embodiment, the data corresponding to the target scene segment can be automatically indexed through the video data index during model training, and automatically and quickly linked into the scene data channel. After the training of the data corresponding to a target scene segment When , it can automatically switch to the data corresponding to the next target scene segment, and continue training, so as to achieve a better effect of automatic training of the model.
在本实施例中,在接收到模型训练指令时,根据所述模型训练指令确定目标场景信息;根据所述目标场景信息和所述视频数据索引从所述待处理视频数据中提取目标场景片段;根据所述目标场景片段进行模型训练。在本方案中,不需要对原始视频数据进行分割存储,在使用时也不需要去查找这些分割存储的视频数据,而是通过目标场景信息来进行场景匹配,并根据视频数据索引来定位当前需要的目标场景片段,并从原始的待处理视频数据中提取出目标场景片段对应的数据进行模型训练,通过索引和快速定位的方式,提高了模型训练的效率,也节约了人力成本。In this embodiment, when a model training instruction is received, target scene information is determined according to the model training instruction; target scene segments are extracted from the video data to be processed according to the target scene information and the video data index; Model training is performed according to the target scene segment. In this solution, there is no need to divide and store the original video data, and it is not necessary to search for these divided and stored video data when using it. Instead, the target scene information is used to perform scene matching, and the current needs are located according to the video data index. The target scene segment, and extract the data corresponding to the target scene segment from the original video data to be processed for model training. Through indexing and fast positioning, the efficiency of model training is improved, and labor costs are also saved.
此外,本申请实施例还提出一种存储介质,所述存储介质上存储有数据索引程序,所述数据索引程序被处理器执行时实现如上文所述的数据索引方法的步骤。In addition, the embodiment of the present application also proposes a storage medium, on which a data index program is stored, and when the data index program is executed by a processor, the steps of the above-mentioned data index method are implemented.
由于本存储介质采用了上述所有实施例的全部技术方案,因此至少具有上述实施例的技术方案所带来的所有有益效果,在此不再一一赘述。Since the storage medium adopts all the technical solutions of all the above-mentioned embodiments, it at least has all the beneficial effects brought by the technical solutions of the above-mentioned embodiments, which will not be repeated here.
此外,参照图9,本申请实施例还提出一种数据索引装置,所述数据索引装置包括:In addition, referring to FIG. 9 , the embodiment of the present application also proposes a data indexing device, the data indexing device includes:
场景分类模块10,用于对待处理视频数据进行场景分类,获得多个场景片段。The scene classification module 10 is configured to perform scene classification on the video data to be processed to obtain a plurality of scene fragments.
需要说明的是,本实施例中的待处理视频数据可为测试车辆在行驶过程中对环境数据进行采集得到的视频数据,可通过测试车辆上的车载摄像设备进行自动采集,也可通过设置在测试车辆上的外部摄像设备进行自动采集,还可通过其他方式对环境数据进行自动采集,以得到待处理视频数据,本实施例对此不作限制。It should be noted that the video data to be processed in this embodiment can be the video data obtained by collecting the environmental data during the driving process of the test vehicle, which can be automatically collected by the vehicle-mounted camera equipment on the test vehicle, or can be collected by setting the The external camera equipment on the test vehicle can automatically collect the environmental data, and other methods can also be used to automatically collect the environmental data to obtain the video data to be processed, which is not limited in this embodiment.
在具体实现中,例如,测试车辆行驶了6个小时,在测试车辆行驶的过程中自动对环境数据进行了采集,得到6个小时的视频数据,在数据采集结束后需要对采集的视频数据进行处理,便可将该视频数据作为待处理视频数据。In the specific implementation, for example, the test vehicle has been driving for 6 hours, and the environmental data has been collected automatically during the driving process of the test vehicle to obtain 6 hours of video data. After the data collection, the collected video data needs to be processed. processing, the video data can be used as video data to be processed.
应当理解的是,在获取待处理视频数据之后,可对待处理视频数据进行场景分类,获得多个场景片段。需要说明的是,在本方案的该步骤中并不需要从待处理视频数据中截取出这些场景片段,而是只需要在待处理视频数据中识别出这些场景片段即可。具体地,根据预设分类类型对待处理视频数据进行场景分类,也可通过其他方式进行场景分类,本实施例对此不作限制。It should be understood that after the video data to be processed is acquired, scene classification may be performed on the video data to be processed to obtain multiple scene segments. It should be noted that, in this step of the solution, it is not necessary to intercept these scene fragments from the video data to be processed, but only need to identify these scene fragments in the video data to be processed. Specifically, the scene classification is performed on the video data to be processed according to the preset classification type, and the scene classification may also be performed in other ways, which is not limited in this embodiment.
在具体实现中,预设分类类型可根据实际使用需求预先进行设置,例如,如果以天气场景进行区分,可划分为晴天、雨天、雾天等,进一步地,可将晴天细分为艳阳、多云、少云等,可将雨天细分为大雨、中雨、小雨等,可将雾天细分为等浓雾、中雾、轻雾等。如果以行驶场景进行区分,可划分为巡航、变道、刹车等,进一步地,可将巡航细分为高速巡航、中速巡航、低速巡航等,可将变道细分为向左变道、向右变道、连续变道等,可将刹车细分为紧急刹车、缓慢刹车、点刹等。应当理解的是,除了上述分类类型外,还可包括其他更多的场景分类类型,例如时间场景、地形场景等,本实施例对此不作限制,在本实施例中,以预设分类类型包括天气场景和行驶场景为例进行说明。In the specific implementation, the preset classification type can be set in advance according to the actual use requirements. For example, if the weather scene is used to distinguish, it can be divided into sunny days, rainy days, foggy days, etc., and further, sunny days can be subdivided into sunny days and cloudy days. , few clouds, etc., rainy days can be subdivided into heavy rain, moderate rain, light rain, etc., and foggy days can be subdivided into equal dense fog, medium fog, light fog, etc. If the driving scene is distinguished, it can be divided into cruising, lane changing, braking, etc. Further, cruising can be subdivided into high-speed cruising, medium-speed cruising, low-speed cruising, etc. Lane changing can be subdivided into left lane changing, Right lane change, continuous lane change, etc. Braking can be subdivided into emergency braking, slow braking, point braking, etc. It should be understood that, in addition to the above classification types, other more scene classification types may also be included, such as time scenes, terrain scenes, etc., which are not limited in this embodiment. In this embodiment, the preset classification types include The weather scene and the driving scene are taken as examples for illustration.
可以理解的是,可对待处理视频数据进行图像分析,可根据图像分析结果获得待处理视频数据中的多个场景片段。It can be understood that image analysis can be performed on the video data to be processed, and multiple scene segments in the video data to be processed can be obtained according to the image analysis results.
在具体实现中,例如,如果以行驶场景进行分类,可如图3所示,图3为行驶场景片段示意图,图3中的横线为待处理视频数据的时间轴,可识别出该待处理视频数据中的场景片段有变道场景片段1、巡航场景片段1、变道场景片段2。In a specific implementation, for example, if the driving scene is used for classification, as shown in Figure 3, Figure 3 is a schematic diagram of a driving scene segment, the horizontal line in Figure 3 is the time axis of the video data to be processed, and the video data to be processed can be identified The scene segments in the video data include a lane change scene segment 1 , a cruising scene segment 1 , and a lane change scene segment 2 .
又例如,如果以天气场景进行分类,可如图4所示,图4为天气场景片段示意图,图4中的横线为待处理视频数据的时间轴,可识别出该待处理视频数据中的场景片段有晴天场景片段1、雨天场景片段1。For another example, if the weather scene is used to classify, as shown in Figure 4, Figure 4 is a schematic diagram of a weather scene segment, the horizontal line in Figure 4 is the time axis of the video data to be processed, and the video data to be processed can be identified The scene segments include a sunny scene segment 1 and a rainy day scene segment 1 .
又例如,如果将行驶场景与天气场景结合起来进行分类,可如图5所示,图5为行驶场景与天气场景结合示意图,图5中的横线为待处理视频数据的时间轴,除了可识别出上述行驶场景和天气场景外,还可将行驶场景与天气场景进行结合,可得到图5中的A1、A2、A3、A4四个场景片段,其中,A1为晴天变道场景片段1,A2为晴天巡航场景片段1,A3为雨天巡航场景片段1,A4为雨天变道场景片段1。For another example, if the driving scene and the weather scene are combined for classification, it can be shown in Figure 5, which is a schematic diagram of the combination of the driving scene and the weather scene, and the horizontal line in Figure 5 is the time axis of the video data to be processed. In addition to identifying the above-mentioned driving scene and weather scene, the driving scene and the weather scene can also be combined to obtain four scene fragments A1, A2, A3, and A4 in Figure 5, where A1 is the lane change scene fragment 1 on a sunny day, A2 is the cruising scene segment 1 in sunny days, A3 is the cruising scene segment 1 in rainy days, and A4 is the lane changing scene segment 1 in rainy days.
可以理解的是,除了上述举例说明的独立场景分类方式以及结合场景分类方式外,还可通过其他独立场景进行分类,或者结合更多种类的场景进行分类,可根据实际使用需求进行设置,本实施例对此不作限制。It can be understood that, in addition to the independent scene classification method and combined scene classification method illustrated above, other independent scene classification methods can also be used for classification, or more types of scene classification can be combined, which can be set according to actual use requirements. This implementation Examples are not limited to this.
信息获取模块20,用于获取各场景片段对应的片段信息。The information acquisition module 20 is configured to acquire segment information corresponding to each scene segment.
应当理解的是,由于视频数据中包含很多种类的信息,而场景片段是视频数据中的一部分数据,因此,场景片段中也同样包含很多种类的信息。所以,在确定待处理视频数据中包含的多个场景片段后,可获取这些场景片段对应的片段信息,其中,片段信息可包括数据帧信息和片段属性信息,除此之外,还可包括其他类型的信息,本实施例对此不作限制。It should be understood that since video data contains many types of information, and scene segments are part of the video data, scene segments also contain many types of information. Therefore, after determining a plurality of scene fragments contained in the video data to be processed, the fragment information corresponding to these scene fragments can be obtained, wherein the fragment information can include data frame information and fragment attribute information, besides, it can also include other Type information, which is not limited in this embodiment.
数据生成模块30,用于根据所述片段信息生成各场景片段对应的数据帧间标签和场景分类信息。The data generation module 30 is configured to generate data inter-frame tags and scene classification information corresponding to each scene segment according to the segment information.
可以理解的是,可根据片段信息对各场景片段分别进行帧间标注,以确定各场景片段对应的开始帧和结束帧,进而生成数据帧间标签。还可根据片段信息中包含的场景类别、时长、存储位置等属性信息,根据这些属性信息生成场景分类信息。It can be understood that the inter-frame labeling can be performed on each scene segment according to the segment information, so as to determine the start frame and the end frame corresponding to each scene segment, and then generate data inter-frame tags. Scene classification information may also be generated according to attribute information such as scene category, duration, and storage location included in the fragment information.
数据索引模块40,用于根据所述数据帧间标签和所述场景分类信息构建所述待处理视频数据中各场景片段对应的视频数据索引。The data index module 40 is configured to construct a video data index corresponding to each scene segment in the video data to be processed according to the inter-frame tag of the data and the scene classification information.
应当理解的是,在生成各场景片段对应的数据帧间标签和场景分类信息后,可根据数据帧间标签和场景分类信息构建待处理视频数据中各场景片段对应的视频数据索引。可将这些视频数据索引进行存储,并且直接将完整的待处理视频数据进行存储,不需要对待处理视频数据进行单独截取、命名存储等耗时的操作,在需要使用待处理视频数据中的场景片段时,直接根据视频数据索引便可确定需要使用的场景片段的相关信息,进而从待处理视频数据中提取出这些场景片段进行使用。It should be understood that, after the data inter-frame tags and scene classification information corresponding to each scene segment are generated, a video data index corresponding to each scene segment in the video data to be processed may be constructed according to the data inter-frame tag and scene classification information. These video data indexes can be stored, and the complete video data to be processed can be stored directly, without time-consuming operations such as separate interception, naming and storage of the video data to be processed, and scene fragments in the video data to be processed can be used , the relevant information of the scene segments to be used can be determined directly according to the video data index, and then these scene segments are extracted from the video data to be processed for use.
可以理解的是,可通过数据库或者表格的方式对视频数据索引进行存储,或以其它编程语言方式实现对视频数据索引进行存储,例如,可将这些视频数据索引存储在EXCLE表格中,除了上述方式外,还可通过其他方式对视频数据索引进行存储,本实施例对此不作限制,在本实施例中,以通过表格对视频数据索引进行存储为例进行说明。由于视频数据索引相较于分割的视频数据而言,需要的存储空间较小,便于存储,而且不需要进行复杂的管理,通过表格就可以很清晰的对这些信息进行记录,可减少数据处理的时间,并且避免耗费大量的人力物力。It can be understood that the video data index can be stored in the form of a database or a table, or can be stored in other programming languages. For example, these video data indexes can be stored in an EXCLE table. In addition, the video data index may also be stored in other ways, which is not limited in this embodiment. In this embodiment, storage of the video data index in a table is taken as an example for illustration. Compared with segmented video data, the video data index requires less storage space, is easy to store, and does not require complicated management. The information can be clearly recorded through the table, which can reduce the cost of data processing. time and avoid wasting a lot of manpower and material resources.
在具体实现中,可将待处理视频数据中的所有场景片段对应的视频数据索引记录在同一个表格中进行存储,也可将不同类型的场景片段对应的视频数据索引记录在不同的表格中进行存储,例如,将天气场景对应的场景片段对应的视频数据索引记录在一个表格中进行存储,将行驶场景对应的场景片段对应的视频数据索引记录在另一个表格中进行存储。又例如,还可更加细化,将晴天场景对应的场景片段对应的视频数据索引记录在一个表格中进行存储,将雨天场景对应的对应的场景片段对应的视频数据索引记录在另一个表格中进行存储。除了上述方式外,还可根据实际使用需求通过其他存储方式进行存储,本实施例对此不作限制。In a specific implementation, the video data indexes corresponding to all scene segments in the video data to be processed can be recorded in the same table for storage, and the video data indexes corresponding to different types of scene segments can also be recorded in different tables for storage. For storage, for example, the video data index corresponding to the scene segment corresponding to the weather scene is recorded in a table for storage, and the video data index corresponding to the scene segment corresponding to the driving scene is recorded in another table for storage. For another example, it can also be further refined, and the video data index corresponding to the scene segment corresponding to the sunny scene is recorded in a table for storage, and the video data index corresponding to the corresponding scene segment corresponding to the rainy scene is recorded in another table. storage. In addition to the above methods, other storage methods may also be used for storage according to actual usage requirements, which is not limited in this embodiment.
在本实施例中,对待处理视频数据进行场景分类,获得多个场景片段;获取各场景片段对应的片段信息;根据所述片段信息生成各场景片段对应的数据帧间标签和场景分类信息;根据所述数据帧间标签和所述场景分类信息构建所述待处理视频数据中各场景片段对应的视频数据索引。从而不需要将待处理视频数据中的场景片段单独截取出来、命名并存储,而是只需要根据各场景片段对应的数据帧间标签和场景分类信息生成对应的视频数据索引,对视频数据索引进行存储即可,根据视频数据索引便可查找到需要使用的场景片段在待处理视频中的位置,节约了视频数据处理的时间,也避免了耗费大量的人力物力,进而提高了视频数据处理的效率。In this embodiment, scene classification is performed on the video data to be processed to obtain a plurality of scene segments; segment information corresponding to each scene segment is obtained; data inter-frame tags and scene classification information corresponding to each scene segment are generated according to the segment information; The data inter-frame tags and the scene classification information construct a video data index corresponding to each scene segment in the video data to be processed. Therefore, it is not necessary to separately intercept, name and store the scene fragments in the video data to be processed, but only need to generate the corresponding video data index according to the data inter-frame labels and scene classification information corresponding to each scene fragment, and carry out the video data index It only needs to be stored, and according to the video data index, the position of the scene segment to be used in the video to be processed can be found, which saves the time of video data processing and avoids the consumption of a lot of manpower and material resources, thereby improving the efficiency of video data processing .
在一实施例中,所述数据生成模块30,还用于根据所述片段信息确定各场景片段对应的数据帧信息和片段属性信息;根据所述数据帧信息生成各场景片段对应的数据帧间标签,并根据所述片段属性信息生成各场景片段对应的场景分类信息。In one embodiment, the data generation module 30 is further configured to determine the data frame information and segment attribute information corresponding to each scene segment according to the segment information; generate the data frame information corresponding to each scene segment according to the data frame information tags, and generate scene classification information corresponding to each scene segment according to the segment attribute information.
在一实施例中,所述数据生成模块30,还用于根据所述数据帧信息确定各场景片段对应的帧头信息和帧尾信息;根据所述帧头信息和所述帧尾信息生成各场景片段对应的数据帧间标签。In an embodiment, the data generating module 30 is further configured to determine frame header information and frame trailer information corresponding to each scene segment according to the data frame information; generate each frame header information and frame trailer information according to the frame header information and the frame trailer information The inter-frame label of the data corresponding to the scene fragment.
在一实施例中,所述数据索引装置还包括模型训练模块,用于在接收到模型训练指令时,根据所述模型训练指令确定目标场景信息;根据所述目标场景信息和所述视频数据索引从所述待处理视频数据中提取目标场景片段;根据所述目标场景片段进行模型训练。In one embodiment, the data indexing device further includes a model training module, configured to determine target scene information according to the model training command when receiving a model training command; Extracting a target scene segment from the video data to be processed; performing model training according to the target scene segment.
在一实施例中,所述模型训练模块,还用于将所述目标场景信息与视频数据索引中的场景分类信息进行匹配,以确定目标场景分类信息;将所述目标场景分类信息对应的场景片段作为目标场景片段,并将所述目标场景片段对应的数据帧间标签作为目标数据帧间标签;根据所述目标场景分类信息和所述目标数据帧间标签从所述待处理视频数据中提取目标场景片段。In one embodiment, the model training module is further configured to match the target scene information with the scene classification information in the video data index to determine the target scene classification information; the scene corresponding to the target scene classification information The segment is used as the target scene segment, and the data inter-frame tag corresponding to the target scene segment is used as the target data inter-frame tag; extract from the video data to be processed according to the target scene classification information and the target data inter-frame tag Target scene fragment.
在一实施例中,所述模型训练模块,还用于根据所述目标场景分类信息和所述目标数据帧间标签进行数据位置定位,以确定目标场景片段的开始时间和结束时间;根据所述开始时间和所述结束时间从所述待处理视频数据中提取目标场景片段。In one embodiment, the model training module is further configured to perform data location positioning according to the target scene classification information and the inter-frame label of the target data, so as to determine the start time and end time of the target scene segment; according to the The start time and the end time extract target scene segments from the video data to be processed.
在一实施例中,所述模型训练模块,还用于在所述目标场景片段为多个时,根据各目标场景片段对应的开始时间和结束时间对目标场景片段进行排序,得到排序后的目标场景片段;依次根据排序后的目标场景片段进行模型训练。In one embodiment, the model training module is further configured to sort the target scene segments according to the start time and end time corresponding to each target scene segment when there are multiple target scene segments, to obtain the sorted target scene segments. Scene fragments; perform model training according to the sorted target scene fragments in turn.
在本申请所述数据索引装置的其他实施例或具体实现方法可参照上述各方法实施例,此处不再赘述。For other embodiments or specific implementation methods of the data indexing device described in this application, reference may be made to the above-mentioned method embodiments, which will not be repeated here.
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者装置不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者装置所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、方法、物品或者装置中还存在另外的相同要素。It should be noted that, in this document, the term "comprising", "comprising" or any other variation thereof is intended to cover a non-exclusive inclusion such that a process, method, article or apparatus comprising a set of elements includes not only those elements, It also includes other elements not expressly listed, or elements inherent in the process, method, article, or device. Without further limitations, an element defined by the phrase "comprising a ..." does not preclude the presence of additional identical elements in the process, method, article, or apparatus comprising that element.
上述本申请实施例序号仅仅为了描述,不代表实施例的优劣。The serial numbers of the above embodiments of the present application are for description only, and do not represent the advantages and disadvantages of the embodiments.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该估算机软件产品存储在如上所述的一个估算机可读存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台智能设备(可以是手机,估算机,数据索引设备,或者网络数据索引设备等)执行本申请各个实施例所述的方法。Through the description of the above embodiments, those skilled in the art can clearly understand that the methods of the above embodiments can be implemented by means of software plus a necessary general-purpose hardware platform, and of course also by hardware, but in many cases the former is better implementation. Based on this understanding, the essence of the technical solution of this application or the part that contributes to the prior art can be embodied in the form of a software product, which is stored in a computer-readable storage medium as described above (such as ROM/RAM, magnetic disk, optical disk), including several instructions to make a smart device (which can be a mobile phone, a computing machine, a data indexing device, or a network data indexing device, etc.) execute the various embodiments described in this application Methods.
以上仅为本申请的优选实施例,并非因此限制本申请的专利范围,凡是利用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利保护范围内。The above are only preferred embodiments of the present application, and are not intended to limit the patent scope of the present application. All equivalent structures or equivalent process transformations made by using the description of the application and the accompanying drawings are directly or indirectly used in other related technical fields. , are all included in the patent protection scope of the present application in the same way.

Claims (10)

  1. 一种数据索引方法,其特征在于,所述数据索引方法包括以下步骤:A data indexing method, characterized in that the data indexing method comprises the following steps:
    对待处理视频数据进行场景分类,获得多个场景片段;Perform scene classification on the video data to be processed to obtain multiple scene fragments;
    获取各场景片段对应的片段信息;Obtain the fragment information corresponding to each scene fragment;
    根据所述片段信息生成各场景片段对应的数据帧间标签和场景分类信息;Generate inter-frame data tags and scene classification information corresponding to each scene segment according to the segment information;
    根据所述数据帧间标签和所述场景分类信息构建所述待处理视频数据中各场景片段对应的视频数据索引。A video data index corresponding to each scene segment in the video data to be processed is constructed according to the data inter-frame tag and the scene classification information.
  2. 如权利要求1所述的数据索引方法,其特征在于,所述根据所述片段信息生成各场景片段对应的数据帧间标签和场景分类信息,包括:The data indexing method according to claim 1, wherein said generating the data inter-frame labels and scene classification information corresponding to each scene segment according to the segment information comprises:
    根据所述片段信息确定各场景片段对应的数据帧信息和片段属性信息;Determining data frame information and segment attribute information corresponding to each scene segment according to the segment information;
    根据所述数据帧信息生成各场景片段对应的数据帧间标签,并根据所述片段属性信息生成各场景片段对应的场景分类信息。Generate data inter-frame tags corresponding to each scene segment according to the data frame information, and generate scene classification information corresponding to each scene segment according to the segment attribute information.
  3. 如权利要求2所述的数据索引方法,其特征在于,所述根据所述数据帧信息生成各场景片段对应的数据帧间标签,包括:The data indexing method according to claim 2, wherein said generating the data inter-frame tags corresponding to each scene segment according to the data frame information comprises:
    根据所述数据帧信息确定各场景片段对应的帧头信息和帧尾信息;determining frame header information and frame tail information corresponding to each scene segment according to the data frame information;
    根据所述帧头信息和所述帧尾信息生成各场景片段对应的数据帧间标签。Generate data inter-frame tags corresponding to each scene segment according to the frame header information and the frame trailer information.
  4. 如权利要求1至3中任一项所述的数据索引方法,其特征在于,所述根据所述数据帧间标签和所述场景分类信息构建所述待处理视频数据中各场景片段对应的视频数据索引之后,还包括:The data indexing method according to any one of claims 1 to 3, wherein the video corresponding to each scene segment in the video data to be processed is constructed according to the data inter-frame label and the scene classification information After data indexing, also includes:
    在接收到模型训练指令时,根据所述模型训练指令确定目标场景信息;When receiving a model training instruction, determine target scene information according to the model training instruction;
    根据所述目标场景信息和所述视频数据索引从所述待处理视频数据中提取目标场景片段;extracting target scene segments from the video data to be processed according to the target scene information and the video data index;
    根据所述目标场景片段进行模型训练。Model training is performed according to the target scene segment.
  5. 如权利要求4所述的数据索引方法,其特征在于,所述根据所述目标场景信息和所述视频数据索引从所述待处理视频数据中提取目标场景片段,包括:The data indexing method according to claim 4, wherein said extracting a target scene segment from said video data to be processed according to said target scene information and said video data index comprises:
    将所述目标场景信息与视频数据索引中的场景分类信息进行匹配,以确定目标场景分类信息;Matching the target scene information with the scene classification information in the video data index to determine the target scene classification information;
    将所述目标场景分类信息对应的场景片段作为目标场景片段,并将所述目标场景片段对应的数据帧间标签作为目标数据帧间标签;Using the scene segment corresponding to the target scene classification information as the target scene segment, and using the data inter-frame tag corresponding to the target scene segment as the target data inter-frame tag;
    根据所述目标场景分类信息和所述目标数据帧间标签从所述待处理视频数据中提取目标场景片段。Extracting target scene segments from the video data to be processed according to the target scene classification information and the target data inter-frame tags.
  6. 如权利要求5所述的数据索引方法,其特征在于,所述根据所述目标场景分类信息和所述目标数据帧间标签从所述待处理视频数据中提取目标场景片段,包括:The data indexing method according to claim 5, wherein said extracting a target scene segment from said video data to be processed according to said target scene classification information and said target data inter-frame label comprises:
    根据所述目标场景分类信息和所述目标数据帧间标签进行数据位置定位,以确定目标场景片段的开始时间和结束时间;performing data location positioning according to the target scene classification information and the target data inter-frame label to determine the start time and end time of the target scene segment;
    根据所述开始时间和所述结束时间从所述待处理视频数据中提取目标场景片段。Extracting target scene segments from the video data to be processed according to the start time and the end time.
  7. 如权利要求6所述的数据索引方法,其特征在于,所述根据所述目标场景片段进行模型训练,包括:The data indexing method according to claim 6, wherein said performing model training according to said target scene segment comprises:
    在所述目标场景片段为多个时,根据各目标场景片段对应的开始时间和结束时间对目标场景片段进行排序,得到排序后的目标场景片段;When there are multiple target scene segments, the target scene segments are sorted according to the corresponding start time and end time of each target scene segment, and the sorted target scene segments are obtained;
    依次根据排序后的目标场景片段进行模型训练。Model training is performed sequentially according to the sorted target scene segments.
  8. 一种数据索引装置,其特征在于,所述数据索引装置包括:A data indexing device, characterized in that the data indexing device includes:
    场景分类模块,用于对待处理视频数据进行场景分类,获得多个场景片段;The scene classification module is used to classify the scene of the video data to be processed, and obtain a plurality of scene fragments;
    信息获取模块,用于获取各场景片段对应的片段信息;An information acquisition module, configured to acquire segment information corresponding to each scene segment;
    数据生成模块,用于根据所述片段信息生成各场景片段对应的数据帧间标签和场景分类信息;A data generation module, configured to generate inter-frame labels and scene classification information corresponding to each scene segment according to the segment information;
    数据索引模块,用于根据所述数据帧间标签和所述场景分类信息构建所述待处理视频数据中各场景片段对应的视频数据索引。A data index module, configured to construct a video data index corresponding to each scene segment in the video data to be processed according to the data inter-frame tags and the scene classification information.
  9. 一种数据索引设备,其特征在于,所述数据索引设备包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的数据索引程序,所述数据索引程序被处理器执行时实现如权利要求1至7中任一项所述的数据索引方法。A data indexing device, characterized in that the data indexing device includes: a memory, a processor, and a data indexing program stored in the memory and operable on the processor, and the data indexing program is executed by the processor When executed, the data indexing method according to any one of claims 1 to 7 is realized.
  10. 一种存储介质,其特征在于,所述存储介质上存储有数据索引程序,所述数据索引程序被处理器执行时实现如权利要求1至7中任一项所述的数据索引方法。A storage medium, characterized in that a data indexing program is stored on the storage medium, and when the data indexing program is executed by a processor, the data indexing method according to any one of claims 1 to 7 is implemented.
PCT/CN2021/113518 2021-08-19 2021-08-19 Data indexing method, apparatus and device, and storage medium WO2023019510A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2021/113518 WO2023019510A1 (en) 2021-08-19 2021-08-19 Data indexing method, apparatus and device, and storage medium
CN202180099925.4A CN117597680A (en) 2021-08-19 2021-08-19 Data indexing method, device, equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2021/113518 WO2023019510A1 (en) 2021-08-19 2021-08-19 Data indexing method, apparatus and device, and storage medium

Publications (1)

Publication Number Publication Date
WO2023019510A1 true WO2023019510A1 (en) 2023-02-23

Family

ID=85239392

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/113518 WO2023019510A1 (en) 2021-08-19 2021-08-19 Data indexing method, apparatus and device, and storage medium

Country Status (2)

Country Link
CN (1) CN117597680A (en)
WO (1) WO2023019510A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1993989A (en) * 2005-05-26 2007-07-04 索尼株式会社 Contents processing device, contents processing method, and computer program
US20100246965A1 (en) * 2009-03-31 2010-09-30 Microsoft Corporation Tagging video using character recognition and propagation
CN111209431A (en) * 2020-01-13 2020-05-29 上海极链网络科技有限公司 Video searching method, device, equipment and medium
CN111797801A (en) * 2020-07-14 2020-10-20 北京百度网讯科技有限公司 Method and apparatus for video scene analysis

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1993989A (en) * 2005-05-26 2007-07-04 索尼株式会社 Contents processing device, contents processing method, and computer program
US20100246965A1 (en) * 2009-03-31 2010-09-30 Microsoft Corporation Tagging video using character recognition and propagation
CN111209431A (en) * 2020-01-13 2020-05-29 上海极链网络科技有限公司 Video searching method, device, equipment and medium
CN111797801A (en) * 2020-07-14 2020-10-20 北京百度网讯科技有限公司 Method and apparatus for video scene analysis

Also Published As

Publication number Publication date
CN117597680A (en) 2024-02-23

Similar Documents

Publication Publication Date Title
US11605226B2 (en) Video data processing method and apparatus, and readable storage medium
CN110210387B (en) Method, system and device for detecting insulator target based on knowledge graph
CN102880854B (en) Distributed processing and Hash mapping-based outdoor massive object identification method and system
CN104537336A (en) Face identification method and system with self-learning function
CN112884005B (en) Image retrieval method and device based on SPTAG and convolutional neural network
US20230259509A1 (en) Data association query method and apparatus, device, and storage medium
WO2022089143A1 (en) Method for generating analog image, and electronic device and storage medium
CN110706314A (en) Element layout method and device, electronic equipment and readable storage medium
CN113627402A (en) Image identification method and related device
CN103970901B (en) A kind of geography information graph data integration method
JP2021170399A (en) Signal lamp identification method and apparatus, device, storage medium and program
CN103473569A (en) Medical image classification method based on SVM
CN105320703B (en) Enquiry expanding method and device in similar vehicle retrieval
CN103488977A (en) Medical image management system based on SVM
WO2023019510A1 (en) Data indexing method, apparatus and device, and storage medium
CN113760242B (en) Data processing method, device, server and medium
CN111797772B (en) Invoice image automatic classification method, system and device
CN112668411B (en) SGAN-based video stream marking method, SGAN-based video stream marking device, SGAN-based video stream marking system and storage medium
CN109446356A (en) A kind of multimedia document retrieval method and device
CN115543809A (en) Method and device for constructing test scene library of automatic driving function
CN111553335A (en) Image generation method and apparatus, and storage medium
CN114580544B (en) Scene task processing method, device, equipment and computer readable storage medium
CN116662559B (en) Case knowledge graph construction platform and method based on big data technology
CN115810183B (en) Traffic sign detection method based on improved VFNet algorithm
CN113627568B (en) Method, device and equipment for supplementing marks and readable storage medium

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 202180099925.4

Country of ref document: CN

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21953756

Country of ref document: EP

Kind code of ref document: A1