WO2016146847A1 - A method of analysing a multimedia file - Google Patents

A method of analysing a multimedia file Download PDF

Info

Publication number
WO2016146847A1
WO2016146847A1 PCT/EP2016/056064 EP2016056064W WO2016146847A1 WO 2016146847 A1 WO2016146847 A1 WO 2016146847A1 EP 2016056064 W EP2016056064 W EP 2016056064W WO 2016146847 A1 WO2016146847 A1 WO 2016146847A1
Authority
WO
WIPO (PCT)
Prior art keywords
multimedia file
objects
analysing
manually
tracker system
Prior art date
Application number
PCT/EP2016/056064
Other languages
French (fr)
Inventor
Ian Kerr
Mervyn Graham
Csorba KRISTÓF
Varga MÁRTON BÁLINT
Lipták LEVENTE
Kundra LÁSZLÓ
Szabolcs Fodor
Istvan ALBERT
Original Assignee
Idaso Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Idaso Limited filed Critical Idaso Limited
Publication of WO2016146847A1 publication Critical patent/WO2016146847A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • G06V20/54Surveillance or monitoring of activities, e.g. for recognising suspicious objects of traffic, e.g. cars on the road, trains or boats

Definitions

  • This invention relates to a method of analysing a multimedia fife,
  • the present invention is directed towards a method for analysing a multimedia file for purposes of observing objects appearing in the multimedia file by detecting, counting, tracking and/or classifying such objects.
  • multimedia file shall be understood to encompass any type of multimedia file which may be stored and downloaded for analysis or streamed from a remote storage for analysis or streamed live from a content source location for real-time analysis.
  • the multimedia file may be a video file and such a video file can be interpreted as containing a plurality of images which together make up the video file. It may not, and is likeiy not to, have any audio component.
  • the multimedia file which is to be analysed is envisaged to be a video file showing a road junction, and the objects to be observed would be vehicles passing through the road junction.
  • the multimedia file will be created using content recorded from a traffic camera, closed-circuit television camera or similar type of camera, which records images of the junction to be analysed.
  • a further aspect of the invention envisages observing people as the objects. This could be used in analysing public spaces to determine the flow of people through a pedestrian junction, or monitor the flow of people through public spaces such as stadiums, shopping centres and the like.
  • the primary implementation of the present invention is envisaged to assist with object ⁇ e.g. vehicles) movement observation, specifically, object detection, object counting, object tracking and object classification from video footage.
  • the detection, counting and tracking of vehicles as they pass through a junction can be of great assistance for traffic management purposes.
  • the present invention is able to count where the vehicles travel from, and travel on to as they pass through a specific junction, during a time period, for example 12 hours.
  • a further aspect of the invention is to classify the vehicles as this is useful from a statistical point of view for planning road extensions and road capacity requirements for heavy loads.
  • junction refers to any roadway, or route which has a plurality of arms, along each of which arms a vehicle may travel and endpoints of the arms meet one another to form the junction.
  • a 3-arm junction is typically known as a T-junction; a 4-arm junction may be a crossroads or a 4-arm roundabout
  • a vehicle passing through the junction will enter along one arm of the junction and will exit along an arm. It is a main object of the present invention to detect and count the number of vehicles passing through a selected junction and to also track the movement of the vehicles through the junction, during a time period, by analysing the entry and exit arms for each of the vehicles passing through the junction. The classification of the vehicle is also determined.
  • the class of a vehicle may be one of: a car, a bus, a taxi, a truck, a van, a lorry and so on.
  • U.S. Patent Number US 7,460,691 (GET TECHNOLOGIES PTE LIMITED) is a further example of the prior art and discloses image processing techniques which are applied video images for the purposes of acquiring traffic data.
  • the disclosure discusses a traffic monitoring system, the basic function of which is for traffic data acquisition and incident detection and in particular the application of image processing techniques for the detection of a vehicle, from sequence of video images, as well as the acquisition of traffic data and detection of any traffic incidents.
  • the present invention is directed to a method of analysing a multimedia file for the detection, counting, tracking and/or classification of objects displayed in images contained within the multimedia file, wherein the method comprises the steps of executing an automated analysis of the multimedia file, wherein, the automated analysis comprises observing the objects by detecting, counting, tracking and/or classifying the objects in the multimedia file, and producing an accuracy rating for at least each observation of an object; marking chosen sections of the multimedia file, where observations of the objects in those chosen sections have an accuracy rating below a predefined threshold; and, manually observing objects in the chosen sections of the multimedia file.
  • the advantage of providing such a semi-automated method is that the object analysis of a multimedia file can be carried out in a relative short amount of time, whilst the accuracy of the detections is maintained at a relatively high standard due to the manual intervention in chosen sections of the multimedia file where the automated detection has been deemed to have an accuracy rating below an acceptable predefined threshold.
  • the method also allows for correction of the automated detection on the fly in non-chosen sections of the multimedia file.
  • the accuracy rating is produced separately for the detection, counting, tracking and/or classification of the objects.
  • the accuracy rating is an accumulated accuracy rating which is produced based on any combination of the detection, counting, tracking and/or classification of the objects.
  • the accuracy rating is calculated on the probability of the detection, counting, tracking, and/or classifying being correct.
  • the probability of the detection, counting, tracking, and/or classifying of the object being correct is based on one or more of: to what extent the motion of the object complies with expected motions of the object, such as kinematic and dynamic capabilities of the object; whether the object has properly approached, entered and exited an internal area, which internal area is specified by a user and is a portion of an image contained in the multimedia file; whether a direction of motion of the object complies with a preset movement path on the image of the multimedia file, which movement path has been defined by the user; whether the object is distinct and separated from other objects by a predefined distance in the image of the multimedia file; and/or, whether a tracked path of the object is distinct from other tracked paths of other objects.
  • the method comprises a further step of producing an object analysis report.
  • the method comprises a further step of manually correcting erroneous detections in non-chosen sections of the multimedia file. In a further embodiment, the method comprises a further step of manually classifying the automatically detected objects and/or the manually detected objects.
  • the step of manually classifying the automatically detected objects and/or the manually detected objects includes displaying a thumbnail image of the object to be classified to a user, and allowing the user to call up a short video of the object to be classified if the thumbnail image is deemed to be insufficient to classify the object.
  • the method comprises a further step of validating portions of the automatically detected objects, the manually detected objects and the manually classified objects, by way of making a comparison with manually detected objects and manually classified objects respectively.
  • the multimedia file is a video file. In a further embodiment, the multimedia file is a video file of a junction.
  • the objects are moving objects.
  • the objects are vehicles.
  • a user manually detects, counts, tracks and/or classifies objects using a game controller suitable for a games console.
  • background areas on images contained in the multimedia file may be selected by a user in advance of executing the automated analysis of the multimedia file such that those selected background areas in the images of the multimedia file can be excluded from the automated analysis.
  • internal areas on images contained in the multimedia file may be selected by a user in advance of executing the automated analysis of the multimedia file so that the selected internal areas in the images of the multimedia file ca be used during the automated analysis to assist in calculating the accuracy rating and/or confirming a valid detection of an object.
  • analysis parameters for analysing the multimedia file are input by a user.
  • the analysis parameters input by the user comprise one or more of: movement paths, camera settings, time settings, background and internal areas, and, tracker configuration.
  • the present invention is further directed towards a tracker system for analysing a multimedia file so as to detect, count, track and/or classify objects displayed in images contained within the multimedia file;
  • the tracker system comprising means for executing an automated analysis of the multimedia file, wherein, the automated analysis comprising observing objects in the multimedia file by detecting, counting, tracking and/or classifying the objects in the multimedia file, and means for producing an accuracy rating for each observation of an object; and, means for marking chosen sections of the multimedia file, where observations of the objects in those chosen sections have an accuracy rating below a predefined threshold; and, means to manually detect objects in the chosen sections of the multimedia file.
  • the means for producing an accuracy rating for each observation of the object produces and accuracy rating separately for each of the detection, counting, tracking and/or classification of the objects. In a further embodiment, the means for producing an accuracy rating for each observation of the object produces an accumulated accuracy rating based on a combination of the accuracy ratings for the detection, counting, tracking and/or classification of the objects.
  • the accuracy rating is calculated on the probability of the detection, counting, tracking, and/or classifying being correct.
  • the probability of the detection, counting, tracking, and/or classifying of the object being correct is based on one or more of: to what extent the motion of the object complies with expected motions of the object, such as kinematic and dynamic capabilities of the object; whether the object has properly approached, entered and exited an internal area, which internal area is specified by a user and is a portion of an image contained in the multimedia file; whether a direction of motion of the object complies with a preset movement path on the image of the multimedia file, which movement path has been defined by the user; whether the object is distinct and separated from other objects by a predefined distance in the image of the multimedia file; and/or, whether a tracked path of the object is distinct from other tracked paths of other objects.
  • the tracker system further comprises means to manually correct erroneous detections in non-chosen sections of the multimedia file. In a further embodiment, the tracker system further comprises means to manually classify the automatically detected objects and/or the manually detected objects.
  • the means to manually classify the automatically detected objects and/or the manually detected objects comprises means to display a thumbnail image of the object to be classified by a user, and means to allow the user to call up a short video of the object to be classified if the thumbnail image is deemed to be insufficient to classify the object.
  • the tracker system further comprises means to validate portions of the automatically detected objects, the manually detected objects and the manually classified objects being validated by way of a comparison with manually detected and classified objects.
  • the tracker system further comprises means to produce an object analysis report.
  • the multimedia file is a video file. In a further embodiment, the multimedia file is a video file of a junction. In a further embodiment, the objects are moving objects. In a further embodiment, the objects are vehicles.
  • the user manually detects, counts, tracks and/or classifies objects using a game controller suitable for a games console.
  • the tracker system further comprises means for selecting background areas on images contained in the multimedia file in advance of executing the automated analysis of the multimedia file such that those selected background areas in the images of the multimedia file can be excluded from the automated analysis.
  • the tracker system further comprises means to select internal areas on images contained in the multimedia file in advance of executing the automated analysis of the multimedia file so that the selected internal areas in the images of the multimedia file can be used during the automated analysis to assist in calculating an accuracy rating and/or confirming a valid detection of an object.
  • the tracker system further comprises input means to allow a user input analysis parameters for analysing the multimedia file to prior to executing the automated analysis of the multimedia file.
  • the analysis parameters input by the user comprise one or more of: movement paths, camera settings, time settings, background and internal areas, and, tracker configuration.
  • the present invention is further directed to a method of analysing a multimedia file for the detection, counting, tracking and/or classification of objects displayed in images contained within the multimedia file, wherein the method comprises the steps of executing an automated analysis of the multimedia file, wherein, the automated analysis detects, counts, tracks and/or classifies objects in the multimedia file by at least detecting an object, and producing an accuracy rating for at least each detection of an object; marking chosen sections of the multimedia file, where detections of the objects in those chosen sections have an accuracy rating below a predefined threshold; manually detecting objects in the chosen sections of the multimedia file; manually correcting erroneous detections in non-chosen sections of the multimedia file; and producing an object analysis report.
  • the present invention is further directed to a method of analysing a multimedia file for the detection, counting, tracking and/or classification of objects displayed in images contained within the multimedia file, wherein the method comprises the steps of executing an automated analysis of the multimedia file, wherein, the automated analysis detects, counts, tracks and/or classifies objects i the multimedia file by at least detecting an object, and producing an accuracy rating for at least each detection of an object; marking chosen sections of the multimedia file, where detections of the objects in those chosen sections have an accuracy rating below a predefined threshold; manually detecting objects in the chosen sections of the multimedia file; manually correcting erroneous detections in non-chosen sections of the multimedia file; manually classifying the automatically detected objects and the manually detected objects; and producing an object analysis report.
  • the present invention is further directed to a method of analysing a multimedia file for the detection, counting, tracking and/or classification of objects displayed in images contained within the multimedia file, wherein the method comprises the steps of executing an automated analysis of the multimedia file, wherein, the automated analysis detects, counts, tracks and/or classifies objects in the multimedia by at least detecting an object, and producing an accuracy rating for at least each detection of an object; marking chosen sections of the multimedia file, where detections of the objects in those chosen sections have an accuracy rating below a predefined threshold; manually detecting objects in the chosen sections of the multimedia file; manually correcting erroneous detections in non-chosen sections of the multimedia file; manually classifying the automatically detected objects and the manually detected objects; validating portions of the automatically detected objects, the manually detected objects and the manually classified objects by way of a comparison with manually detected and classified objects; and producing an object analysis report.
  • the multimedia file is a video file
  • the objects are moving objects.
  • the objects are vehicles.
  • the multimedia file is a video file of a junction.
  • the accuracy rating is calculated on the probability of the detection, counting, tracking, and/or classifying being correct.
  • a user manually detects, counts, tracks and/or classifies objects using a game controller.
  • the game controller can be programmed.
  • the manual classification of the automatically detected objects and the manually detected objects includes displaying a thumbnail image of the object to be classified to a user, and further allows a user to call up a short video of the object to be classified if the thumbnail image is deemed to be insufficient to classify the object.
  • background areas on images from the multimedia file may be selected by a user in advance of executing the automated analysis of the multimedia file so that background areas in the images of the multimedia file can be excluded from the automated analysis.
  • internal areas on images from the multimedia file may be selected b a user in advance of executing the automated analysis of the multimedia file so that the internal areas in the images of the multimedia file can be used during the automated analysis to assist in calculating the accuracy rating and/or confirming a valid detection of an object.
  • analysis parameters for analysing the multimedia file are input by a user prior to executing the automated analysis of the multimedia file.
  • the analysis parameters input by the user comprise one or more of: movement paths, camera settings, time settings, background and internal areas, and, tracker configuration.
  • the present invention is further directed towards a tracker system for analysing a multimedia file so as to detect, count, track and/or classify objects displayed in images contained within the multimedia file, characterised in that the tracker system comprises means for executing an automated analysis of the multimedia file, wherein, the automated analysis detects, counts, tracks and/or classifies objects in the multimedia file in the form of a detection of an object, and produces an accuracy rating for each detection of an object; and, means for marking chosen sections of the multimedia file, where detections of the objects in those chosen sections have an accuracy rating below a predefined threshold; means to manually detect objects in the chosen sections of the multimedia file.
  • the present invention is further directed towards a tracker system for analysing a multimedia file so as to detect, count, track and/or classify objects displayed in images contained within the multimedia file, characterised in that the tracker system comprises means for executing an automated analysis of the multimedia file, wherein, the automated analysis detects, counts, tracks and/or classifies objects in the multimedia file in the form of a detection of an object, and produces an accuracy rating for each detection of an object; means for marking chosen sections of the multimedia file, where detections of the objects in those chosen sections have an accuracy rating below a predefined threshold; means to manually detect objects in the chosen sections of the multimedia file; means to manually correct erroneous detections in non-chosen sections of the multimedia file; means to manually classify the automatically detected objects and the manually detected objects; means to validate portions of the automatically detected objects, the manually detected objects and the manually classified objects being validated by way of a comparison with manually detected and classified objects; and means to produce an object analysis report.
  • the tracker system comprises means for executing an automated analysis of the multimedia file,
  • Figure 1 is a flow diagram detailing some of the steps involved in the present invention.
  • Figure 2 is an image taken from a multimedia file, whereby, objects appearing in the image are to be analysed in accordance with the present invention
  • Figure 3 is a screenshot of a user interface for manually counting and tracking objects using a tracker system in accordance with the present invention.
  • Figure 4 is a screenshot of a user interface for manually classifying objects using a tracker system in accordance with the present invention.
  • the multimedia file analysis method 100 comprises a plurality of steps, 102 to 116, which enable the analysis of objects which appear in the multimedia file.
  • step 102 a user will define analysis parameters for analysing a multimedia file.
  • step 104 an automated observation of objects displayed in the multimedia file is carried out, based on defined analysis parameters.
  • Figure 1 refers to an automated count as opposed to an automated observation; however, it will be understood that throughout this specification, the term Observation of objects', and its grammatical equivalents, shall be interpreted as detecting objects displayed in one or more images which make up a part of the multimedia file, counting the objects displayed in the one or more images which make up a part of the multimedia file, tracking the objects displayed in the one or more images which make up a part of the multimedia file, and/or, classifying the objects displayed in the one or more images which make up a part of the multimedia file.
  • the automated observation of objects may be carried out over the full length of the multimedia file, which will typically be a video file; or, the automated observation may be carried out during a predefined selected portion of the multimedia file.
  • step 106 automated count results from step 104 are transferred for correction and verification by a user.
  • step 08 the user will manually count objects, which shall be understood to be manually observing objects, for chosen sections of the multimedia file.
  • the chosen sections will be sections of the multimedia file where the automated observation has a relatively low accuracy rating as the probability of the observations regarding the observed object being correct have been adjudged to be low.
  • the method of the present invention then specifies for these sections to be manually counted.
  • step 110 the user will manually correct the automated count in non-chosen sections of the multimedia file, if the user notices that there are any errors in these non-chosen sections when having a brief look through the multimedia file.
  • This step is of course not mandatory and may be omitted in some embodiments of the invention.
  • step 1 12 the user manually classifies objects in the multimedia file according to preset criteria, or, objects are automatically classified according to the preset criteria. It will be appreciated that it is foreseen to automatically classify at least some of the objects detected in the multimedia file based on assessing the size and/or shape of the object detected in the multimedia file. As will be discussed in greater detail hereinbelow, the angle of view which the camera has relative to the objects passing through the field of vision of the camera will be taken into account when determining the criteria for assessing the shape and size of objects appearing in the images of the multimedia file.
  • step 114 of the method the automated and manual analysis results are verified by have a user check randomly chosen portions of the multimedia file. ln the final step 116, an analysis report is created.
  • steps such as the creation of a report, or the verification of data, or the classification of objects may be omitted.
  • the implementation of the method will focus on the observation of objects, thus incorporating the detection, counting, tracking and classification of objects appearing in a multimedia file, such as detecting, counting, tracking and classifying vehicles which appear on a video file of a junction with the vehicles passing through the junction.
  • Alternative implementations for this methodology are envisaged for observing the movement of people, animals and other types of transportation at junctions or open public spaces.
  • one or more users may be used to complete the steps given hereinbefore; the invention is thus not limited to the same user completing all of the steps.
  • the system will allows manually implemented steps to be carried out by users derived from a crowd sou reed community of workers.
  • an individual or group of individuals would log into the system, and possibly complete a training exercise, prior to then carrying out the manual detection, counting, tracking, classifying and/or verifying of objects in an assigned multimedia file.
  • the system would assign work to the logged in users and would track the work carried out by each user to ensure it is of a sufficient quality rating before authorising a payment for the user. Now turning to consider each of the steps in more detail.
  • FIG. 200 there is shown an image indicated generally by reference numeral 200, from a multimedia file (not shown) in which a number of vehicles 202, 204 are seen to be passing through a three-armed junction indicated generally by reference numeral 206.
  • the three-armed junction 206 is a round-about.
  • a background area (discussed hereinbelow) 208 has been defined by a user.
  • An internal area (discussed further below) 210 has also been defined by the user.
  • a movement path (discussed hereinunder) 212 which is indicated by a plurality of arrows, has been defined by the user.
  • a user will define the analysis parameters for a tracker system which will be used for the multimedia file analysts of the image of Figure 2.
  • the analysis parameters are required by the tracker system, in addition to the multimedia file, to be able to detect the objects, in this case vehicles, in the multimedia file.
  • the analysis parameters comprise, but are not limited to, one or more of the following: a) Movement routes; each possible movement through the junction is represented as a matrix of movement directions.
  • the matrix of movement directions is shown in a top left comer of Figure 2, where each direction of each arm is accorded a separate reference letter.
  • a vehicle entering the junction from the bottom of the Figure 2 and turning left, from a driver's perspective, will travel through cell C and then pass through cell A, and exit the junction through cell D. If the vehicle entering the junction from the bottom of the Figure 2 turns right off of the round-about, again from a driver's perspective, the vehicle will travel through cell C and then pass through cell A, and exit the junction through cell F.
  • Cell A is the junction/round-about itself.
  • Camera settings a plurality of settings which indicate the position and/or orientation of the camera are used to allow the method to determine the view contained in the multimedia file so that a plane may be established. For example, one or more of: a height of installation of the camera from ground level (e.g. 5m), a field of view of the camera (e.g. 65°), a horizon height which refers to the location of the horizon relative to a chosen border of the camera view (e.g. 23% below the top border), a cell size for the movement matrix (e.g. 1m), and a camera angle which is the angle formed between the direction in which camera faces and the horizontal (e.g.
  • a height of installation of the camera from ground level e.g. 5m
  • a field of view of the camera e.g. 65°
  • a horizon height which refers to the location of the horizon relative to a chosen border of the camera view (e.g. 23% below the top border)
  • a cell size for the movement matrix e.g
  • 30° downward are provided by a user in order to allow the view of the junction to be contextualised by the method, for example, to understand how far into the distance the camera view can see.
  • This can assist with detecting and classifying objects as the size and shape of the objects will vary according to the angle/plane of view of the camera, and the distance towards the horizon which the camera view can see, and hence provides in the multimedia file.
  • Time settings; frames per second in real time are made known to the method to allow the method to be able to calculate the real velocity (m/s) of the objects appearing in the multimedia file, which will be the velocity of the vehicles passing through the junction.
  • Background and internal areas the user can select areas of the image from the multimedia file which can be ignored from the point of view of the analysis. For example, flags or trees fluttering due to wind, or, a pedestrian walking on the footpath may be excluded, if desired by using a background area. The flags/trees could be confused for objects that are to be tracked, and, it may not always be desired to detect and analyse pedestrians in some circumstances. Although it will be understood, that in other embodiments, the roadways and carriageways may be excluded so that the method can focus on analysing pedestrians on the footpaths, if desired.
  • the internal areas can be used to help trigger an observation of an object by setting the parameters such that an object will only be counted, tracked and or classified if it at least enters the internal area, or enters and subsequently leaves the internal area.
  • Tracker configuration the method employed by the tracker system can be adjusted to take account of several pre-defined conditions. For example, different video formats will require different configurations due to the difference image contrasts that exist between the different configurations.
  • the tracker system may also be used during different lighting conditions, such as low light at dusk or dawn, or at night time.
  • the tracker system provides the user with a user interface to allow the user to easily input and set these parameters.
  • the background areas, the internal areas and the movement paths can be defined on a drawing area which can be overlaid on an image from the multimedia file.
  • the movement paths can be drawn using a stroke of a brush- type software tool which can define a possible travel path. In this manner, each of the possible travel paths can be input quickly.
  • the software tool may also take account of the angle/plane of the camera when accepting the path drawing stroke so that the depth perspective of the camera view is taken into account by automatically narrowing the path appropriately when the drawing stroke is made towards the horizon.
  • the background areas and the internal areas may be automatically defined by the tracker system.
  • the internal area 210 is needed for the tracker system to be able to distinguish between valid vehicles and invalid vehicles.
  • a valid vehicle is defined as an object/vehicle which is initially detected outside of the internal area 210, is detected inside the internal area 210 and hence is considered to pass through the internal area 210 through the internal area 210, and subsequently is detected outside of the internal area 210 again.
  • step 104 the automated count of objects in the multimedia file is executed based on defined analysis parameters from step 102.
  • the tracker system exports the analysis parameters set by the user defined in the previous section for the tracker and then executes the tracker system to begin scrolling through the multimedia file and counting the objects shown in the multimedia file.
  • the tracker system provides the user with progress information and estimates the total time of execution. This is important as the tracking execution time for a longer video (12 hr) can last many hours. The user can also cancel/restart the tracking at any time.
  • the tracker runs through the entire multimedia file and runs the object/vehicle detection algorithm to count the objects/vehicle.
  • An object or vehicle is detected by comparing pixels of adjacent frames to detect the presence of an object, and its motion relative to the background pixels which will not substantively change.
  • the tracker system exports the tracking results.
  • the results includes one or more of: the frame numbers) in the multimedia file (e.g. video file) where the object entered and exited the junction; the detected object movement (e.g.
  • the estimated accuracy is the probability that the detected object is actually an object. The probability is determined by assessing:
  • the vehicle will have a maxima! acceleration level, a constraint against motion sideways without turning, and so on;
  • the object has a size and/or shape which is within an expected range of sizes and/or set of expected shape types, it could be that a saloon car, an estate car, a SUV, a track, a van, a motorbike, a pushbike, and so on all have different expected sizes and shapes and the detected size and shape would be compared to the expected sizes and shapes to determine if one of them substantially matches.
  • the position of the object on the image can be used to determine the distance of the object from the camera as this will clearly have a bearing on the expected size and shape of the object;
  • the object size should reduce if the object moves towards a horizon and vice versa; and/or,
  • the estimated accuracy is an important aspect of the invention and as many information sources (such as information provided by the user, information retrieved by the tracking system, and, information extracted from the images) are used to determine the estimated accuracy of the observation of the object in successive images from the camera.
  • information sources such as information provided by the user, information retrieved by the tracking system, and, information extracted from the images
  • the tracker system estimates the accuracy of the observation of an object by estimating the probability that the object has been detected, counted, tracked and/or classified correctly.
  • the tracker will mark chosen sections of the multimedia file, in the form of time ranges during the multimedia file, where the tracker was not able to observe the object with a high accuracy rating, as compared with an acceptable preset threshold. These are the sections of the multimedia file where there are either objects with low accuracy ratings (doubtful cases), or objects which do not seem to be vehicles but nonetheless have a higher accuracy than conventional noise (possible noise cases). These possible noise cases are where the tracker system definitely sees motion but it is not clearly an expected object, such as a vehicle.
  • the chosen section of the multimedia file will be grouped in accordance with the arms, or cells, of the junction under review. In this manner, it is easier for a manual count of the chosen sections of the multimedia file to be done for several objects all emanating from the same arm on the junction.
  • the tracker also determines a number in the form of a percentage value, which is referred to as a certainty ratio, for each arm of the junction.
  • This certainty ratio indicates how much of the multimedia file requires manual processing, which is an indicator of the quality of the automated analysis. This ratio is calculated based on the number of chosen sections of the multimedia file, the duration of the chosen sections and the distribution of the chosen sections. If the chosen sections are relatively long and are distributed relatively far from each other, then the user of the tracker system will not have any issues to operate the system as the user can simply use the time ratio of the chosen intervals, as chosen, to manually process the chosen sections.
  • the tracker system merges chosen sections which are close to each other, even though there will be a short certain section between the two chosen sections.
  • the user can set a parameter, such as five seconds, which determines if chosen sections are too close to one another and need to be merged to make the manual processing easier for the user.
  • the tracker system will merge these chosen sections (and any short certain sections falling therebetween) into a single longer, continuous section for manual processing. It should be noted that if an object or vehicle is correctly observed by the tracker system in a short certain section which is subsequently merged with chosen sections for manual processing, then the count, classification, detection and so on for that vehicle is removed so as to avoid double counting.
  • a user will manually count objects for the chosen sections of video, which had a relatively low accuracy estimated value. In this manner, any objects which the tracker system was uncertain about are manual counted by a user. Moreover, the user can correct non-chosen sections of the multimedia file and correct any errors that may be present, even though the tracker system would have evaluated these non-chosen sections to have a sufficiently high accuracy rating. As the multimedia is played back to the user, during these non-chosen sections a user can still override the automated result and correct errors in this way. After the automated analysis has been executed, the manual analysis may be carried out. The manual analysis consists of the following tasks:
  • the first two tasks can be completed on the same screen, which is shown in Figure 3.
  • the third task is completed on a separate screen as shown in Figure 4.
  • Detect noise as a valid vehicle For example, pedestrians, cyclists, trees, flags and other items may cause false detections which must then be manually deleted by the user.
  • the correction user interface 300 which allows a user to manually analyse objects in the chosen section(s) of the multimedia file which has been analysed automatically already, and, correct any errors that may be present in the non-chosen sections of the multimedia file is shown.
  • the correction interface 300 comprises a main panel 302 which contains the video of the junction with an overlay of detections and the thumbnail of the junction in the upper left corner of the main panel 302, which indicates the cells of the matrix which represent the arms of the junction, which is the given example are represented by reference letters A to G.
  • a progress bar indicated generally by reference numeral 304 is provided below the main panel 302. The progress bar 304 indicated the overall length of playback of the multimedia file, and where the user currently is within that playback period.
  • a classification table 306 is provided to one side of the main panel 302.
  • the classification table 306 lists the different type of classifications which may be accorded to the detected objects and may also provide a running count of each classification type.
  • An origin arm and destination arm indicator 308 is provided on the correction user interface 300 also. As described hereinbefore, playback will be carried out for one origin arm at a time, so for the most part the origin arm will be locked. If the user changes the origin arm on the origin arm and destination arm indicator 308, then the progress bar 304 and the main panel 302 will be updated to reflect the relevant information, such as the chosen and non-chosen sections for the newly selected origin arm.
  • Information on the frame number of the multimedia file, and a recording time may be provided as indicated by reference numeral 310.
  • each detected object 312 which in this embodiment will be a vehicle, is displayed on the main panel 302 with an overlay of a bounding box 314, a detected and tracked path 316 for that object and additional information such as the origin arm/destination arm pair reference letters 318 and the object ID number for that object 320 which is useful for identification purposes.
  • the bounding box 314 and the detected and tracked path 316 of the object 312 is visible and may be colour coded to differentiate numerous objects from one another when there are several objects being shown together on the same frame. In this manner, the user is able to quickly understand and recognise which objects have been detected and what is the detected and tracked path for that object is. Any errors are easily decipherable and the user can take manual correction action as required.
  • the user can add the object and classify the object, at the same time, using just one input device.
  • the user can be altered by way of a red border, flashing border or some other audio, visual or tactile alarm so as to encourage the user to pay particular attention during those periods.
  • the user will carry out the following steps of playing the multimedia file for a selected origin arm.
  • the user will manually count and classify the vehicles not have been detected, and, correct the tracker system errors. The user repeats this process for all of the origin arms at the junction.
  • the manual analysis of an object requires the user to select the destination arm of the vehicle, bearing in mind that the origin arm will be fixed, and, select the class of the object.
  • These selections can be made using any known input devices such as a mouse, by clicking on the appropriate destination arm button and the number next to the selected class for the object; using a keyboard in a similar fashion by way of shortcuts with predefined keys, or with a game controller which is used with a games console.
  • the game controller is seen as a preferred input device as the inputs can be made in a very quick and ergonomic manner.
  • the destination arm can be selected using a D-pad or control stick, by pushing the D-pad or control stick in one of the four available directions with can be associated with the different arms. Up to eight different arms could be accommodated using such a D-pad or control stick using the horizontal, vertical and diagonal movements of the D-pad or control stick.
  • the class of the object can also be specified with the game controller by using the buttons on the game controller, whereby the different buttons are each associated with one of the different classes of object.
  • the deletion can be executed with mouse by right clicking for example, or by using a keyboard with a key such as the 'delete' key, or by using a game controller by selecting the object to be deleted with a dedicated controller button.
  • the selected object is then highlighted and additional information appears on the vehicle overlay. If there are multiple detections on the current frame, the detection which has the highest probability that it needs to be deleted (by having the lowest accuracy rating, for example) is selected first although a user can scroll to other detected objects within that frame. The user can accomplish this by scrolling to the next detected object by clicking the same dedicated controller button again.
  • the dedicated controller button could be a shoulder button or rear touch pad as is found on many game controllers now.
  • an object classification user interface indicated generally by reference numeral 400, which allows a user to manually classify objects which have been either manually detected by a user or automatically detected by the tracker system in the multimedia file.
  • the classification interface 400 displays a plurality of objects to be classified 402.
  • Each object to be classified 402 comprises a completed field box 404 and an automatic classification indicator 408. If the automatic classification is accurate enough, it may not be necessary for manual classification, although at present it is currently envisaged to manually classify each detected object.
  • the plurality of objects to be classified 402 are exported by the tracker system as thumbnail images. These thumbnail images are presented in a grid on the classification user interface 400.
  • a classification table 410 is provided on the classification user interface 400.
  • the classification table 410 lists the different type of classifications which may be accorded to the detected objects and may also provide a running count of each classification type. Progress data 408 may also be provided on the classification user interface 400.
  • the user will classify each detected object separately.
  • the selected object can be classified using an input device such as a mouse, a keyboard or a game controller.
  • an input device such as a mouse, a keyboard or a game controller.
  • the user would click on the required class, and the selected object would become the next object to be classified and the use would move through the entire grid of objects to be classified in this manner.
  • a keyboard may be used, with predesignated keys relating to each of the classification types.
  • the classification can be very effective. There can be some situations where it is not easy to recognise the class of the object from the thumbnail image alone. For these situations the classification user interface 400 provides a keyboard shortcut for the user to open a video playback around the frame of the thumbnail so the object to be classified can be seen in a short moving video as this may assist i the classification.
  • the object to be classified will be marked on the video playback using a bounding box overlay in a preferred embodiment.
  • a final verification step is carried out in accordance with step 114 of Figure 1. This is to ensure that the analysis results comply with predefined accuracy levels.
  • the validation algorithm marks a plurality of relatively short time periods randomly selected from the multimedia file and preferably distributed over the length of the multimedia file. A user then manually counts, tracks and classifies the objects in these short time periods and sends the results to the tracker system.
  • the validation algorithm compares the original analysis results with the results of the validation analysis. Using this comparison, the validation algorithm can estimate the accuracy of the original analysis results.
  • the acceptable levels of accuracy of the original analysis results var from market to market, but may be approximately 5% in the United Kingdom and Ireland for the enumeration of vehicles, and, approximately 15% for the classification of vehicles. There is no published standard and different levels of accuracy of the original analysis will be accepted in different jurisdictions.
  • An analysis report is then created as detailed in step 1 16 of Figure 1 and this report can be used to ascertain traffic flow, or object movement in a videoed space/area.
  • the automated detection, counting, tacking and classification of objects in the multimedia file may be accomplished in a number of well-known ways. For example, an algorithm may detect an object by comparing the pixels of one frame to those in an adjacent frame and thus detect movement of objects through a sequence of frames. The shape and size of the objects, along with their trajectory and approximate velocity can be determined when taking into account the parameters input by a user with regard to the camera location and angle and so on.
  • analysis shall be understood to refer to the detection and/or tracking and/or counting and/or classification of objects in a multimedia file, and has been used interchangeably as such with these latter terms.
  • the objects shall be understood to be moving objects, such as vehicles as being the primary example and embodiment described hereinbefore.
  • vehicle shall be afforded the broadest possible scope and shall refer to motorised vehicles such as cars, trucks, vans, lorries, motorbikes, mopeds, scooters, self-balancing wheeled hooverboards and so on, and, also to non-motorised vehicles such as pushbikes, wheelchairs, scooters, skateboards, and so on.
  • motorised vehicles such as cars, trucks, vans, lorries, motorbikes, mopeds, scooters, self-balancing wheeled hooverboards and so on
  • non-motorised vehicles such as pushbikes, wheelchairs, scooters, skateboards, and so on.
  • the invention can also be applied to people and can be adapted to accommodate people pushing trolleys, buggies, pushchairs and so on. It will be understood that the components shown in any of the drawings are not necessarily drawn to scale, and, like parts shown in several drawings are designated the same reference numerals.

Abstract

The present invention is directed to a method of analysing a multimedia file for the detection, counting, tracking and/or classification of objects displayed in images contained within the multimedia file. The method comprises the steps of executing an automated analysis of the multimedia file and then performing a manual analysis of selected portions of the multimedia file. The analysis comprises observing objects seen in images of the multimedia file by detecting, counting, tracking and/or classifying the objects. An accuracy rating for each observed object is produced. When the accuracy rating is below a predefined threshold, that observed object is selected for manual analysis. The advantage of providing such a semi-automated method is that the object analysis of a multimedia file can be carried out in a relative short amount of time, whilst the accuracy of the detections is maintained at a relatively high standard due to the manual intervention in chosen sections of the multimedia file.

Description

"A method of analysing a multimedia file"
Introduction
This invention relates to a method of analysing a multimedia fife, In particular, the present invention is directed towards a method for analysing a multimedia file for purposes of observing objects appearing in the multimedia file by detecting, counting, tracking and/or classifying such objects.
Throughout this specification, the term "multimedia file" shall be understood to encompass any type of multimedia file which may be stored and downloaded for analysis or streamed from a remote storage for analysis or streamed live from a content source location for real-time analysis. The multimedia file may be a video file and such a video file can be interpreted as containing a plurality of images which together make up the video file. It may not, and is likeiy not to, have any audio component.
The multimedia file which is to be analysed is envisaged to be a video file showing a road junction, and the objects to be observed would be vehicles passing through the road junction. In such cases, the multimedia file will be created using content recorded from a traffic camera, closed-circuit television camera or similar type of camera, which records images of the junction to be analysed. A further aspect of the invention envisages observing people as the objects. This could be used in analysing public spaces to determine the flow of people through a pedestrian junction, or monitor the flow of people through public spaces such as stadiums, shopping centres and the like. The primary implementation of the present invention is envisaged to assist with object {e.g. vehicles) movement observation, specifically, object detection, object counting, object tracking and object classification from video footage.
The detection, counting and tracking of vehicles as they pass through a junction can be of great assistance for traffic management purposes. The present invention is able to count where the vehicles travel from, and travel on to as they pass through a specific junction, during a time period, for example 12 hours. A further aspect of the invention is to classify the vehicles as this is useful from a statistical point of view for planning road extensions and road capacity requirements for heavy loads.
It will be readily understood that the term "junction" refers to any roadway, or route which has a plurality of arms, along each of which arms a vehicle may travel and endpoints of the arms meet one another to form the junction. A 3-arm junction is typically known as a T-junction; a 4-arm junction may be a crossroads or a 4-arm roundabout
A vehicle passing through the junction will enter along one arm of the junction and will exit along an arm. It is a main object of the present invention to detect and count the number of vehicles passing through a selected junction and to also track the movement of the vehicles through the junction, during a time period, by analysing the entry and exit arms for each of the vehicles passing through the junction. The classification of the vehicle is also determined. The class of a vehicle may be one of: a car, a bus, a taxi, a truck, a van, a lorry and so on.
Currently in most cases the detecting, counting and classifying of vehicles is carried out entirely manually. In this manner, a traffic survey company will install a camera at a specific junction and a person will be employed to watch the video recorded from that camera. The person will count, track and classify the vehicles. The results are normally written by the person on paper sheets. It is common that the person will have to view the recorded video several times in order to generate accurate results, particularly where the junction is a complicated junction that may have a large number of vehicles passing through it. The person usually has to rewind and watch forward a section of video several times if a number of cars are passing through the junction concurrently. This manual process is extremely time intensive and labour intensive. The efficiency of the process is relatively low and can be prone to human error also.
Automated detection, counting and tracking of vehicles has been considered in the prior art. The objects (e.g. vehicles) shown in the video file can be automatically detected using a tracking algorithm. The tracking algorithm requires some configuration to allow the algorithm to analyse the video and produce the results. Such automated systems are described in U.S. Patent Number US 8,204,955 (MIOVISION TECHNOLOGIES INCORPORATED). US 8,204,955 discloses a method and system which are provided for remotely analysing multimedia content, in particular video.
U.S. Patent Number US 7,460,691 (GET TECHNOLOGIES PTE LIMITED) is a further example of the prior art and discloses image processing techniques which are applied video images for the purposes of acquiring traffic data. The disclosure discusses a traffic monitoring system, the basic function of which is for traffic data acquisition and incident detection and in particular the application of image processing techniques for the detection of a vehicle, from sequence of video images, as well as the acquisition of traffic data and detection of any traffic incidents.
However, there are a number of issues with fully automated systems such as those described in U.S. Patent Number US 8,204,955 and U.S. Patent Number US 7,460,691. Oftentimes, count errors occur due to weather conditions, poor camera views, high density traffic and so on, In such instances, the accuracy of the automated systems drops dramatically and the system is ineffectual as the count errors which are unavoidable due to the weather conditions, poor camera views, high density traffic, etc. are incorporated into the automated analysis. It is also known for automated systems to miss some shapes/sizes of objects (such as small mopeds for example), or, produce alternative errors such as false positives, duplicated detections and so on.
In short, using the fully automated tracking produces several errors which results in a lower accuracy than the manual counting, although the fully automated systems are time efficient. The manual tracking is better from an accuracy point of view but is both labour intensive and time intensive and therefore relatively costly in comparison to the automated process described in the prior art.
It is a goal of the present invention to provide a system and method that overcomes at least one of the above mentioned problems. Symmary of the Invention
The present invention is directed to a method of analysing a multimedia file for the detection, counting, tracking and/or classification of objects displayed in images contained within the multimedia file, wherein the method comprises the steps of executing an automated analysis of the multimedia file, wherein, the automated analysis comprises observing the objects by detecting, counting, tracking and/or classifying the objects in the multimedia file, and producing an accuracy rating for at least each observation of an object; marking chosen sections of the multimedia file, where observations of the objects in those chosen sections have an accuracy rating below a predefined threshold; and, manually observing objects in the chosen sections of the multimedia file.
The advantage of providing such a semi-automated method is that the object analysis of a multimedia file can be carried out in a relative short amount of time, whilst the accuracy of the detections is maintained at a relatively high standard due to the manual intervention in chosen sections of the multimedia file where the automated detection has been deemed to have an accuracy rating below an acceptable predefined threshold. The method also allows for correction of the automated detection on the fly in non-chosen sections of the multimedia file.
In a further embodiment, the accuracy rating is produced separately for the detection, counting, tracking and/or classification of the objects. In a further embodiment, the accuracy rating is an accumulated accuracy rating which is produced based on any combination of the detection, counting, tracking and/or classification of the objects. in a further embodiment, the accuracy rating is calculated on the probability of the detection, counting, tracking, and/or classifying being correct.
In a further embodiment, the probability of the detection, counting, tracking, and/or classifying of the object being correct is based on one or more of: to what extent the motion of the object complies with expected motions of the object, such as kinematic and dynamic capabilities of the object; whether the object has properly approached, entered and exited an internal area, which internal area is specified by a user and is a portion of an image contained in the multimedia file; whether a direction of motion of the object complies with a preset movement path on the image of the multimedia file, which movement path has been defined by the user; whether the object is distinct and separated from other objects by a predefined distance in the image of the multimedia file; and/or, whether a tracked path of the object is distinct from other tracked paths of other objects.
In a further embodiment, the method comprises a further step of producing an object analysis report.
In a further embodiment, the method comprises a further step of manually correcting erroneous detections in non-chosen sections of the multimedia file. In a further embodiment, the method comprises a further step of manually classifying the automatically detected objects and/or the manually detected objects.
In a further embodiment, the step of manually classifying the automatically detected objects and/or the manually detected objects includes displaying a thumbnail image of the object to be classified to a user, and allowing the user to call up a short video of the object to be classified if the thumbnail image is deemed to be insufficient to classify the object.
In a further embodiment, the method comprises a further step of validating portions of the automatically detected objects, the manually detected objects and the manually classified objects, by way of making a comparison with manually detected objects and manually classified objects respectively.
In a further embodiment, the multimedia file is a video file. In a further embodiment, the multimedia file is a video file of a junction.
In a further embodiment, the objects are moving objects. In a further embodiment, the objects are vehicles. In a further embodiment, a user manually detects, counts, tracks and/or classifies objects using a game controller suitable for a games console. in a further embodiment, background areas on images contained in the multimedia file may be selected by a user in advance of executing the automated analysis of the multimedia file such that those selected background areas in the images of the multimedia file can be excluded from the automated analysis.
In a further embodiment, internal areas on images contained in the multimedia file may be selected by a user in advance of executing the automated analysis of the multimedia file so that the selected internal areas in the images of the multimedia file ca be used during the automated analysis to assist in calculating the accuracy rating and/or confirming a valid detection of an object.
In a further embodiment, prior to executing the automated analysis of the multimedia file, analysis parameters for analysing the multimedia file are input by a user.
In a further embodiment, the analysis parameters input by the user comprise one or more of: movement paths, camera settings, time settings, background and internal areas, and, tracker configuration.
The present invention is further directed towards a tracker system for analysing a multimedia file so as to detect, count, track and/or classify objects displayed in images contained within the multimedia file; the tracker system comprising means for executing an automated analysis of the multimedia file, wherein, the automated analysis comprising observing objects in the multimedia file by detecting, counting, tracking and/or classifying the objects in the multimedia file, and means for producing an accuracy rating for each observation of an object; and, means for marking chosen sections of the multimedia file, where observations of the objects in those chosen sections have an accuracy rating below a predefined threshold; and, means to manually detect objects in the chosen sections of the multimedia file.
In a further embodiment, the means for producing an accuracy rating for each observation of the object produces and accuracy rating separately for each of the detection, counting, tracking and/or classification of the objects. In a further embodiment, the means for producing an accuracy rating for each observation of the object produces an accumulated accuracy rating based on a combination of the accuracy ratings for the detection, counting, tracking and/or classification of the objects.
In a further embodiment, the accuracy rating is calculated on the probability of the detection, counting, tracking, and/or classifying being correct. In a further embodiment, the probability of the detection, counting, tracking, and/or classifying of the object being correct is based on one or more of: to what extent the motion of the object complies with expected motions of the object, such as kinematic and dynamic capabilities of the object; whether the object has properly approached, entered and exited an internal area, which internal area is specified by a user and is a portion of an image contained in the multimedia file; whether a direction of motion of the object complies with a preset movement path on the image of the multimedia file, which movement path has been defined by the user; whether the object is distinct and separated from other objects by a predefined distance in the image of the multimedia file; and/or, whether a tracked path of the object is distinct from other tracked paths of other objects.
In a further embodiment, the tracker system further comprises means to manually correct erroneous detections in non-chosen sections of the multimedia file. In a further embodiment, the tracker system further comprises means to manually classify the automatically detected objects and/or the manually detected objects.
In a further embodiment, the means to manually classify the automatically detected objects and/or the manually detected objects comprises means to display a thumbnail image of the object to be classified by a user, and means to allow the user to call up a short video of the object to be classified if the thumbnail image is deemed to be insufficient to classify the object.
In a further embodiment, the tracker system further comprises means to validate portions of the automatically detected objects, the manually detected objects and the manually classified objects being validated by way of a comparison with manually detected and classified objects.
In a further embodiment, the tracker system further comprises means to produce an object analysis report.
In a further embodiment, the multimedia file is a video file. In a further embodiment, the multimedia file is a video file of a junction. In a further embodiment, the objects are moving objects. In a further embodiment, the objects are vehicles.
In a further embodiment, the user manually detects, counts, tracks and/or classifies objects using a game controller suitable for a games console.
In a further embodiment, the tracker system further comprises means for selecting background areas on images contained in the multimedia file in advance of executing the automated analysis of the multimedia file such that those selected background areas in the images of the multimedia file can be excluded from the automated analysis.
In a further embodiment, the tracker system further comprises means to select internal areas on images contained in the multimedia file in advance of executing the automated analysis of the multimedia file so that the selected internal areas in the images of the multimedia file can be used during the automated analysis to assist in calculating an accuracy rating and/or confirming a valid detection of an object.
In a further embodiment, the tracker system further comprises input means to allow a user input analysis parameters for analysing the multimedia file to prior to executing the automated analysis of the multimedia file.
In a further embodiment, the analysis parameters input by the user comprise one or more of: movement paths, camera settings, time settings, background and internal areas, and, tracker configuration. The present invention is further directed to a method of analysing a multimedia file for the detection, counting, tracking and/or classification of objects displayed in images contained within the multimedia file, wherein the method comprises the steps of executing an automated analysis of the multimedia file, wherein, the automated analysis detects, counts, tracks and/or classifies objects in the multimedia file by at least detecting an object, and producing an accuracy rating for at least each detection of an object; marking chosen sections of the multimedia file, where detections of the objects in those chosen sections have an accuracy rating below a predefined threshold; manually detecting objects in the chosen sections of the multimedia file; manually correcting erroneous detections in non-chosen sections of the multimedia file; and producing an object analysis report.
The present invention is further directed to a method of analysing a multimedia file for the detection, counting, tracking and/or classification of objects displayed in images contained within the multimedia file, wherein the method comprises the steps of executing an automated analysis of the multimedia file, wherein, the automated analysis detects, counts, tracks and/or classifies objects i the multimedia file by at least detecting an object, and producing an accuracy rating for at least each detection of an object; marking chosen sections of the multimedia file, where detections of the objects in those chosen sections have an accuracy rating below a predefined threshold; manually detecting objects in the chosen sections of the multimedia file; manually correcting erroneous detections in non-chosen sections of the multimedia file; manually classifying the automatically detected objects and the manually detected objects; and producing an object analysis report.
The present invention is further directed to a method of analysing a multimedia file for the detection, counting, tracking and/or classification of objects displayed in images contained within the multimedia file, wherein the method comprises the steps of executing an automated analysis of the multimedia file, wherein, the automated analysis detects, counts, tracks and/or classifies objects in the multimedia by at least detecting an object, and producing an accuracy rating for at least each detection of an object; marking chosen sections of the multimedia file, where detections of the objects in those chosen sections have an accuracy rating below a predefined threshold; manually detecting objects in the chosen sections of the multimedia file; manually correcting erroneous detections in non-chosen sections of the multimedia file; manually classifying the automatically detected objects and the manually detected objects; validating portions of the automatically detected objects, the manually detected objects and the manually classified objects by way of a comparison with manually detected and classified objects; and producing an object analysis report.
In a further embodiment, the multimedia file is a video file, In a further embodiment, the objects are moving objects. In a further embodiment, the objects are vehicles. In a further embodiment, the multimedia file is a video file of a junction.
In a further embodiment, the accuracy rating is calculated on the probability of the detection, counting, tracking, and/or classifying being correct. In a further embodiment, a user manually detects, counts, tracks and/or classifies objects using a game controller. In a further embodiment, the game controller can be programmed.
In a further embodiment, the manual classification of the automatically detected objects and the manually detected objects includes displaying a thumbnail image of the object to be classified to a user, and further allows a user to call up a short video of the object to be classified if the thumbnail image is deemed to be insufficient to classify the object. In a further embodiment, background areas on images from the multimedia file may be selected by a user in advance of executing the automated analysis of the multimedia file so that background areas in the images of the multimedia file can be excluded from the automated analysis. In a further embodiment, internal areas on images from the multimedia file may be selected b a user in advance of executing the automated analysis of the multimedia file so that the internal areas in the images of the multimedia file can be used during the automated analysis to assist in calculating the accuracy rating and/or confirming a valid detection of an object. In a further embodiment, prior to executing the automated analysis of the multimedia file, analysis parameters for analysing the multimedia file are input by a user. In a further embodiment, the analysis parameters input by the user comprise one or more of: movement paths, camera settings, time settings, background and internal areas, and, tracker configuration.
The present invention is further directed towards a tracker system for analysing a multimedia file so as to detect, count, track and/or classify objects displayed in images contained within the multimedia file, characterised in that the tracker system comprises means for executing an automated analysis of the multimedia file, wherein, the automated analysis detects, counts, tracks and/or classifies objects in the multimedia file in the form of a detection of an object, and produces an accuracy rating for each detection of an object; and, means for marking chosen sections of the multimedia file, where detections of the objects in those chosen sections have an accuracy rating below a predefined threshold; means to manually detect objects in the chosen sections of the multimedia file. The present invention is further directed towards a tracker system for analysing a multimedia file so as to detect, count, track and/or classify objects displayed in images contained within the multimedia file, characterised in that the tracker system comprises means for executing an automated analysis of the multimedia file, wherein, the automated analysis detects, counts, tracks and/or classifies objects in the multimedia file in the form of a detection of an object, and produces an accuracy rating for each detection of an object; means for marking chosen sections of the multimedia file, where detections of the objects in those chosen sections have an accuracy rating below a predefined threshold; means to manually detect objects in the chosen sections of the multimedia file; means to manually correct erroneous detections in non-chosen sections of the multimedia file; means to manually classify the automatically detected objects and the manually detected objects; means to validate portions of the automatically detected objects, the manually detected objects and the manually classified objects being validated by way of a comparison with manually detected and classified objects; and means to produce an object analysis report. Detailed Description of Embodiments
The invention will be more clearly understood from the following description of some embodiments thereof, given by way of example only, with reference to the accompanying drawings, in which:
Figure 1 is a flow diagram detailing some of the steps involved in the present invention;
Figure 2 is an image taken from a multimedia file, whereby, objects appearing in the image are to be analysed in accordance with the present invention; Figure 3 is a screenshot of a user interface for manually counting and tracking objects using a tracker system in accordance with the present invention; and,
Figure 4 is a screenshot of a user interface for manually classifying objects using a tracker system in accordance with the present invention.
Referring to Figure 1 , there is provided a multimedia file analysis method indicated generally by reference numeral 100. The multimedia file analysis method 100 comprises a plurality of steps, 102 to 116, which enable the analysis of objects which appear in the multimedia file.
In step 102, a user will define analysis parameters for analysing a multimedia file.
In step 104, an automated observation of objects displayed in the multimedia file is carried out, based on defined analysis parameters. Figure 1 refers to an automated count as opposed to an automated observation; however, it will be understood that throughout this specification, the term Observation of objects', and its grammatical equivalents, shall be interpreted as detecting objects displayed in one or more images which make up a part of the multimedia file, counting the objects displayed in the one or more images which make up a part of the multimedia file, tracking the objects displayed in the one or more images which make up a part of the multimedia file, and/or, classifying the objects displayed in the one or more images which make up a part of the multimedia file. The automated observation of objects may be carried out over the full length of the multimedia file, which will typically be a video file; or, the automated observation may be carried out during a predefined selected portion of the multimedia file.
In step 106, automated count results from step 104 are transferred for correction and verification by a user.
In step 08 the user will manually count objects, which shall be understood to be manually observing objects, for chosen sections of the multimedia file. The chosen sections will be sections of the multimedia file where the automated observation has a relatively low accuracy rating as the probability of the observations regarding the observed object being correct have been adjudged to be low. The method of the present invention then specifies for these sections to be manually counted.
In step 110, the user will manually correct the automated count in non-chosen sections of the multimedia file, if the user notices that there are any errors in these non-chosen sections when having a brief look through the multimedia file. This step is of course not mandatory and may be omitted in some embodiments of the invention.
In step 1 12, the user manually classifies objects in the multimedia file according to preset criteria, or, objects are automatically classified according to the preset criteria. It will be appreciated that it is foreseen to automatically classify at least some of the objects detected in the multimedia file based on assessing the size and/or shape of the objet detected in the multimedia file. As will be discussed in greater detail hereinbelow, the angle of view which the camera has relative to the objects passing through the field of vision of the camera will be taken into account when determining the criteria for assessing the shape and size of objects appearing in the images of the multimedia file.
In step 114 of the method, the automated and manual analysis results are verified by have a user check randomly chosen portions of the multimedia file. ln the final step 116, an analysis report is created.
It will be understood that steps such as the creation of a report, or the verification of data, or the classification of objects may be omitted.
For the purposes of this description, the implementation of the method will focus on the observation of objects, thus incorporating the detection, counting, tracking and classification of objects appearing in a multimedia file, such as detecting, counting, tracking and classifying vehicles which appear on a video file of a junction with the vehicles passing through the junction. Alternative implementations for this methodology are envisaged for observing the movement of people, animals and other types of transportation at junctions or open public spaces.
It will be further understood that one or more users may be used to complete the steps given hereinbefore; the invention is thus not limited to the same user completing all of the steps. Indeed, in one embodiment of the present invention, it is envisaged that the system will allows manually implemented steps to be carried out by users derived from a crowd sou reed community of workers. In this embodiment, an individual or group of individuals would log into the system, and possibly complete a training exercise, prior to then carrying out the manual detection, counting, tracking, classifying and/or verifying of objects in an assigned multimedia file. The system would assign work to the logged in users and would track the work carried out by each user to ensure it is of a sufficient quality rating before authorising a payment for the user. Now turning to consider each of the steps in more detail.
With reference to Figure 2, there is shown an image indicated generally by reference numeral 200, from a multimedia file (not shown) in which a number of vehicles 202, 204 are seen to be passing through a three-armed junction indicated generally by reference numeral 206. In this case the three-armed junction 206 is a round-about. A background area (discussed hereinbelow) 208 has been defined by a user. An internal area (discussed further below) 210 has also been defined by the user. And, a movement path (discussed hereinunder) 212, which is indicated by a plurality of arrows, has been defined by the user. As per step 102 of Figure 1 , a user will define the analysis parameters for a tracker system which will be used for the multimedia file analysts of the image of Figure 2. The analysis parameters are required by the tracker system, in addition to the multimedia file, to be able to detect the objects, in this case vehicles, in the multimedia file.
The analysis parameters comprise, but are not limited to, one or more of the following: a) Movement routes; each possible movement through the junction is represented as a matrix of movement directions. The matrix of movement directions is shown in a top left comer of Figure 2, where each direction of each arm is accorded a separate reference letter. A vehicle entering the junction from the bottom of the Figure 2 and turning left, from a driver's perspective, will travel through cell C and then pass through cell A, and exit the junction through cell D. If the vehicle entering the junction from the bottom of the Figure 2 turns right off of the round-about, again from a driver's perspective, the vehicle will travel through cell C and then pass through cell A, and exit the junction through cell F. Cell A is the junction/round-about itself. b) Camera settings; a plurality of settings which indicate the position and/or orientation of the camera are used to allow the method to determine the view contained in the multimedia file so that a plane may be established. For example, one or more of: a height of installation of the camera from ground level (e.g. 5m), a field of view of the camera (e.g. 65°), a horizon height which refers to the location of the horizon relative to a chosen border of the camera view (e.g. 23% below the top border), a cell size for the movement matrix (e.g. 1m), and a camera angle which is the angle formed between the direction in which camera faces and the horizontal (e.g. 30° downward) are provided by a user in order to allow the view of the junction to be contextualised by the method, for example, to understand how far into the distance the camera view can see. This can assist with detecting and classifying objects as the size and shape of the objects will vary according to the angle/plane of view of the camera, and the distance towards the horizon which the camera view can see, and hence provides in the multimedia file.
Time settings; frames per second in real time are made known to the method to allow the method to be able to calculate the real velocity (m/s) of the objects appearing in the multimedia file, which will be the velocity of the vehicles passing through the junction.
Background and internal areas; the user can select areas of the image from the multimedia file which can be ignored from the point of view of the analysis. For example, flags or trees fluttering due to wind, or, a pedestrian walking on the footpath may be excluded, if desired by using a background area. The flags/trees could be confused for objects that are to be tracked, and, it may not always be desired to detect and analyse pedestrians in some circumstances. Although it will be understood, that in other embodiments, the roadways and carriageways may be excluded so that the method can focus on analysing pedestrians on the footpaths, if desired. The internal areas can be used to help trigger an observation of an object by setting the parameters such that an object will only be counted, tracked and or classified if it at least enters the internal area, or enters and subsequently leaves the internal area.
Tracker configuration; the method employed by the tracker system can be adjusted to take account of several pre-defined conditions. For example, different video formats will require different configurations due to the difference image contrasts that exist between the different configurations. The tracker system may also be used during different lighting conditions, such as low light at dusk or dawn, or at night time. The tracker system provides the user with a user interface to allow the user to easily input and set these parameters. The background areas, the internal areas and the movement paths can be defined on a drawing area which can be overlaid on an image from the multimedia file. The movement paths can be drawn using a stroke of a brush- type software tool which can define a possible travel path. In this manner, each of the possible travel paths can be input quickly. The software tool may also take account of the angle/plane of the camera when accepting the path drawing stroke so that the depth perspective of the camera view is taken into account by automatically narrowing the path appropriately when the drawing stroke is made towards the horizon. In further embodiments, the background areas and the internal areas may be automatically defined by the tracker system.
As can be seen form Figure 2, the internal area 210 is needed for the tracker system to be able to distinguish between valid vehicles and invalid vehicles. A valid vehicle is defined as an object/vehicle which is initially detected outside of the internal area 210, is detected inside the internal area 210 and hence is considered to pass through the internal area 210 through the internal area 210, and subsequently is detected outside of the internal area 210 again.
With reference now to step 104, the automated count of objects in the multimedia file is executed based on defined analysis parameters from step 102. Firstly, the tracker system exports the analysis parameters set by the user defined in the previous section for the tracker and then executes the tracker system to begin scrolling through the multimedia file and counting the objects shown in the multimedia file. During the tracking the tracker system provides the user with progress information and estimates the total time of execution. This is important as the tracking execution time for a longer video (12 hr) can last many hours. The user can also cancel/restart the tracking at any time.
The tracker runs through the entire multimedia file and runs the object/vehicle detection algorithm to count the objects/vehicle. An object or vehicle is detected by comparing pixels of adjacent frames to detect the presence of an object, and its motion relative to the background pixels which will not substantively change. After the detection is complete the tracker system exports the tracking results. The results includes one or more of: the frame numbers) in the multimedia file (e.g. video file) where the object entered and exited the junction; the detected object movement (e.g. the number of the movement based on the cells of the movement matrix); an estimated accuracy; the detected object class, if the classification is being done automaticall based on the size and/or shape of a cluster of pixels which are determined to represent an object; and/or, a screenshot of each of the objects for manual classification, if manual classification is being done. The estimated accuracy is the probability that the detected object is actually an object. The probability is determined by assessing:
- to what extent the motion of the object comp!ies with expected movements of the object. The expected movements could be based on the kinematic and dynamic capabilities of the object. Taking the example of a vehicle, the vehicle will have a maxima! acceleration level, a constraint against motion sideways without turning, and so on;
- whether the object has properly approached and entered the internal area 210, exited the internal area 210, and did not re-enter the internal area 210 again;
- whether a direction of motion of the object complies with the movement path 212 defined by the user;
- whether the object is distinct from other objects (so as to avoid, for example, counting a front cab and a rear trailer of a truck as two separate vehicies); - whether a tracked path of the object is distinct from other tracked paths of other objects, so as to determine whether there are multiple objects which are occluding each other for a period of time;
- whether the object has a size and/or shape which is within an expected range of sizes and/or set of expected shape types, it could be that a saloon car, an estate car, a SUV, a track, a van, a motorbike, a pushbike, and so on all have different expected sizes and shapes and the detected size and shape would be compared to the expected sizes and shapes to determine if one of them substantially matches. As the plane of view of the camera will be known, the position of the object on the image can be used to determine the distance of the object from the camera as this will clearly have a bearing on the expected size and shape of the object;
- whether the object is observed to move in a manner consistent with the known plane of view of the camera. Such that, the object size should reduce if the object moves towards a horizon and vice versa; and/or,
- whether the object enters and leaves the cells of the movement matrix in an expected direction(s).
The estimated accuracy is an important aspect of the invention and as many information sources (such as information provided by the user, information retrieved by the tracking system, and, information extracted from the images) are used to determine the estimated accuracy of the observation of the object in successive images from the camera.
As mentioned above, the tracker system estimates the accuracy of the observation of an object by estimating the probability that the object has been detected, counted, tracked and/or classified correctly. The tracker will mark chosen sections of the multimedia file, in the form of time ranges during the multimedia file, where the tracker was not able to observe the object with a high accuracy rating, as compared with an acceptable preset threshold. These are the sections of the multimedia file where there are either objects with low accuracy ratings (doubtful cases), or objects which do not seem to be vehicles but nonetheless have a higher accuracy than conventional noise (possible noise cases). These possible noise cases are where the tracker system definitely sees motion but it is not clearly an expected object, such as a vehicle. For example, traffic jams cause issues as there will be many objects moving slowly and stopping behind each other; the tracker system thus detects a single, large object moving in complex ways and the tracker system will be unable to identify the motion of individual objects. However, the tracker system will understand that there are vehiclelike objects moving and the system needs to ask for the user's manual assistance. These chosen sections of the multimedia file will then be processed manually in accordance with step 108.
In a preferred embodiment, the chosen section of the multimedia file will be grouped in accordance with the arms, or cells, of the junction under review. In this manner, it is easier for a manual count of the chosen sections of the multimedia file to be done for several objects all emanating from the same arm on the junction.
The tracker also determines a number in the form of a percentage value, which is referred to as a certainty ratio, for each arm of the junction. This certainty ratio indicates how much of the multimedia file requires manual processing, which is an indicator of the quality of the automated analysis. This ratio is calculated based on the number of chosen sections of the multimedia file, the duration of the chosen sections and the distribution of the chosen sections. If the chosen sections are relatively long and are distributed relatively far from each other, then the user of the tracker system will not have any issues to operate the system as the user can simply use the time ratio of the chosen intervals, as chosen, to manually process the chosen sections. However, if the chosen intervals are relative short and are relatively close to one another, then the user cannot easily manually process two seconds of video, then skip one second of video, then manually process three seconds of video, and so on. So before reporting the final chosen sections and calculating the certainty ratio, the tracker system merges chosen sections which are close to each other, even though there will be a short certain section between the two chosen sections. The user can set a parameter, such as five seconds, which determines if chosen sections are too close to one another and need to be merged to make the manual processing easier for the user. In this manner, if the distribution of the chosen intervals is many relative short sections that are relative close to each other, the tracker system will merge these chosen sections (and any short certain sections falling therebetween) into a single longer, continuous section for manual processing. It should be noted that if an object or vehicle is correctly observed by the tracker system in a short certain section which is subsequently merged with chosen sections for manual processing, then the count, classification, detection and so on for that vehicle is removed so as to avoid double counting.
All of the results of the automated analysis are transfer, following step 106, for correction and verification by a user.
In step 108, a user will manually count objects for the chosen sections of video, which had a relatively low accuracy estimated value. In this manner, any objects which the tracker system was uncertain about are manual counted by a user. Moreover, the user can correct non-chosen sections of the multimedia file and correct any errors that may be present, even though the tracker system would have evaluated these non-chosen sections to have a sufficiently high accuracy rating. As the multimedia is played back to the user, during these non-chosen sections a user can still override the automated result and correct errors in this way. After the automated analysis has been executed, the manual analysis may be carried out. The manual analysis consists of the following tasks:
i) Manually observe, including one or more of: detect, count, track and/or classify objects in the chosen section(s) of the multimedia file which has been analysed automatically already;
ii) Correct any errors that may be present in the non-chosen sections of the multimedia file; and,
iii) Manually classify both automatically detected vehicles and manually detected vehicles.
The first two tasks can be completed on the same screen, which is shown in Figure 3. The third task is completed on a separate screen as shown in Figure 4.
There are different of errors which can occur during the automated analysis of a multimedia file. The types of known errors comprise:
a) Duplication: this is where a single object is detected as two objects. This is typical where the objects may be a vehicle with a trailer. In this situation the user needs to delete one of the detections.
b) Incorrectly detected origin arm or incorrectly detected destination arm: in such cases the user needs to delete the detection and add the vehicle manually with the correct origin arm and the correct destination arm. c) The detected and tracked paths of two different objects become either joined or switched. For example, if two vehicles are moving in the opposite direction to one another, and they meet at the center of the junction, the tracker system may interpret this as being two U-turns. In such cases, the user needs to manually delete both detections and add the two vehicles correctly.
d) Detect noise as a valid vehicle. For example, pedestrians, cyclists, trees, flags and other items may cause false detections which must then be manually deleted by the user.
Referring now to Figure 3, the correction user interface indicated generally by reference numeral 300, which allows a user to manually analyse objects in the chosen section(s) of the multimedia file which has been analysed automatically already, and, correct any errors that may be present in the non-chosen sections of the multimedia file is shown. The correction interface 300 comprises a main panel 302 which contains the video of the junction with an overlay of detections and the thumbnail of the junction in the upper left corner of the main panel 302, which indicates the cells of the matrix which represent the arms of the junction, which is the given example are represented by reference letters A to G. A progress bar indicated generally by reference numeral 304 is provided below the main panel 302. The progress bar 304 indicated the overall length of playback of the multimedia file, and where the user currently is within that playback period. Chosen and non-chosen sections of the multimedia file can be differentiated along the progress bar 304 using different colours or the like. A classification table 306 is provided to one side of the main panel 302. The classification table 306 lists the different type of classifications which may be accorded to the detected objects and may also provide a running count of each classification type. An origin arm and destination arm indicator 308 is provided on the correction user interface 300 also. As described hereinbefore, playback will be carried out for one origin arm at a time, so for the most part the origin arm will be locked. If the user changes the origin arm on the origin arm and destination arm indicator 308, then the progress bar 304 and the main panel 302 will be updated to reflect the relevant information, such as the chosen and non-chosen sections for the newly selected origin arm. Information on the frame number of the multimedia file, and a recording time, may be provided as indicated by reference numeral 310.
Within the main panel 302, each detected object 312, which in this embodiment will be a vehicle, is displayed on the main panel 302 with an overlay of a bounding box 314, a detected and tracked path 316 for that object and additional information such as the origin arm/destination arm pair reference letters 318 and the object ID number for that object 320 which is useful for identification purposes. The bounding box 314 and the detected and tracked path 316 of the object 312 is visible and may be colour coded to differentiate numerous objects from one another when there are several objects being shown together on the same frame. In this manner, the user is able to quickly understand and recognise which objects have been detected and what is the detected and tracked path for that object is. Any errors are easily decipherable and the user can take manual correction action as required. Where the object has not been detected at ail, such as during the chosen sections of the multimedia file, the user can add the object and classify the object, at the same time, using just one input device. During the chosen sections of the multimedia file, the user can be altered by way of a red border, flashing border or some other audio, visual or tactile alarm so as to encourage the user to pay particular attention during those periods.
Therefore, as can be ascertained from the above, the user will carry out the following steps of playing the multimedia file for a selected origin arm. During playback for the selected origin arm, the user will manually count and classify the vehicles not have been detected, and, correct the tracker system errors. The user repeats this process for all of the origin arms at the junction.
The manual analysis of an object requires the user to select the destination arm of the vehicle, bearing in mind that the origin arm will be fixed, and, select the class of the object. These selections can be made using any known input devices such as a mouse, by clicking on the appropriate destination arm button and the number next to the selected class for the object; using a keyboard in a similar fashion by way of shortcuts with predefined keys, or with a game controller which is used with a games console. The game controller is seen as a preferred input device as the inputs can be made in a very quick and ergonomic manner. For example, the destination arm can be selected using a D-pad or control stick, by pushing the D-pad or control stick in one of the four available directions with can be associated with the different arms. Up to eight different arms could be accommodated using such a D-pad or control stick using the horizontal, vertical and diagonal movements of the D-pad or control stick.
The class of the object can also be specified with the game controller by using the buttons on the game controller, whereby the different buttons are each associated with one of the different classes of object. As the correction of errors, may require a deletion of a detected object, the deletion can be executed with mouse by right clicking for example, or by using a keyboard with a key such as the 'delete' key, or by using a game controller by selecting the object to be deleted with a dedicated controller button. The selected object is then highlighted and additional information appears on the vehicle overlay. If there are multiple detections on the current frame, the detection which has the highest probability that it needs to be deleted (by having the lowest accuracy rating, for example) is selected first although a user can scroll to other detected objects within that frame. The user can accomplish this by scrolling to the next detected object by clicking the same dedicated controller button again. The dedicated controller button could be a shoulder button or rear touch pad as is found on many game controllers now.
With reference now to step 112 in Figure 1 and to Figure 4, there is provided in Figure 4 an object classification user interface indicated generally by reference numeral 400, which allows a user to manually classify objects which have been either manually detected by a user or automatically detected by the tracker system in the multimedia file. The classification interface 400 displays a plurality of objects to be classified 402. Each object to be classified 402 comprises a completed field box 404 and an automatic classification indicator 408. If the automatic classification is accurate enough, it may not be necessary for manual classification, although at present it is currently envisaged to manually classify each detected object. The plurality of objects to be classified 402 are exported by the tracker system as thumbnail images. These thumbnail images are presented in a grid on the classification user interface 400. A classification table 410 is provided on the classification user interface 400. The classification table 410 lists the different type of classifications which may be accorded to the detected objects and may also provide a running count of each classification type. Progress data 408 may also be provided on the classification user interface 400.
In use, the user will classify each detected object separately. The selected object can be classified using an input device such as a mouse, a keyboard or a game controller. With a mouse, the user would click on the required class, and the selected object would become the next object to be classified and the use would move through the entire grid of objects to be classified in this manner. Alternatively a keyboard may be used, with predesignated keys relating to each of the classification types. With these shortcuts the classification can be very effective. There can be some situations where it is not easy to recognise the class of the object from the thumbnail image alone. For these situations the classification user interface 400 provides a keyboard shortcut for the user to open a video playback around the frame of the thumbnail so the object to be classified can be seen in a short moving video as this may assist i the classification. The object to be classified will be marked on the video playback using a bounding box overlay in a preferred embodiment.
Once the multimedia file has been thoroughly analysed and all of the objects have been detected, counted, tracked and/or classified, a final verification step is carried out in accordance with step 114 of Figure 1. This is to ensure that the analysis results comply with predefined accuracy levels. The validation algorithm marks a plurality of relatively short time periods randomly selected from the multimedia file and preferably distributed over the length of the multimedia file. A user then manually counts, tracks and classifies the objects in these short time periods and sends the results to the tracker system. The validation algorithm compares the original analysis results with the results of the validation analysis. Using this comparison, the validation algorithm can estimate the accuracy of the original analysis results. The acceptable levels of accuracy of the original analysis results var from market to market, but may be approximately 5% in the United Kingdom and Ireland for the enumeration of vehicles, and, approximately 15% for the classification of vehicles. There is no published standard and different levels of accuracy of the original analysis will be accepted in different jurisdictions. An analysis report is then created as detailed in step 1 16 of Figure 1 and this report can be used to ascertain traffic flow, or object movement in a videoed space/area. A skilled person in the art will understand that the automated detection, counting, tacking and classification of objects in the multimedia file may be accomplished in a number of well-known ways. For example, an algorithm may detect an object by comparing the pixels of one frame to those in an adjacent frame and thus detect movement of objects through a sequence of frames. The shape and size of the objects, along with their trajectory and approximate velocity can be determined when taking into account the parameters input by a user with regard to the camera location and angle and so on.
The terms "comprise" and "include", and any variations thereof required for grammatical reasons, are to be considered as interchangeable and accorded the widest possible interpretation.
Throughout the preceding specification, the term "analysis" shall be understood to refer to the detection and/or tracking and/or counting and/or classification of objects in a multimedia file, and has been used interchangeably as such with these latter terms. The objects shall be understood to be moving objects, such as vehicles as being the primary example and embodiment described hereinbefore.
It will be understood that throughout the preceding specification, the term "vehicle" shall be afforded the broadest possible scope and shall refer to motorised vehicles such as cars, trucks, vans, lorries, motorbikes, mopeds, scooters, self-balancing wheeled hooverboards and so on, and, also to non-motorised vehicles such as pushbikes, wheelchairs, scooters, skateboards, and so on. As discussed hereinbefore, the invention can also be applied to people and can be adapted to accommodate people pushing trolleys, buggies, pushchairs and so on. It will be understood that the components shown in any of the drawings are not necessarily drawn to scale, and, like parts shown in several drawings are designated the same reference numerals.
The invention is not limited to the embodiments hereinbefore described which may be varied in both construction and detail

Claims

A method of analysing a multimedia file for the detection, counting, tracking and/or classification of objects displayed in images contained within the multimedia file, wherein the method comprises the steps of executing an automated analysis of the multimedia file, wherein, the automated analysis comprises observing the objects by detecting, counting, tracking and/or classifying the objects in the multimedia file, and producing an accuracy rating for at least each observation of an object; marking chosen sections of the multimedia file, where observations of the objects in those chosen sections have an accuracy rating below a predefined threshold; and, manually observing objects in the chosen sections of the multimedia file.
A method of analysing a multimedia file as claimed in claim , wherein, the accuracy rating is produced separately for the detection, counting, tracking and/or classification of the objects.
A method of analysing a multimedia file as claimed in claim 1 , wherein, the accuracy rating is an accumulated accuracy rating which is produced based on any combination of the detection, counting, tracking and/or classification of the objects.
A method of analysing a multimedia file as claimed in any preceding claims, wherein, the accuracy rating is calculated on the probability of the detection, counting, tracking, and/or classifying being correct.
A method of analysing a multimedia file as claimed in claim 4, wherein, the probability of the detection, counting, tracking, and/or classifying of the object being correct is based on one or more of:
- to what extent the motion of the object complies with expected motions of the object, such as kinematic and dynamic capabilities of the object;
- whether the object has properly approached, entered and exited an internal area, which internal area is specified by a user and is a portion of an image contained in the multimedia file;
- whether a direction of motion of the object complies with a preset movement path on the image of the multimedia file, which movement path has been defined by the user;
- whether the object is distinct and separated from other objects by a predefined distance in the image of the multimedia file; and/or,
- whether a tracked path of the object is distinct from other tracked paths of other objects.
A method of analysing a multimedia file as claimed in any of the preceding claims, wherein, the method comprises a further step of producing an object analysis report.
A method of analysing a multimedia file as claimed in any of the preceding claims, wherein, the method comprises a further step of manually correcting erroneous detections in non-chosen sections of the multimedia file.
A method of analysing a multimedia file as claimed in any of the preceding claims, wherein, the method comprises a further step of manually classifying the automatically detected objects and/or the manually detected objects.
A method of analysing a multimedia file as claimed in any of claim 8, wherein, the step of manually classifying the automatically detected objects and/or the manually detected objects includes displaying a thumbnail image of the object to be classified to a user, and allowing the user to call up a short video of the object to be classified if the thumbnail image is deemed to be insufficient to classify the object.
A method of analysing a multimedia file as claimed in any of the preceding claims, wherein, the method comprises a further step of validating portions of the automatically detected objects, the manually detected objects and the manually classified objects, by way of making a comparison with manually detected objects and manually classified objects respectively.
A method of analysing a multimedia file as claimed in any preceding claims, wherein, the multimedia file is a video file.
A method of analysing a multimedia file as claimed in any preceding claims, wherein, the multimedia file is a video file of a junction.
A method of analysing a multimedia file as claimed in any preceding claims, wherein, the objects are moving objects.
A method of analysing a multimedia file as claimed in any preceding claims, wherein, the objects are vehicles.
A method of analysing a multimedia file as claimed in any preceding claims, wherein, a user manually detects, counts, tracks and/or classifies objects using a game controller suitable for a games console.
A method of analysing a multimedia file as claimed in any preceding claims, wherein, background areas on images contained in the multimedia file may be selected by a user in advance of executing the automated analysis of the multimedia file such that those selected background areas in the images of the multimedia file can be excluded from the automated analysis.
A method of analysing a multimedia file as claimed in any preceding claims, wherein, internal areas on images contained in the multimedia file may be selected by a user in advance of executing the automated analysis of the multimedia file so that the selected internal areas in the images of the multimedia file can be used during the automated analysis to assist in calculating the accuracy rating and/or confirming a valid detection of an object.
18. A method of analysing a multimedia file as claimed in any preceding claims, wherein, prior to executing the automated analysis of the multimedia file, analysis parameters for analysing the multimedia file are input by a user.
19. A method of analysing a multimedia file as claimed in claim 18, wherein, the analysis parameters input by the user comprise one or more of: movement paths, camera settings, time settings, background and internal areas, and, tracker configuration.
A tracker system for analysing a multimedia file so as to detect, count, track and/or classify objects displayed in images contained within the multimedia file; the tracker system comprising means for executing an automated analysis of the multimedia file, wherein, the automated analysis comprising observing objects in the multimedia file by detecting, counting, tracking and/or classifying the objects in the multimedia file, and means for producing an accuracy rating for each observation of an object; and, means for marking chosen sections of the multimedia file, where observations of the objects in those chosen sections have an accuracy rating below a predefined threshold; and, means to manually detect objects in the chosen sections of the multimedia file.
A tracker system for analysing a multimedia file as claimed in claim 20, wherein, the means for producing an accuracy rating for each observation of the object produces and accuracy rating separately for each of the detection, counting, tracking and/or classification of the objects.
A tracker system for analysing a multimedia file as claimed in claim 20, wherein, the means for producing an accuracy rating for each observation of the object produces an accumulated accuracy rating based on a combination of the accuracy ratings for the detection, counting, tracking and/or classification of the objects.
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 22, wherein, the accuracy rating is calculated on the probability of the detection, counting, tracking, and/or classifying being correct.
A tracker system for analysing a multimedia fife as claimed in claim 23, wherein, the probability of the detection, counting, tracking, and/or classifying of the object being correct is based on one or more of;
- to what extent the motion of the object complies with expected motions of the object, such as kinematic and dynamic capabilities of the object;
- whether the object has properly approached, entered and exited an internal area, which internal area is specified by a user and is a portion of an image contained in the multimedia file;
- whether a direction of motion of the object complies with a preset movement path on the image of the multimedia file, which movement path has been defined by the user;
- whether the object is distinct and separated from other objects by a predefined distance in the image of the multimedia file; and/or,
- whether a tracked path of the object is distinct from other tracked paths of other objects.
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 24, wherein, the tracker system further comprises means to manually correct erroneous detections in non-chosen sections of the multimedia file.
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 25, wherein, the tracker system further comprises means to manually classify the automatically detected objects and/or the manually detected objects,
A tracker system for analysing a multimedia file as claimed in claim 26, wherein, the means to manually classify the automatically detected objects and/or the manually detected objects comprises means to display a thumbnail image of the object to be classified by a user, and means to allow the user to call up a short video of the object to be classified if the thumbnail image is deemed to be insufficient to classify the object.
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 27, wherein, the tracker system further comprises means to validate portions of the automatically detected objects, the manually detected objects and the manually classified objects being validated by way of a comparison with manually detected and classified objects.
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 23, wherein, the tracker system further comprises means to produce an object analysis report.
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 29, wherein, the multimedia file is a video file.
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 30, wherein, the multimedia file is a video file of a junction.
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 31 , wherein, the objects are moving objects.
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 32, wherein, the objects are vehicles.
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 33, wherein, the user manually detects, counts, tracks and/or classifies objects using a game controller suitable for a games console,
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 34, wherein, the tracker system further comprises means for selecting background areas on images contained in the multimedia file in advance of executing the automated analysis of the multimedia file such that those selected background areas in the images of the multimedia file can be excluded from the automated analysis.
A tracker system for analysing a multimedia file as claimed in any of claims 20 to 35, wherein, the tracker system further comprises means to select internal areas on images contained in the multimedia file in advance of executing the automated analysis of the multimedia file so that the selected internal areas in the images of the multimedia file can be used during the automated analysis to assist in calculating an accuracy rating and/or confirming a valid detection of an object.
A tracker system for analysing a multimedia file as claimed in any of claims
20 to 36, wherein, the tracker system further comprises input means to allow a user input analysis parameters for analysing the multimedia file to prior to executing the automated analysis of the multimedia file.
A tracker system for analysing a multimedia file as claimed in claim 37, wherein, the analysis parameters input by the user comprise one or more of: movement paths, camera settings, time settings, background and internal areas, and, tracker configuration.
PCT/EP2016/056064 2015-03-19 2016-03-18 A method of analysing a multimedia file WO2016146847A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
IES20150085 2015-03-19
IES2015/0085 2015-03-19

Publications (1)

Publication Number Publication Date
WO2016146847A1 true WO2016146847A1 (en) 2016-09-22

Family

ID=55637347

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2016/056064 WO2016146847A1 (en) 2015-03-19 2016-03-18 A method of analysing a multimedia file

Country Status (1)

Country Link
WO (1) WO2016146847A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1921581A1 (en) * 2006-11-08 2008-05-14 Smarteree GmbH Method and apparatus for monitoring a moving object
US7460691B2 (en) 1999-11-03 2008-12-02 Cet Technologies Pte Ltd Image processing techniques for a video based traffic monitoring system and methods therefor
US20090244291A1 (en) * 2008-03-03 2009-10-01 Videoiq, Inc. Dynamic object classification
US8204955B2 (en) 2007-04-25 2012-06-19 Miovision Technologies Incorporated Method and system for analyzing multimedia content

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7460691B2 (en) 1999-11-03 2008-12-02 Cet Technologies Pte Ltd Image processing techniques for a video based traffic monitoring system and methods therefor
EP1921581A1 (en) * 2006-11-08 2008-05-14 Smarteree GmbH Method and apparatus for monitoring a moving object
US8204955B2 (en) 2007-04-25 2012-06-19 Miovision Technologies Incorporated Method and system for analyzing multimedia content
US20090244291A1 (en) * 2008-03-03 2009-10-01 Videoiq, Inc. Dynamic object classification

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
COIFMAN B ET AL: "A real-time computer vision system for vehicle tracking and traffic surveillance", TRANSPORTATION RESEARCH. PART C, EMERGING TECHNOLOGIES, PERGAMON, NEW YORK, NY, GB, vol. 6C, no. 4, 1 August 1998 (1998-08-01), pages 271 - 288, XP008088782, ISSN: 0968-090X, DOI: 10.1016/S0968-090X(98)00019-9 *

Similar Documents

Publication Publication Date Title
Siebert et al. Detecting motorcycle helmet use with deep learning
Messelodi et al. Intelligent extended floating car data collection
JP6180482B2 (en) Methods, systems, products, and computer programs for multi-queue object detection and analysis (multi-queue object detection and analysis)
US8855361B2 (en) Scene activity analysis using statistical and semantic features learnt from object trajectory data
CN103069434B (en) For the method and system of multi-mode video case index
CN103824452B (en) A kind of peccancy parking detector based on panoramic vision of lightweight
US9014429B2 (en) Analyzing a segment of video
Zhang et al. Automated detection of grade-crossing-trespassing near misses based on computer vision analysis of surveillance video data
KR101375583B1 (en) Object Density Estimation in Video
CN106297278A (en) A kind of method and system shedding thing vehicle for inquiry
Ismail Application of computer vision techniques for automated road safety analysis and traffic data collection
CN113674523A (en) Traffic accident analysis method, device and equipment
US20200034626A1 (en) Method for recognizing an object of a mobile unit
CN113515968A (en) Method, device, equipment and medium for detecting street abnormal event
Reyad et al. Traffic conflict–based before–after study with use of comparison groups and the empirical Bayes method
Malinovskiy et al. Model‐free video detection and tracking of pedestrians and bicyclists
Suttiponpisarn et al. Detection of wrong direction vehicles on two-way traffic
Koetsier et al. Trajectory extraction for analysis of unsafe driving behaviour
WO2016146847A1 (en) A method of analysing a multimedia file
US20230048304A1 (en) Environmentally aware prediction of human behaviors
JP6997471B2 (en) Information processing system, information processing device, terminal device, server device, program, or method
Ismail et al. Automated pedestrian safety analysis using video data in the context of scramble phase intersections
Modi Counting Empty Parking Spots at Truck Stops Using Computer Vision.
MUNAJAT et al. A NEW METHOD FOR ANALYZING CONGESTION LEVELS BASED ON ROAD DENSITY AND VEHICLE SPEED.
US11727580B2 (en) Method and system for gathering information of an object moving in an area of interest

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16712008

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16712008

Country of ref document: EP

Kind code of ref document: A1