EP3449463A1 - Motion analysis system and motion tracking system comprising same of moved or moving objects that are thermally distinct from their surroundings - Google Patents
Motion analysis system and motion tracking system comprising same of moved or moving objects that are thermally distinct from their surroundingsInfo
- Publication number
- EP3449463A1 EP3449463A1 EP17736558.2A EP17736558A EP3449463A1 EP 3449463 A1 EP3449463 A1 EP 3449463A1 EP 17736558 A EP17736558 A EP 17736558A EP 3449463 A1 EP3449463 A1 EP 3449463A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- camera
- unit
- motion analysis
- thermal
- model
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
- G06T7/251—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/292—Multi-camera tracking
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/12—Edge-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/174—Segmentation; Edge detection involving the use of two or more images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/194—Segmentation; Edge detection involving foreground-background segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/564—Depth or shape recovery from multiple images from contours
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/90—Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/60—Editing figures and text; Combining figures or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/04—Indexing scheme for image data processing or generation, in general involving 3D image data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/08—Indexing scheme for image data processing or generation, in general involving all processing steps from image acquisition to 3D model generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10048—Infrared image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30108—Industrial image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30221—Sports video; Sports image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30232—Surveillance
Definitions
- Motion analysis system and a comprehensive motion tracking system of moving or moving, thermally contrasting from their environment objects
- the present invention relates to a motion analysis system and a motion tracking system comprising moving or moving objects that are thermally different from their surroundings.
- movement analyzes are carried out on living objects such as humans in order to improve biomechanics in medicine or sport and to uncover weak points in a movement sequence.
- a comparable goal is pursued in industrial objects with the analysis of the movements of robot arms or similar grippers.
- the basis of any movement analysis is the reliable detection, in particular of distance and angle / orientation data, if possible in real time.
- the object to be analyzed is provided with several marker elements.
- the data acquisition takes place with the aid of video cameras which record the changes in position of the marker elements attached to the musculoskeletal system of the object by means of continuous digital storage of 2D video images by means of at least one video image recorder and make these recordings available to a data processing system for evaluation.
- One difficulty with these applications is tracking the movement of each individual marker element in the 2D video image in real time and automatically assigning it a unique identity.
- a system suitable for this purpose was known commercially under the name Aktisys® and described in detail in WO 2011/141531 AI.
- Markerless systems In addition to marker-based systems, there is also a high demand in the field of motion analysis and tracking for flexibly deployable systems that can do without attaching additional markers to the target object.
- moving objects in the background of video images such as spectators at a sporting event or trees moving in the wind, etc .
- illumination intensities such as moving light sources (headlights), moving shadow sources (clouds) or reflections from moving surfaces (water) etc .
- headlights moving light sources
- shadow sources clouds
- reflections from moving surfaces water etc .
- insufficient illumination intensities such as in particular in caves, chambers, or in twilight / night shots etc.
- the object on which the present invention is based is achieved by a motion analysis system and a movement tracking system comprising moving or moving objects that are thermally lifted from their surroundings with the features of the independent patent claims 1 and 12, respectively.
- a motion analysis system is characterized by a camera group with at least one thermal imaging camera, a calibration unit, a synchronization unit, a segmentation unit, a reconstruction unit, a projection unit and an identification unit.
- a motion tracking system comprises such a motion analysis system and is characterized by a motion tracking unit which realigns a model of the object (s) from associated correspondences.
- the present invention opens up hitherto closed application possibilities for motion analysis and / or tracking systems of interest here, in particular in the areas of sports competition analysis, security technology and animal research:
- sport science has the problem that it analyzes movements primarily in laboratories, but not where the sports movements actually take place: under competitive conditions and in the open field.
- markerless acquisition and thermal based segmentation as with the present invention. teaches, it is now possible for the first time to analyze the exact biomechanics of, for example, a football player at the moment of his cruciate ligament injury and to follow its movement. This and similar information is of high relevance to sport, in particular with regard to explanation and illustration of performance and injury issues.
- thermographic analysis provides the opportunity to directly determine the average skin temperature during a sport activity, as well as to map the muscle groups involved in a movement.
- physiological processes of thermoregulation of the body can not only be tracked directly, but also sports-specific questions can be considered in detail during the course of physical activity.
- thermographic analysis is primarily used to detect local sources of inflammation. Since the heat generation and radiation of a healthy body is relatively symmetrical, deviations from this symmetry may indicate injuries and possibly diseases. Thus, for example, diseased blood vessels, the formation of certain cancer cells, dysfunction of the thyroid gland, but also fractures or, in a comparatively reduced heat radiation, circulatory disorders can be detected in thermographic images.
- FIG. 1 shows schematically by way of example a flow analysis diagram of a movement analysis and / or tracking system according to the invention; for example, the arrangement of a first group of cameras, which comprises at least one thermal imaging camera and at least two video image cameras;
- FIG. 3 shows, by way of example, the arrangement of a second group of cameras, which comprises at least two thermal imaging cameras and optionally video image cameras;
- FIG. 4 shows by way of example the arrangement of a third group of cameras, which exclusively comprises two or more thermal imaging cameras; and
- FIG. 5 shows, by way of example, the arrangement of a fourth group of cameras, which comprises a multiplicity of heat and video image cameras arranged arbitrarily with one another.
- FIG. 1 shows (distributed over three sheets: FIGS. 1 a, 1 b, 1 c) a movement analysis and / or tracking system 1 according to the invention by way of example with reference to a flowchart.
- video image cameras 21, 22,... Refer to devices which record electromagnetic radiation in the visible light range (wavelength range from 400 to 780 nm) by means of special detectors (video image pickup 20) and from the received electrical signals Generating signals (two-dimensional) 2D video images VB These 2D video images VB are then in the form of pixel graphics.
- a pixel graphic is a computer readable form of describing an image in which the pixels are arranged in a raster pattern and each pixel is assigned a pixel value.
- the pixels are typically assigned a particular color (a particular wavelength of visible light) as the pixel value.
- voxel volumetric pixel
- 3D voxel models 93 of the object or objects 90 are made up of segmented 2D pixel regions 91; 92 is reconstructed from recordings of at least two cameras by means of a reconstruction unit 62 according to the invention, i. designed.
- the design (reconstruction) of a 3D voxel model 93 of segmented 2D pixel regions is sometimes referred to as "space carving".
- thermal imaging cameras 1 1, 12,... Designate devices which record electromagnetic radiation in the infrared range ("heat radiation”, wavelength range: 780 nm to 1 mm), which is particularly of living objects (humans, animals ) is emitted (emitted).
- the pixel values of the resulting 2D Thermal images WB represent temperature values, which can advantageously also be used for the plausibility of depth information in the 2D heat WB and / or 2D video image VB.
- the invention makes use of the fact that the body temperature of living objects (humans, animals) normally differs from the temperature of inanimate objects 90 in the environment, so that silhouettes become more vivid from thermal images via so-called thresholding Extracting objects well from the inanimate environment (as a background), environmental factors such as the exposure conditions, the color similarity between the object and the background as well as shading, as they are disturbing for video cameras, are irrelevant in thermal imaging cameras.
- silhouettes can be advantageously distinguished from the background, in particular by calibrating the temperature range of the image recording to a narrow range around the respective body temperature.
- the movement analysis system is initially characterized by a group of cameras 11, 12, ..., arranged arbitrarily to one another. 21, 22, ... that the field of view 111, 121, ...; 211, 221, ... of each camera 11, 12, ...; 21, 22, ... with the field of view 111, 121, ...; 211, 221, ... at least one other camera 11, 12, ...; 21, 22, ... of the group intersects so that the set of all fields of view 111, 121,: 211, 211, ... is at least indirectly connected, and
- the camera group comprises at least one first and second camera whose objectives 112, 122, ...; 212, 222, ... arranged at a distance x of at least two meters and / or their optical axes 113, 123, ...; 213, 223, ... are aligned with each other at an angle ⁇ of at least 45 °,
- the first camera is a thermal imaging camera 11 for recording thermal radiation by continuous digital storage of 2D Thermal images WB by means of at least one thermal image niehehmers 10, and
- a video image camera 21 is for recording light radiation by continuous digital storage of 2D video images VB by means of at least one video image pickup 20.
- a calibration unit 51 is - for example, according to known prior art - a simultaneous 3D spatial calibration of all heat- 1 1, 12, ... and possibly video cameras 21, 22, ... with overlapping fields of view 1 1 1, 121, 21 1, 221, ... ensured.
- a synchronization unit 52 ensures that the recording, ie the continuous digital storage of 2D heat WB and any 2D video images VB takes place at the same time and / or the recording times of the 2D heat WB and any 2D video images VB are known , In the case of a mixed camera group comprising thermal heat 1 1, 12,... And video image cameras 21, 22, the frequency of the video cameras 21, 22,... Has proven to be an integer multiple of the frequency of the thermal cameras 11 , 12, ... to put.
- the recording times can be controlled for example by an external trigger signal.
- a segmentation unit 61 which in the 2D heat WB and any 2D video images VB associated 2D pixel regions 91; 92 of the object or objects 90 according to predefined homogeneity criteria segmented, ie determined.
- the term “segmentation” refers to the generation of contiguous regions by combining adjacent pixels or voxels according to predefined homogeneity criteria. ⁇ br/> ⁇ br/>
- image processing methods 80 such as background subtraction, edge detection, threshold value methods, region-based methods, and orientation to model silhouettes can be used
- 2D methods of thermal imaging WB can optionally use methods other than 2D video images VB - for example Bayesian classifiers.
- "homogeneity criteria" are in particular the pixel and / or voxel values "color” and / or "temperature”.
- segmented 2D pixel regions 91; 92 reconstruct a 3D voxel model 93 of the object or objects 90, i. design.
- the design (reconstruction) of a 3D voxel model 93 of segmented 2D pixel regions is sometimes referred to as "space carving."
- the present invention is further distinguished by a projection unit 63, by means of which the 3D voxel model 93, which combines pixels from a plurality of synchronously present 2D heat WB and / or any 2D video images VB, as a reference for a search space SR is backprojected into the 2D heat WB and any 2D video images VB.
- the back-projected pixels of the 3D voxel model 93 correspond to the fields of view 111, 121, ...; 211, 221, ...
- the present invention finally features an identification unit 64, by means of which silhouettes 94 of the object or objects 90 are identified, ie recognized, in synchronously present 2D heat WB and possible 2D video images VB on the basis of the search space SR defined by the backprojection , to let.
- the motion analysis system advantageously allows thermally assisted segmentation of both the 2D thermal images and any 2D video images, regardless of the environmental conditions of an object 90 to be analyzed and / or tracked, and without requiring marker elements to be attached to the object 90.
- the present invention thus opens up hitherto closed application possibilities for motion analysis and / or tracking systems 1 of interest here, in particular in the areas of sports competition analysis, security technology and animal research.
- 2D thermal images WB is in a preferred embodiment of Invention proposed a 2D supplementary unit 53, which complements missing 2D thermal images WB so that for each 2D video image VB always a synchronous 2D thermal image WB is present.
- a (not shown) Key Frame Inte ⁇ olation device for the data of the 2D thermal image WB has been proven.
- the 3D voxel models 93 generated by the reconstruction unit 62 is less than the frame rate of the 2D thermal images WB recorded by means of a thermal imaging camera 11, 12, a video camera 21, 22, ... recorded 2D video images VB
- an SD supplementation unit 54 is proposed, which complements missing 3D voxel models 93 so that each 2D heat WB and possibly 2D video image VB always a synchronous 3D voxel model 93 is present.
- this has in practice, for example, a (not shown) Key Frame Inte ⁇ olation device for the 3D data of the voxel model 93 proven.
- FIG. 2 shows, by way of example, the arrangement of a first group of cameras which comprises at least one thermal imaging camera 11, 12,... And at least two video image cameras 21, 22,.
- a first group of cameras which comprises at least one thermal imaging camera 11, 12,... And at least two video image cameras 21, 22,.
- a thermal imaging camera 11 is provided as first camera and a video image camera 21 as second camera, whose objectives 112; 212 arranged at a distance x of at least two meters and / or their optical axes 113; 213 are aligned with each other at an angle ⁇ of at least 45 °,
- a video image camera 22 is provided, the lens 222 is disposed immediately adjacent to the lens 112 of the thermal imaging camera 11 so that the optical axes 113; 223 of both cameras 11, 22 are aligned substantially parallel to each other.
- the arrangement of at least one thermal imaging camera 11 in a group of cameras comprising at least two video cameras 21, 22, advantageously permits at least a first plausibility check of insufficiently segmentable 2D video images VB by the segmentation unit 61 and thus the advantageous reconstruction of less faulty 3D cameras.
- Voxel models 93 are found in the prior art.
- FIG. 1 optionally shows an iterative sequence (process) in which the segmentation unit 61 and the reconstruction unit 62 are run through repeatedly.
- the movement analysis and / or tracking system 1 comprises a further segmentation unit 61 which additionally takes into account restrictions of the search space SR based on the results of the 3D voxel model of the preceding iteration step and adapts homogeneity criteria to the current iteration step.
- the robustness of the system 1 can be further increased by a reconstruction unit 62, which additionally uses in an iterative sequence the segmented 2D pixels regions 91, 92 used for the reconstruction of the 3D voxel model 93 as a function of the current iteration step , the type of camera 11; 21; 31 and / or the quality criteria of the 2D pixel regions.
- a reconstruction unit 62 which reconstructs the 3D voxel model 93 on the basis of the depth image TB of a depth image camera 31 has proven itself.
- an iterative sequence of search space restrictions is advantageously available, consisting of segmentation unit 61, reconstruction unit 62 and projection unit 63.
- FIG. 3 shows, by way of example, the arrangement of a second group of cameras, which encompasses at least two thermal imaging cameras 1 1, 12,... And possibly video image cameras 21, 22,.
- Fig. 3 it can be seen, as for example
- a thermal imaging camera 11 as a first camera, a thermal imaging camera 11 and as a second camera, a thermal imaging camera 12 are provided, the lenses 112; 122 arranged at a distance x of at least two meters and / or their optical axes 113; 123 are aligned with each other at an angle ⁇ of at least 45 °,
- a video image camera 21 is provided, the lens 212 is disposed immediately adjacent to the lens 122 of the second thermal imaging camera 12 so that the optical axes 123; 213 of both camera 12, 21 are aligned substantially parallel to each other.
- FIG. 4 shows, by way of example, the arrangement of a third group of cameras which comprises a plurality of video cameras 21, 22,..., Arranged in any desired manner, in particular two to three, heat 11, 12,..., And in particular 5 to 6 , With fewer cameras, one would conveniently occupy the positions of the cameras in the order of the camera numbers 11, 12, 21, 22, ... or adapt them to the specific needs of the application-specific motion analysis and / or tracking.
- the arrangement of a group of cameras encompassing at least two thermal imaging cameras 11, 12,... - as suggested by way of example in FIG. 3 or FIG. 4 - advantageously permits a particularly reliable segmentation of 2D thermal images WB by means of the navigation system.
- mentation unit 61 two thermal cameras 1 1, 12 whose optical axes 1, 13, 123, ... are arranged at a suitable angle ⁇ already enable the reconstruction of a 3D voxel model 93 purely from 2D thermal images WB.
- a reconstruction unit 62 which first reconstructs a 3D voxel model 93 of the object or objects 90 from segmented 2D WB pixel regions 91 alone is preferred.
- the 3D WB voxel model 93 thus obtained purely from data of the 2D thermal image WB is particularly reliable insofar as there is a particularly robust restriction of a search space SR in the 2D video images VB of the video cameras 21, 22,. .. supplies.
- a projection unit 63 is preferred, which first projects back an SD thermal image WB voxel model 93 as a reference for a search space SR into the synchronously present 2D heat WB and any 2D video images VB.
- the motion analysis and / or tracking system 1 further comprises an allocation unit 65 which assigns points of the identified silhouettes 94 points of known silhouettes 95 of a model MO of the object or objects 90 as correspondence and / or points of the known silhouettes 95 of a Model MO of the object (s) 90 assigns points of the identified silhouettes 94 as correspondence.
- the model MO advantageously represents a virtual image of the object or objects 90. As a rule, it will be designed as a kinematic chain with associated point grid and possibly further references to sensors. It is thus possible to project the model MO in its current orientation by means of the calibration unit 51 into the 2D heat WB and possibly 2D video images VB and to display the outline, i. the Silhouette 95, to determine.
- an allocation unit 65 has proven to be useful which, in addition to the correspondences obtained from data of the silhouettes 95, optionally uses data, in particular further sensors 40, any image processing units 80 and / or a depth camera 31, 32, to generate correspondence. These are related to state variables of a model MO, ie properties related to the current orientation of a model MO.
- an allocation unit is available, which advantageously additional correspondences by assignment of further state variables of a model MO of the object or objects 90 to data, in particular of further sensors 40, any image processing units 80 and / or a depth image camera 31, 32, ... created.
- orientation sensors geoscopes
- acceleration sensors or active heat markers have proven to be suitable for further sensors 40.
- a depth image camera 31, 32 ... all cameras can be used, which allow the pictorial representation of distances. In this case, each pixel does not receive the color of the object 90 to be seen, as in a video camera 21, 22,... Or the temperature of the object as in a thermal imaging camera 11, 12,..., But the distance of the point of the object 90, is visible in the corresponding pixel.
- Depth image cameras 31, 32, ... are available in different versions such as:
- Structured Light in which a light pattern, produced by light of the visible or infrared wavelength range, is projected onto the scene to be recorded, recorded with a camera and the depth information is calculated from the distortion of the pattern with respect to the undistorted pattern;
- TOF cameras time of flight, which close the distance from transit time measurements of the light
- Light field cameras which also determines the angle in addition to position and intensity of the incident light and thus allows the calculation of depth information.
- weighting unit 66 has proven itself, which weighs correspondences created according to fixed predefined and / or variable parameters.
- weighting criteria can be implemented, which a user can possibly adjust by means of parameters.
- the movement analysis and / or tracking system 1 further comprises a movement tracking unit 71, which realigns a model MO of the object or objects 90 from associated correspondences.
- a movement tracking unit 71 has proved particularly useful, which in an iterative procedure after each iteration with already existing correspondences and / or re-aligning a model MO of the object (s) 90 with correspondence created again based on an updated model MO until the orientation of the model MO meets a predefined criterion.
- a criterion is met, for example, if the orientation of the model MO changes less than a predefined threshold or if a certain number of iterations has been reached. This advantageously provides new silhouettes 95.
- the movement analysis and / or tracking system 1 further comprises a movement analysis unit 72, which from a final alignment of a model MO of the object or objects 90 positions, in particular existing knee or other joint angle, and / or movements of or of the objects 90 analyzed.
- the motion analysis and / or tracking system 1 can be supplemented by a visualization unit 73.
- a visualization unit 73 has proved successful with which either the temperature data of the segmented 2D WB pixel regions 91 can be displayed on the 3D voxel model 93 or on a, in particular final, aligned model MO by means of texture mapping.
- Such a visualization of the temperature data can advantageously enable thermographic analyzes, in particular during the course of movement of an object 90 to be examined.
- FIG. 5 shows by way of example the arrangement of a fourth group of cameras which exclusively comprises two or more thermal imaging cameras 11, 12, 13,.
- FIG. 5 shows how the objectives 1 12, 122, 132 of the exemplary three thermal imaging cameras 11, 12, 13 are arranged at a distance x of at least two meters and / or their optical axes 113, 123, 133 to each other below a Angle ⁇ of at least 45 ° are aligned.
- the arrangement of a group of cameras formed exclusively of thermal imaging cameras 11, 12, 13,... Advantageously allows the movement analysis and / or tracking of objects 90 even in complete darkness, which is particularly useful for the investigation of nocturnal animals or in the fight against crime is of interest.
- the motion analysis and / or tracking system 1 of the present invention advantageously allows for thermally-assisted segmentation of both the 2D thermal images and any 2D video images, regardless of the environmental conditions of an object 90 to be analyzed and / or tracked, and without Object 90 marker elements to be attached requires.
- the present invention thus opens up hitherto closed application possibilities for motion analysis and / or tracking systems 1 of interest here, in particular in the areas of sports competition analysis, security technology and animal research.
- Lens (122) or (212) of the second camera (12) or (21) ⁇ angle between the optical axis (113) of the first camera (11) and the optical axis (123) or (213) of the second camera (12) or (21)
Abstract
Description
Claims
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE102016107667.8A DE102016107667A1 (en) | 2016-04-25 | 2016-04-25 | Motion analysis and / or tracking system of moving or moving objects thermally lifted from their environment |
PCT/DE2017/100331 WO2017186225A1 (en) | 2016-04-25 | 2017-04-24 | Motion analysis system and motion tracking system comprising same of moved or moving objects that are thermally distinct from their surroundings |
Publications (1)
Publication Number | Publication Date |
---|---|
EP3449463A1 true EP3449463A1 (en) | 2019-03-06 |
Family
ID=59294885
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP17736558.2A Withdrawn EP3449463A1 (en) | 2016-04-25 | 2017-04-24 | Motion analysis system and motion tracking system comprising same of moved or moving objects that are thermally distinct from their surroundings |
Country Status (5)
Country | Link |
---|---|
US (1) | US20220237808A1 (en) |
EP (1) | EP3449463A1 (en) |
CA (1) | CA3061269A1 (en) |
DE (2) | DE102016107667A1 (en) |
WO (1) | WO2017186225A1 (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10885630B2 (en) * | 2018-03-01 | 2021-01-05 | Intuitive Surgical Operations, Inc | Systems and methods for segmentation of anatomical structures for image-guided surgery |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7257237B1 (en) | 2003-03-07 | 2007-08-14 | Sandia Corporation | Real time markerless motion tracking using linked kinematic chains |
KR20090130003A (en) | 2007-03-02 | 2009-12-17 | 오가닉 모션 | System and method for tracking three dimensional objects |
DE102010028904A1 (en) | 2010-05-11 | 2011-11-17 | Movolution Gmbh | Motion analysis and / or tracking system |
EP2710557B1 (en) | 2011-05-16 | 2020-04-29 | Max-Planck-Gesellschaft zur Förderung der Wissenschaften e.V. | Fast articulated motion tracking |
US10033979B2 (en) * | 2012-03-23 | 2018-07-24 | Avigilon Fortress Corporation | Video surveillance systems, devices and methods with improved 3D human pose and shape modeling |
-
2016
- 2016-04-25 DE DE102016107667.8A patent/DE102016107667A1/en not_active Withdrawn
-
2017
- 2017-04-24 CA CA3061269A patent/CA3061269A1/en not_active Abandoned
- 2017-04-24 DE DE112017002158.8T patent/DE112017002158A5/en not_active Withdrawn
- 2017-04-24 EP EP17736558.2A patent/EP3449463A1/en not_active Withdrawn
- 2017-04-24 WO PCT/DE2017/100331 patent/WO2017186225A1/en active Application Filing
- 2017-04-24 US US16/607,744 patent/US20220237808A1/en not_active Abandoned
Also Published As
Publication number | Publication date |
---|---|
US20220237808A1 (en) | 2022-07-28 |
DE102016107667A1 (en) | 2017-10-26 |
CA3061269A1 (en) | 2017-11-02 |
DE112017002158A5 (en) | 2019-01-24 |
WO2017186225A1 (en) | 2017-11-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2422297B1 (en) | Device and method for recording a plant | |
Saberioon et al. | Automated multiple fish tracking in three-dimension using a structured light sensor | |
DE69910757T2 (en) | WAVELET-BASED FACIAL MOTION DETECTION FOR AVATAR ANIMATION | |
EP2791903B1 (en) | Method and device for estimating a pose | |
DE112017000642T5 (en) | Systems and methods for a possible collision between an object and detecting a patient | |
WO2015117906A1 (en) | 2d image analyzer | |
DE102015013031B4 (en) | Device, method and computer program for determining a position of at least two sub-segments of a patient support device | |
DE112010001224T5 (en) | Motion compensation based CT device and CT method | |
DE102012112322A1 (en) | Method for optically scanning and measuring an environment | |
CN108731587A (en) | A kind of the unmanned plane dynamic target tracking and localization method of view-based access control model | |
WO2008083869A1 (en) | Method, apparatus and computer program for the self-calibration of a surveillance camera | |
DE102011054658A1 (en) | Method for distinguishing between a real face and a two-dimensional image of the face in a biometric capture process | |
DE102019212978A1 (en) | Monitoring device for person recognition and method | |
DE112019007390T5 (en) | Method, device and computer program for tracking moving objects | |
DE102016123149A1 (en) | IMAGE-DATA-BASED RECONSTRUCTION OF THREE-DIMENSIONAL SURFACES | |
DE102014106718B4 (en) | System that presents a field of view representation in a physical position in a changeable solid angle range | |
DE102020103575B4 (en) | Feature point detection device and method for detecting feature points in image data | |
EP3449463A1 (en) | Motion analysis system and motion tracking system comprising same of moved or moving objects that are thermally distinct from their surroundings | |
DE10145608B4 (en) | Model-based object classification and target recognition | |
DE102014110570A1 (en) | An imaging apparatus and method combining functional imaging and ultrasound imaging | |
DE102019133515B3 (en) | Method and device for parallax determination of recordings of a multi-lens camera system | |
DE102021101336A1 (en) | Method for evaluating sensor data from a distance sensor, determination device, computer program and electronically readable data carrier | |
DE102015112411A1 (en) | System for the stereoscopic display of images of an endoscope | |
EP2940624A1 (en) | Method for creating a three-dimensional virtual model of an environment for applications for determining positions | |
DE102012100848A1 (en) | System for stereoscopic display of images of endoscope using dimensional map, has analyzing unit to test whether detected overexposure is reflection of light by object in spatial area based on three-dimensional model of spatial area |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: UNKNOWN |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE |
|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20181124 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
DAV | Request for validation of the european patent (deleted) | ||
DAX | Request for extension of the european patent (deleted) | ||
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
17Q | First examination report despatched |
Effective date: 20200716 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20210127 |