WO2023176562A1 - Appareil de traitement d'informations, procédé de traitement d'informations et programme de traitement d'informations - Google Patents

Appareil de traitement d'informations, procédé de traitement d'informations et programme de traitement d'informations Download PDF

Info

Publication number
WO2023176562A1
WO2023176562A1 PCT/JP2023/008411 JP2023008411W WO2023176562A1 WO 2023176562 A1 WO2023176562 A1 WO 2023176562A1 JP 2023008411 W JP2023008411 W JP 2023008411W WO 2023176562 A1 WO2023176562 A1 WO 2023176562A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
information processing
extraction
tracking target
statistical information
Prior art date
Application number
PCT/JP2023/008411
Other languages
English (en)
Japanese (ja)
Inventor
雄二 永松
Original Assignee
ソニーグループ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニーグループ株式会社 filed Critical ソニーグループ株式会社
Publication of WO2023176562A1 publication Critical patent/WO2023176562A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules

Definitions

  • the present disclosure relates to an information processing device, an information processing method, and an information processing program.
  • Patent Document 1 A technique for recognizing surrounding people and objects is disclosed in Patent Document 1, for example.
  • Patent Document 1 a rectangular filter is used when extracting a person or object from an acquired surrounding image.
  • a rectangular filter is used, a large amount of background images that become noise are included, and for example, the accuracy of identifying the same tracking target between frames (tracking accuracy) may be reduced. Therefore, it is desirable to provide an information processing device, an information processing method, and an information processing program that can suppress noise to a low level.
  • An information processing device includes an object detection section and an object image extraction section.
  • the object detection unit generates statistical information of a pixel region to be tracked included in the input image.
  • the object image extraction unit determines an extraction region to be tracked based on the statistical information, and extracts an object image from the input image as an image of the extraction region.
  • An information processing method includes the following two. (A1) Generating statistical information on the pixel area of the tracking target included in the input image (A2) Determining the extraction area of the tracking target based on the statistical information, and extracting an object image from the input image as an image of the extraction area to do
  • the information processing program causes a computer to execute the following two things.
  • (B1) Generating statistical information on the pixel area of the tracking target included in the input image
  • (B2) Determining the extraction area of the tracking target based on the statistical information, and extracting an object image from the input image as an image of the extraction area to do
  • an extraction region of a tracking target is determined based on statistical information of a pixel region of a tracking target included in an input image, and , the object image is extracted as the image of the extraction area.
  • the proportion of the background image included in the object image can be reduced compared to the case of extracting the tracking target image using a rectangular filter. It can be suppressed.
  • FIG. 1 is a diagram illustrating an example of functional blocks of an information processing system according to a first embodiment of the present disclosure.
  • FIG. 2 is a diagram illustrating a method for determining the same object using two object images.
  • FIG. 3 is a diagram illustrating a method of acquiring an object image from an input image.
  • FIG. 4 is a diagram illustrating an example of information processing in the information processing system.
  • FIG. 5 is a diagram showing an example in which an offset is provided in the extraction area.
  • FIG. 6 is a diagram showing an example of the t distribution.
  • FIG. 7 is a diagram illustrating a method for determining identical objects using a histogram.
  • FIG. 8 is a diagram illustrating a modified example of the functional blocks of the information processing system in FIG. 1.
  • FIG. 9 is a diagram illustrating an example of functional blocks of an information processing system according to the second embodiment of the present disclosure.
  • a pixel region corresponding to a tracking target is extracted for each frame, and the same tracking target is identified between frames using the degree of matching of pixel regions as an index.
  • a method is known in which a rectangular or specific shaped filter is swept over an image.
  • a filter is sometimes called a kernel.
  • FIG. 1 shows an example of functional blocks of an information processing system 1.
  • the information processing system 1 includes a sensor device section 10, an object detection section 20, a storage section 30, an object image extraction section 40, and an object tracking section 50.
  • the object detection section 20, object image extraction section 40, and object tracking section 50 are configured to include, for example, a CPU (Central Processing Unit) and a GPU (Graphics Processing Unit).
  • the object detecting section 20, the object image extracting section 40, and the object tracking section 50 are configured to read information written in the program 33 by loading various programs (for example, a program 33 described later) stored in the storage section 30 into the CPU. Perform a series of steps.
  • the object detection section 20, object image extraction section 40, and object tracking section 50 may be configured with an MPU (Micro-Processing Unit) that executes the respective functions.
  • MPU Micro-Processing Unit
  • the sensor device section 10 includes, for example, a sensor element that recognizes the external environment and acquires environmental data corresponding to the recognized external environment.
  • the sensor element outputs the acquired environmental data to the object detection section 20.
  • the sensor element is, for example, an RGB camera, an RGB-D camera, a depth sensor, an infrared sensor, or an event camera.
  • the RGB camera is, for example, a single-application visible light image sensor that outputs an RGB image obtained by receiving visible light and converting it into an electrical signal.
  • the RGB-D camera is, for example, a binocular visible light image sensor, and outputs an RGB-D image (an RGB image and a distance image obtained from parallax).
  • the depth sensor is, for example, a ToF (Time of Flight) sensor or a Lider (Laser Imaging Detection and Ranging) sensor, and outputs a distance image obtained by measuring scattered light in response to pulsed laser irradiation.
  • the infrared sensor outputs an infrared image obtained by, for example, receiving infrared light and converting it into an electrical signal.
  • the event camera is, for example, a single-purpose visible light image sensor, and outputs a difference between RGB images (difference image) between frames.
  • the sensor device unit 10 outputs various images (for example, an RGB image, an RGB-D image, a distance image, an infrared image, or a difference image) obtained from the external environment as an input image Iin (see FIG. 2(A)). ).
  • the object detection unit 20 generates statistical information of the pixel region of the tracking target included in the input image Iin obtained by the sensor device unit 10 and type information of the tracking target.
  • the object detection unit 20 stores the generated statistical information and type information, and the input image Iin in the storage unit 30.
  • the statistical information generated by the object detection unit 20 includes the average position ( ⁇ x, ⁇ y) and variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy) of the pixel region to be tracked (see FIG. 2(B)).
  • the statistical information generated by the object detection section 20 is stored in the statistical information 31 in the storage section 30.
  • the average position ( ⁇ x, ⁇ y) is, for example, two-dimensional coordinates corresponding to the center position in the X-axis direction and the center position in the Y-axis direction in the pixel region of the tracking target included in the input image Iin.
  • the variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy) are the variance-covariance values of the pixel region to be tracked included in the input image Iin.
  • ⁇ xx is the (1, 1) element in the variance-covariance matrix.
  • ⁇ yy is the (2,2) element in the variance-covariance matrix.
  • ⁇ xy is the (2,1), (1,2) element in the variance-covariance matrix.
  • the type information generated by the object detection unit 20 includes, for example, the name of the tracked object, such as a person or a car.
  • the name of the tracked object roughly indicates the shape, size, and other characteristics of the tracked object.
  • the type information generated by the object detection section 20 is stored in the type information 32 in the storage section 30.
  • the object detection unit 20 has a machine learning model such as a neural network.
  • This neural network uses, for example, a learning image (an image containing a tracking target) and statistical information (average position ( ⁇ x, ⁇ y) and variance/covariance value ( ⁇ xx , ⁇ xy, ⁇ yy)) and tracking target type information included in the learning image as teaching data.
  • this neural network calculates statistical information (average position ( ⁇ x, ⁇ y) and variance/covariance values ( ⁇ xx, ⁇ xy, ⁇ yy )) and the tracking target type information included in the input image Iin.
  • the object detection unit 20 uses a neural network to output statistical information (average position ( ⁇ x, ⁇ y) and variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy)) and tracking target type information from the input image Iin. do.
  • the storage unit 30 is, for example, a recording medium such as a semiconductor memory or a hard disk.
  • the storage unit 30 stores various programs (for example, a program 33).
  • the program 33 describes a series of procedures for realizing the respective functions of the object detection section 20, object image extraction section 40, and object tracking section 50.
  • the storage unit 30 stores various data (for example, input image Iin, object image Iob, statistical information 31, and type information 32) generated by executing the program 33.
  • the object image extraction unit 40 determines the extraction area PA of the tracking target based on the statistical information generated by the object detection unit 20 and the type information of the tracking target (see FIG. 2(C)). Specifically, the object image extraction unit 40 determines the extraction area PA based on the average position ( ⁇ x, ⁇ y), the variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy), and the tracking target type information.
  • the extraction area PA has an elliptical shape.
  • the object image extraction unit 40 derives the radius of the ellipse based on the variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy), and corrects the derived radius of the ellipse based on the tracking target type information.
  • the value derived in this way is the Mahalanobis distance.
  • the ellipse whose radius is the Mahalanobis distance has a spread corresponding to the pixel distribution of the tracking target (see FIG. 2C). Therefore, by using such an elliptical filter, it is possible to obtain an image (object image Iob) in which the background image that becomes noise is suppressed (see FIG. 2(D)).
  • the vehicle to be tracked can be surrounded by an ellipse. Therefore, for example, by determining an ellipse with a Mahalanobis distance of 3 as the extraction area PA, it is possible to cover approximately 99.7% of the pixel area to be tracked with the extraction area PA, and to suppress the introduction of noise.
  • the extraction area PA when the tracking target is a person, for example, if an ellipse with a Mahalanobis distance of 3 is determined as the extraction area PA, the extraction area PA will also cover the ends of the person's limbs, which may cause noise to be mixed in. There is a possibility that the amount will be large and cause problems in tracking the target. In this way, by changing the setting of the Mahalanobis distance depending on the type of the tracked object, it is possible to perform optimal object tracking for each tracked object.
  • the object image extraction unit 40 corrects the radius of the ellipse derived based on the variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy) to be smaller.
  • the object image extraction unit 40 greatly corrects the radius of the ellipse derived based on the variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy).
  • the object image extraction unit 40 outputs the value derived in this way as the ellipse radius (rx, ry) of the extraction area PA.
  • the derived ellipse radius (rx, ry) is stored in the statistical information 31 in the storage unit 30 in association with the identifier of the tracked object, along with the average position ( ⁇ x, ⁇ y) and variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy). Ru.
  • the object image extraction unit 40 further extracts an object image Iob from the input image Iin as an image of the extraction area PA (see FIG. 2(D)). That is, the object image extraction unit 40 extracts the object image Iob from the input image Iin based on the average position ( ⁇ x, ⁇ y) and the variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy).
  • the object tracking unit 50 tracks a tracking target included in the input image Iin. In order to perform such tracking, the object tracking unit 50 identifies the same tracking target between frames. Specifically, the object tracking unit 50 determines whether the same tracking target is included in a plurality of input images Iin obtained at different times.
  • the object tracking unit 50 determines whether or not the two input images Iin(t) and Iin(t+1) include the same tracking target based on the degree of matching. As a result, if the same tracking target is included in the two input images Iin(t) and Iin(t+1), the object tracking unit 50, for example, selects an object image Iob(t+1) that includes the tracking target. and the average position ( ⁇ x(t+1), ⁇ y(t+1)) of the pixel area to be tracked are output to the outside.
  • FIG. 3 shows an example of information processing in the information processing system 1.
  • the sensor device section 10 acquires the input image Iin (step S101).
  • the object detection unit 20 generates statistical information of the pixel region of the tracking target included in the input image Iin obtained by the sensor device unit 10 and type information of the tracking target (step S102).
  • the object image extraction unit 40 determines the extraction area PA of the pixels of the tracking target based on the statistical information generated by the object detection unit 20 and the type information of the tracking target (step S103).
  • the object image extraction unit 40 extracts the object image Iob from the input image Iin as an image of the extraction area PA (step S104).
  • the object tracking unit 50 identifies the same tracking target between frames. Specifically, the object tracking unit 50 determines whether the same tracking target is included in a plurality of input images Iin obtained at different times (step S105). As a result, if the same tracking target is included in the plurality of input images Iin (step S105; Y), the object tracking unit 50 tracks the tracking target (step S106). The object tracking unit 50 outputs, for example, an object image Iob including the tracking target and the average position ( ⁇ x, ⁇ y) of the pixel region of the tracking target to the outside. On the other hand, if the same tracking target is not included in the plurality of input images Iin (step S105; N), the object tracking unit 50 ends tracking of the tracking target (step S107). For example, the object tracking unit 50 outputs an error code to the outside.
  • the extraction area PA of the tracking target is determined based on the statistical information of the pixel area of the tracking target included in the input image Iin, and the object image Iob is extracted from the input image Iin as an image of the extraction area PA. Ru.
  • the proportion of the background image included in the object image Iob can be reduced compared to the case of extracting the area of the tracking target using a rectangular filter. can be suppressed. As a result, noise can be kept low.
  • the radius of the extraction area PA (ellipse radius (rx, ry)) is determined based on the variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy), and the average position ( ⁇ x, ⁇ y) and ellipse radius
  • the extraction area PA is determined based on (rx, ry).
  • the extraction area PA is determined based on the statistical information of the pixel area of the tracking target and the type information of the tracking target.
  • the ellipse radius (rx, ry) can be adjusted depending on the type of tracking target, and the proportion of the background image included in the object image Iob can be effectively suppressed. As a result, noise can be kept low.
  • statistical information and tracking target type information are generated from the input image Iin using a neural network.
  • the amount of calculation can be kept low compared to the case where the filter is swept.
  • the object image extraction unit 40 determines not only the radius of the extraction area PA but also the average position ( ⁇ x, An offset ⁇ of ⁇ y) may be determined.
  • facial information can be important information for tracking.
  • the entire face can be included in the object image Iob by moving the extraction area PA to a position above the average position ( ⁇ x, ⁇ y) by + ⁇ .
  • the position of the face can be predicted from the size of the ellipse radius (rx, ry)
  • the value of ⁇ can be determined based on the ellipse radius (rx, ry) of the ellipse.
  • the object image extraction unit 40 determines the offset ⁇ of the average position ( ⁇ x, ⁇ y) based on the size of the ellipse radius (rx, ry). good. By doing this, it is possible to perform tracking using a face image.
  • the object detection unit 20 corrects the generated variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy) using the t-distribution. ') may be derived.
  • the Mahalanobis distance represents the distribution of data when the pixel distribution of the tracking target is assumed to be a normal distribution.
  • the ellipse radius (rx, ry) may be derived using a distribution with a larger variance than the normal distribution. Specifically, for a tracking target with a small pixel area size, the variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy) obtained assuming that the pixel distribution of the tracking target is a normal distribution are calculated using the t distribution.
  • the ellipse radius (rx, ry) may be derived using the correction values ( ⁇ xx', ⁇ xy', ⁇ yy') obtained thereby.
  • the t distribution is, for example, a probability density function as shown in FIG.
  • This probability density function is a conditional probability density function in which the average position ( ⁇ x, ⁇ y) and variance-covariance values ( ⁇ xx, ⁇ xy, ⁇ yy), which are parameters of a normal distribution, are known.
  • This probability density function further has a property that as the degree of freedom ⁇ approaches infinity, it approaches a normal distribution.
  • the degree of freedom ⁇ is small, that is, when the number of pixels included in the pixel region of the tracking target is small, the pixel distribution of the tracking target has a gentle distribution that is a squashed normal distribution.
  • the object tracking unit 50 generates a histogram Hg for each object image Iob, and based on the generated histogram Hg, as shown in FIG. It may be determined whether the same tracking target is included in a plurality of input images Iin having different times.
  • the object tracking unit 50 calculates the average position ( ⁇ x(t), ⁇ y(t)) and ellipse radius (rx(t), ry(t)) at time t, and the time t+1.
  • the average position ( ⁇ x(t+1), ⁇ y(t+1)) and the ellipse radius (rx(t+1), ry(t+1)) are compared.
  • the object tracking unit 50 compares the histogram Hg(t) at time t and the histogram Hg(t+1) at time t+1, for example, as shown in FIG.
  • the object tracking unit 50 determines whether or not the two input images Iin(t) and Iin(t+1) include the same tracking target based on the degree of matching. As a result, if the same tracking target is included in the two input images Iin(t) and Iin(t+1), the object tracking unit 50, for example, selects an object image Iob(t+1) that includes the tracking target. and the average position ( ⁇ x(t+1), ⁇ y(t+1)) of the pixel area to be tracked are output to the outside. By using the histogram Hg for discrimination in this way, it becomes possible to perform the above discrimination with higher accuracy.
  • the information processing system 1 may further include a frame interpolation unit 60, for example, as shown in FIG.
  • the frame interpolation unit 60 includes, for example, a CPU and a GPU.
  • the frame interpolation unit 60 executes a series of procedures described in the program 33 by loading various programs (for example, the program 33) stored in the storage unit 30 into the CPU.
  • the frame interpolation unit 60 may be configured with an MPU that executes its functions.
  • the object detection unit 20 generates statistical information at a predetermined operating frequency.
  • the frame interpolation unit 60 performs frame interpolation of the statistical information generated by the object detection unit 20 using a Kalman filter.
  • the Kalman filter When the statistical information generated by the object detecting section 20 is input, the Kalman filter generates an estimated value of the statistical information at a frequency higher than the operating frequency of the object detecting section 20 based on the input statistical information.
  • the object detection unit 20 outputs the estimated value of the statistical information generated using the Kalman filter to the object image extraction unit 40.
  • the object image extraction unit 40 determines the extraction area PA of the tracking target based on the estimated value of the statistical information generated by the frame interpolation unit 60 and the tracking target type information generated by the object detection unit 20.
  • the operating frequency of the neural network can be reduced to 1/N.
  • the amount of calculation of the neural network can be reduced.
  • FIG. 9 shows an example of functional blocks of the information processing system 2.
  • the information processing system 2 includes an image DB (DataBase) generation unit instead of the object tracking unit 50 in the information processing system 2 according to the above embodiment and modifications A to D thereof. 70 and an image DB 80 are provided.
  • image DB DataBase
  • the image DB generation unit 70 stores the generated object image Iob in the image DB 80 every time the object image Iob is generated by the object image extraction unit 40.
  • the image DB 80 stores a large number of object images Iob.
  • the image DB 80 can be used, for example, as an image for learning a neural network (an image including a tracking target).
  • the object detection section 20, the storage section 30, the object image extraction section 40, and the object tracking section 50 may all be installed in a common information processing device.
  • the sensor device section 10, the object detection section 20, the storage section 30, the object image extraction section 40, and the object tracking section 50 may all be installed in a common information processing device.
  • the object detection section 20, the storage section 30, the object image extraction section 40, the image DB generation section 70, and the image DB 80 may all be installed in a common information processing device.
  • the sensor device section 10, the object detection section 20, the storage section 30, the object image extraction section 40, the image DB generation section 70, and the image DB 80 are all installed in a common information processing device. You can leave it there.
  • the present disclosure can take the following configuration.
  • an object detection unit that generates statistical information of a pixel region to be tracked included in the input image;
  • An information processing apparatus comprising: an object image extraction unit that determines an extraction area of the tracking target based on the statistical information and extracts an object image from the input image as an image of the extraction area.
  • the statistical information includes an average position and a variance-covariance value of the pixel region to be tracked,
  • the extraction area has an elliptical shape
  • the information processing device determines a radius of the extraction region based on the variance-covariance value, and determines the extraction region based on the average position and the radius.
  • the object detection unit derives a correction value by correcting the generated variance-covariance value using a t-distribution, The information processing device according to (2), wherein the object image extraction unit determines the radius of the extraction area based on the correction value.
  • the object detection unit generates the statistical information and the tracking target type information from the input image, The information processing device according to (1), wherein the object image extraction unit determines the extraction area based on the statistical information and the type information.
  • the statistical information includes an average position and a variance-covariance value of the pixel distribution of the tracking target,
  • the extraction area has an elliptical shape,
  • the object image extraction unit determines a radius of the extraction area and an offset of the average position based on the variance-covariance value and the type information, and determines the radius of the extraction area and the offset of the average position based on the average position, the radius, and the offset.
  • the information processing device according to (4), which determines an extraction area.
  • the object detection unit derives a correction value by correcting the generated variance-covariance value using a t-distribution, The information processing device according to (5), wherein the object image extraction unit determines the radius of the extraction area and the offset of the average position based on the correction value and the type information. (7) The information processing device according to any one of (1) to (3), wherein the object detection unit outputs the statistical information from the input image using a neural network. (8) The information processing device according to any one of (4) to (6), wherein the object detection unit outputs the statistical information and the type information from the input image using a neural network.
  • the standard object detection unit generates the statistical information at a predetermined operating frequency
  • the information processing device according to any one of (1) to (8), further comprising a frame interpolation unit that performs frame interpolation of the statistical information generated by the systematic object detection unit using a Kalman filter.
  • Information processing device (10) According to any one of (1) to (9), further comprising an object tracking unit that determines whether the same tracking target is included in the plurality of input images obtained at different times. information processing equipment.
  • Generating statistical information of a pixel region to be tracked included in an input image An information processing method comprising: determining an extraction region of the tracking target based on the statistical information, and extracting an object image from the input image as an image of the extraction region.
  • An information processing program that causes a computer to execute the following steps: determining an extraction region of the tracking target based on the statistical information, and extracting an object image from the input image as an image of the extraction region.
  • an extraction region of a tracking target is determined based on statistical information of a pixel region of a tracking target included in an input image, and , the object image is extracted as the image of the extraction area.
  • the proportion of the background image included in the object image can be reduced compared to the case of extracting the image of the tracking target using a rectangular filter. It can be suppressed. As a result, noise can be kept low.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

Un appareil de traitement d'informations selon un mode de réalisation de la présente divulgation comprend une unité de détection d'objet et une unité d'extraction d'image d'objet. L'unité de détection d'objet génère des informations statistiques relatives à une région de pixel à suivre, incluse dans une image d'entrée. L'unité d'extraction d'image d'objet détermine une région d'extraction à suivre, sur la base des informations statistiques, et extrait une image d'objet, en tant qu'image de la région d'extraction, à partir de l'image d'entrée.
PCT/JP2023/008411 2022-03-17 2023-03-06 Appareil de traitement d'informations, procédé de traitement d'informations et programme de traitement d'informations WO2023176562A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022042856 2022-03-17
JP2022-042856 2022-03-17

Publications (1)

Publication Number Publication Date
WO2023176562A1 true WO2023176562A1 (fr) 2023-09-21

Family

ID=88023031

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2023/008411 WO2023176562A1 (fr) 2022-03-17 2023-03-06 Appareil de traitement d'informations, procédé de traitement d'informations et programme de traitement d'informations

Country Status (1)

Country Link
WO (1) WO2023176562A1 (fr)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005122617A (ja) * 2003-10-20 2005-05-12 Advanced Telecommunication Research Institute International リアルタイムの物体検出及び認識システム並びにコンピュータで実行可能なプログラム
US20060187305A1 (en) * 2002-07-01 2006-08-24 Trivedi Mohan M Digital processing of video images
JP2018084802A (ja) * 2016-11-11 2018-05-31 株式会社東芝 撮像装置、撮像システム及び距離情報取得方法
JP2018101165A (ja) * 2015-04-27 2018-06-28 国立大学法人 奈良先端科学技術大学院大学 色画像処理方法、色画像処理プログラム、物体認識方法及び装置

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060187305A1 (en) * 2002-07-01 2006-08-24 Trivedi Mohan M Digital processing of video images
JP2005122617A (ja) * 2003-10-20 2005-05-12 Advanced Telecommunication Research Institute International リアルタイムの物体検出及び認識システム並びにコンピュータで実行可能なプログラム
JP2018101165A (ja) * 2015-04-27 2018-06-28 国立大学法人 奈良先端科学技術大学院大学 色画像処理方法、色画像処理プログラム、物体認識方法及び装置
JP2018084802A (ja) * 2016-11-11 2018-05-31 株式会社東芝 撮像装置、撮像システム及び距離情報取得方法

Similar Documents

Publication Publication Date Title
US10216979B2 (en) Image processing apparatus, image processing method, and storage medium to detect parts of an object
US7912253B2 (en) Object recognition method and apparatus therefor
JP4295799B2 (ja) データ駆動確率伝播を伴う人間の姿勢推定
JP4459137B2 (ja) 画像処理装置及びその方法
US9378422B2 (en) Image processing apparatus, image processing method, and storage medium
US20140307917A1 (en) Robust feature fusion for multi-view object tracking
EP2339507B1 (fr) Procédé de détection et localisation de tête
Tavakkoli et al. Non-parametric statistical background modeling for efficient foreground region detection
US9082000B2 (en) Image processing device and image processing method
JP2018022360A (ja) 画像解析装置、画像解析方法およびプログラム
JP7272024B2 (ja) 物体追跡装置、監視システムおよび物体追跡方法
US20140204238A1 (en) Feature extraction apparatus, feature extraction program, and image processing apparatus
US9443137B2 (en) Apparatus and method for detecting body parts
CN110097050B (zh) 行人检测方法、装置、计算机设备及存储介质
US9305359B2 (en) Image processing method, image processing apparatus, and computer program product
KR20170056860A (ko) 이미지 생성 방법 및 장치
EP3772037A1 (fr) Appareil de traitement d'images, procédé de suivi d'un objet cible et programme
JP7086878B2 (ja) 学習装置、学習方法、プログラムおよび認識装置
JP2019536164A (ja) 画像処理装置、画像処理方法および画像処理プログラム
KR20210133880A (ko) 이미지 심도 결정 방법 및 생체 식별 방법, 회로, 디바이스, 및 매체
CN104063709A (zh) 视线检测装置及方法、图像捕获装置及其控制方法
US11462052B2 (en) Image processing device, image processing method, and recording medium
JP2014021602A (ja) 画像処理装置及び画像処理方法
WO2023176562A1 (fr) Appareil de traitement d'informations, procédé de traitement d'informations et programme de traitement d'informations
US10572753B2 (en) Outside recognition device for vehicle

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23770518

Country of ref document: EP

Kind code of ref document: A1