WO2023012920A1 - Estimation device, estimation method, and program - Google Patents

Estimation device, estimation method, and program Download PDF

Info

Publication number
WO2023012920A1
WO2023012920A1 PCT/JP2021/028897 JP2021028897W WO2023012920A1 WO 2023012920 A1 WO2023012920 A1 WO 2023012920A1 JP 2021028897 W JP2021028897 W JP 2021028897W WO 2023012920 A1 WO2023012920 A1 WO 2023012920A1
Authority
WO
WIPO (PCT)
Prior art keywords
space
reverberation
value based
simulation
impulse response
Prior art date
Application number
PCT/JP2021/028897
Other languages
French (fr)
Japanese (ja)
Inventor
達也 加古
賢一 野口
Original Assignee
日本電信電話株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電信電話株式会社 filed Critical 日本電信電話株式会社
Priority to JP2023539439A priority Critical patent/JPWO2023012920A1/ja
Priority to PCT/JP2021/028897 priority patent/WO2023012920A1/en
Publication of WO2023012920A1 publication Critical patent/WO2023012920A1/en

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01HMEASUREMENT OF MECHANICAL VIBRATIONS OR ULTRASONIC, SONIC OR INFRASONIC WAVES
    • G01H17/00Measuring mechanical vibrations or ultrasonic, sonic or infrasonic waves, not provided for in the preceding groups

Definitions

  • the present invention relates to an estimation technique for estimating the reflectance of walls and the like.
  • Non-Patent Document 1 describes a device capable of acquiring point cloud data inside and outside a building.
  • the point cloud data refers to data representing the presence or absence of a point for each coordinate, with a predetermined real space as a three-dimensional coordinate space. It is possible to add additional information to the point cloud data in addition to the presence or absence of a point for each coordinate, but the additional information is generally information relating to the color of the point.
  • the reflectance represents the attenuation value of the wave of the waveform when the sound is reflected. For example, if a sound hits a wall with a reflectance of 0.5, a wave with half the amplitude of the wave before reflection is reflected from the wall.
  • impulse response is sometimes used, but it needs to be done by a skilled person taking a long time.
  • the purpose of the present invention is to provide a technology that enables estimation of reflectance in a short period of time using simulation space and real space.
  • an estimating device estimates the reflectance of a wall surface within a predetermined space.
  • the estimating device includes a simulation acquisition unit that acquires a value based on the reverberation of an acoustic signal emitted from a predetermined position in a simulation space that simulates a predetermined space, and a simulation acquisition unit that corresponds to the predetermined position in the predetermined space that is the real space.
  • a real space acquisition unit that acquires a value based on the reverberation of an acoustic signal emitted from a position where the object is located, and an evaluation value that uses the difference between the value based on the reverberation in the simulation space and the value based on the reverberation in the real space. and a parameter estimator for estimating the reflectance of the wall surface.
  • FIG. 3 is a functional block diagram of an acquisition unit; The figure which shows the example of the processing flow of an acquisition part. The figure showing the example of the point cloud data after noise removal. The figure showing the example of the point cloud data after noise removal.
  • the functional block diagram of the room shape creation part The figure which shows the example of the processing flow of a room shape preparation part.
  • FIG. 4 is a diagram for explaining processing of a loss complementing unit; FIG. 4 is a diagram for explaining processing of a loss complementing unit; FIG. 4 is a diagram for explaining parameter estimation by a Gaussian process; The figure which shows the table
  • the reflectivity of the wall surface is estimated so that the difference from the value based on the reverberation of the received acoustic signal is minimized.
  • RT60 As a value based on the reverberation of the acoustic signal, it is possible to reduce the amount of calculation by reducing the influence of the position of the sound source and the position of the microphone compared to the impulse response.
  • FIG. 1 is a functional block diagram of an estimation system according to the first embodiment, and FIG. 2 shows its processing flow.
  • the estimation system includes an acquisition unit 110, an estimation device 100, and an impulse response output unit 190.
  • the estimation system acquires point cloud data of the space in which sound is to be collected via the acquisition unit 110, estimates the reflectance of a predetermined space using the point cloud data, and uses the estimated value i of the reflectance to obtain a predetermined It estimates and outputs the impulse response of an arbitrary sound source position and an arbitrary microphone position in the space of .
  • the estimating device 100 is, for example, a special computer configured by reading a special program into a known or dedicated computer having a central processing unit (CPU: Central Processing Unit), a main memory (RAM: Random Access Memory), etc. It is a device.
  • the estimating apparatus 100 executes each process under the control of a central processing unit, for example. Data input to the estimating device 100 and data obtained in each process are stored, for example, in a main memory device, and the data stored in the main memory device are read out to the central processing unit as necessary and used for other purposes. used to process At least a part of each processing unit of the estimation device 100 may be configured by hardware such as an integrated circuit.
  • Each storage unit included in the estimation device 100 can be configured by, for example, a main storage device such as RAM (Random Access Memory), or middleware such as a relational database or a key-value store.
  • a main storage device such as RAM (Random Access Memory), or middleware such as a relational database or a key-value store.
  • middleware such as a relational database or a key-value store.
  • each storage unit does not necessarily have to be provided inside the estimating apparatus 100, and is configured by an auxiliary storage device configured by a semiconductor memory device such as a hard disk, an optical disk, or a flash memory, and the estimating apparatus 100 may be provided outside.
  • the acquiring unit 110 acquires the point cloud data R of the space in which sound is to be collected (S110) and outputs it.
  • FIG. 3 is a functional block diagram of the acquisition unit 110, and FIG. 4 shows an example of its processing flow.
  • the acquisition unit 110 includes a spatial sensing unit 111 , a noise removing unit 113 and a spatial model combining unit 115 .
  • the space sensing unit 111 acquires point cloud data of the space in which sound is to be collected (S111) and outputs it.
  • the space sensing unit 111 consists of a LiDAR installed in a space where sound is collected. and find the direction of the object from the launch direction. This distance and direction are represented by point cloud data.
  • Various conventional techniques can be used as spatial sensing techniques.
  • Reference 1 is known as an existing space sensing technology.
  • the noise removal unit 113 receives the point cloud data, removes noise included in the point cloud data (S113), and outputs the removed point cloud data.
  • Various conventional techniques can be used as the noise removal technique.
  • Reference 2 is known as an existing noise removal technique.
  • FIG. 5 shows point cloud data acquired with the spatial sensing unit 111 (for example, LiDAR) at an elevation angle of 0 degrees
  • FIG. 6 shows point cloud data obtained with an elevation angle of ⁇ 90 degrees.
  • the space model combining unit 115 receives a plurality of point cloud data, combines the plurality of point cloud data, restores the shape of the space (S115), and outputs point cloud data R representing the shape of the space. For example, while rotating the LiDAR, multiple point cloud data with different elevation angles are obtained, and the multiple point cloud data are combined using the Iterative Closest Point (ICP) algorithm to reconstruct a 3D scene.
  • ICP Iterative Closest Point
  • Various conventional techniques can be used as the spatial model combination technique.
  • Reference 3 is known as an existing spatial model combining technique.
  • the estimating apparatus 100 receives point cloud data R indicating the shape of a space, estimates the reflectance of a given space using the point cloud data R, and outputs an estimated value i of the reflectance.
  • the estimation device 100 includes a room shape generator 120, a parameter input unit 130, an impulse response estimator 140, a reverberation time calculator 160, a real environment impulse response acquirer 150, an evaluation function calculator 170, and a parameter estimator 180. (See Figure 1).
  • the estimating device 100 may or may not include the obtaining unit 110 in the computer that constitutes the estimating device 100 .
  • the estimation device 100 includes an input interface for inputting point cloud data (not shown).
  • the processing S113 and the processing S115, or the processing S115 may be performed within the computer that constitutes the estimation device 100 .
  • an input unit (not shown) is provided before the part corresponding to the processing performed in the computer. , or noise-removed point cloud data output from the noise removal unit 113).
  • the room shape generator 120 receives the point cloud data R indicating the shape of the space, complements the shape of the space that does not appear in the point cloud data R (S120), and outputs the shape U of the complemented space.
  • FIG. 7 is a functional block diagram of the room shape creation unit 120, and FIG. 8 shows an example of its processing flow.
  • the room shape creation unit 120 includes a wall surface extraction unit 121 and a loss complement unit 123.
  • the wall surface extracting unit 121 receives the point cloud data R indicating the shape of the space, uses the shape of the space to extract the wall surfaces forming the space (S121), and outputs them.
  • the wall surface extraction unit 121 extracts the wall surface by acquiring the parameters of the wall surface forming the space using an algorithm that is an improvement of RANSAC (Random Sample Consensus) for wall surface extraction. Examples are shown below.
  • RANSAC Random Sample Consensus
  • n is the normal vector of the plane containing the three points a, b, c, and p is each point of the point cloud belonging to the ROI.
  • the normal vector can be obtained by SVD (Singular Value Decomposition) or PCA (Principal Component Analysis) of the covariance matrix.
  • the ratio of the point cloud data belonging to the largest wall surface to the entire point cloud data R is less than or equal to a predetermined value, it is assumed that there is no wall surface there. That is, if the ratio of the number nq of points below the threshold in (5) above in a certain direction is below a predetermined value, it is assumed that there is no wall there. Furthermore, in (10) above, since the point cloud data belonging to the largest wall surface is removed from the data, the number of wall surfaces that can be extracted gradually decreases and converges automatically. With such a configuration, not only the walls and ceiling but also the plane shape of the point cloud that is likely to reflect are extracted to some extent, and as a result, furniture and equipment are simulated as a space. For example, the process may be controlled to be repeated until the plane formed by the point cloud data belonging to the largest wall surface becomes equal to or less than a predetermined threshold value (for example, 20 cm square).
  • the predetermined threshold may be estimated according to the amount of processing.
  • the loss complementing unit 123 receives the wall surface forming the space, and if there is an object other than the wall surface in the space and the shape of the part of the wall surface cannot be estimated, the shape of the part of the wall surface can be estimated from the estimated shape of the other part of the wall surface. The shape is estimated, the shape of the space is complemented using the estimated part of the wall surface (S123), and the shape U of the space after complementation is output.
  • LiDAR uses laser reflection to measure distance
  • data loss occurs in areas where the laser cannot reach due to obstructions. Therefore, the missing data is supplemented from the information of the estimated wall surface, and the vertex (corner of the room) is obtained.
  • FIG. 9 shows point cloud data on a horizontal plane with an elevation angle of 0 degrees, and point O is the installation position of the LiDAR.
  • the parameters of the wall surface (a, b, c, d) are extracted by plane approximation, and the vertex coordinates of the wall are obtained from the straight line. That is, as shown in FIG. 10, the point of intersection P of the straight lines is determined as the vertex coordinate of the wall, and the broken line portion in FIG. 10 is assumed to be the wall surface.
  • the parameter input unit 130 receives the sound source position S and the microphone position M as inputs (S 130 ) and outputs them to the impulse response estimation unit 140 .
  • a speaker and a microphone are arranged at positions in the real space corresponding to the sound source position S and the microphone position M in the simulation space simulating the inside of a predetermined space. Therefore, the sound source position S and the microphone position M are positions where the speaker and the microphone can be arranged in the real space. One or more positions can be envisaged as the sound source position S and the microphone position M respectively.
  • a parameter input unit 130 is an input device such as a keyboard or a mouse, and the user inputs the sound source position via the parameter input unit 130 .
  • the parameter input unit 130 receives the shape of the space after complementing, which is the output of the loss complementing unit 123 (indicated by the dashed line in FIG. 1), and uses the shape of the space after complementing via a display device such as a display.
  • a display device such as a display.
  • a configuration may also be adopted in which the sound source is presented to the user, and the user uses a mouse or the like to specify where in the space after complementation the sound source is to be placed.
  • it may be configured to automatically input using object recognition such as Reference 4 from point cloud data R acquired from LiDAR, or to manually select from the position of the recognized object.
  • the parameter input unit 130 is a storage medium reading device, and the estimation apparatus 100 reads the storage medium storing the sound source positions and the microphone positions via the parameter input unit 130, thereby accepting the sound source positions and the microphone positions as inputs. good too.
  • the sound source position and the microphone position may be manually input to the estimation device 100 in advance.
  • the impulse response estimation unit 140 receives the sound source position S, the space shape U, the listening position M, and the estimated value R or the initial value R ini of the reflectance, and uses these values to calculate the sound source position in the space shape U.
  • the impulse response of S and the listening position M is estimated (S140), and the estimated value i e of the impulse response is output.
  • the reflectance estimate value R is the output of the parameter estimator 180, which will be described later.
  • the impulse response estimating unit 140 uses the sound source position S, the space shape U, the estimated value R or initial value R ini of the reflectance, and the listening position M to calculate the space of the listening position M in the space at the sound source position S. Calculate the transfer function.
  • Various conventional techniques can be used as the spatial transfer function calculation technique.
  • the FDTD method finite-difference time-domain method
  • the impulse response estimator 140 estimates the impulse response using the spatial transfer function.
  • the real environment impulse response acquisition unit 150 picks up and acquires an acoustic signal emitted from a position corresponding to the sound source position S at a position corresponding to the listening position M in a predetermined space that is the real space, and obtains the sound source position S and the position corresponding to the listening position M are measured (S150), and the measured value im of the impulse response is output.
  • a TSP Time Stretched Pulse
  • a speaker placed at a position corresponding to the sound source position S in a given space, which is the real space, and placed at a position corresponding to the listening position in the given space, which is the real space.
  • the microphone and speaker may be separate devices, or may be an integrated device such as a smart speaker.
  • the reverberation time calculator 160 receives the impulse response estimated value i e and the impulse response measured value im as inputs, and calculates the reverberation time r e of the impulse response estimated value i e and the reverberation time of the impulse response measured value im Calculate time r i (S160) and output.
  • RT60 is used as the reverberation time. RT60 is the time it takes for the reverberation in the room to drop by 60dB. By using the RT60, unlike the impulse response, it does not change sensitively depending on the location or sound source position.
  • a combination of the impulse response estimator 140 and the reverberation time calculator 160 is used to obtain a value (RT60) based on reverberation of an acoustic signal emitted from a predetermined position (sound source position) in a simulation space simulating a predetermined space. , is also called a simulation acquisition unit.
  • the combination of the real environment impulse response acquisition unit 150 and the reverberation time calculation unit 160 is used to obtain a value (RT60) based on the reverberation of an acoustic signal emitted from a position corresponding to a given position in a given space that is the real space. Since it acquires, it is also called a real space acquisition unit.
  • the evaluation function calculator 170 receives the reverberation time r e and the reverberation time r m as inputs, calculates the evaluation value E using the difference between the reverberation time r e and the reverberation time r m (S170), and outputs it.
  • an evaluation value for example, the square error between RT60 of the estimated value i e of the impulse response and RT60 of the measured value im of the impulse response can be considered, and is represented by the following equation.
  • the parameter estimator 180 receives the evaluation value E, estimates the reflectance of the wall surface using the evaluation value (S180), and outputs the estimated value R.
  • the parameter estimation unit 180 performs parameter estimation using a Gaussian process.
  • the evaluation value E and the parameter to be manipulated are used as items necessary for parameter estimation by Gaussian process.
  • the manipulated parameter is the reflectance of each wall.
  • FIG. 11 is a diagram for explaining parameter estimation by a Gaussian process.
  • the vertical axis of the figure represents the evaluation value, and the horizontal axis represents the reflectance index.
  • the parameter estimating unit 180 uniquely obtains an evaluation function from each parameter, predicts a regression curve using a Gaussian distribution, sequentially updates the predicted parameter (reflectance) that minimizes the evaluation value E, and uses a high-speed evaluation function. It is possible to search for the lowest value.
  • the parameter estimation section 180 When the number of iterations is equal to or less than the predetermined number (NO in S185), the parameter estimation section 180 outputs the updated parameters to the impulse response estimation section 140, performs processes S140 to S180, and repeats parameter estimation.
  • the parameter estimator 180 When the number of iterations exceeds a predetermined number (YES in S185), the parameter estimator 180 outputs the parameter with the minimum evaluation function as the final reflectance estimated value Rf .
  • the impulse response output unit 190 receives the sound source position S′, the spatial shape U, the listening position M′, and the final reflectance estimate value R f as inputs, and uses these values to calculate the response at the sound source position S′.
  • the impulse response of the listening position M' in space is estimated (S190), and the estimated value i of the impulse response is output.
  • the impulse response is estimated by the same method as the impulse response estimator 140 .
  • Impulse responses at various positions can be estimated while changing the sound source position S' and the listening position M'.
  • the estimating device 100 may or may not include the impulse response output unit 190 in the computer configuring the estimating device 100 .
  • the reflectance can be estimated in a short time.
  • RT60 it is possible to reduce the amount of calculation by reducing the influence of the position of the sound source and the position of the microphone compared to the impulse response.
  • the spatial shape U which is the output value of the acquisition unit 110, is used as the input of the impulse response estimation unit 140. may be used as an input for In this case, acquisition unit 110 and room shape creation unit 120 may not be included.
  • RT60 is used as the value based on the reverberation of the acoustic signal in this embodiment
  • the impulse response may be used as the value based on the reverberation of the acoustic signal.
  • estimation device 100 does not need to include reverberation time calculation section 160 .
  • RT60 by using RT60, the influence of the position of the sound source and the position of the microphone can be reduced compared to the impulse response, and the amount of calculation can be reduced.
  • the parameter estimator 180 outputs the parameter with the maximum evaluation function as the final reflectance estimated value R f .
  • the difference between the value based on the reverberation in the simulation space and the value based on the reverberation in the real space is used.
  • the final estimated reflectance value R f is sufficient.
  • ⁇ Simulation result> Estimation by the FDTD method takes a very long time, so the mirror image method is used in this simulation. Pyroomacoustics is used for the mirror image method. The true value of the reflectance is set to 0.15, and the RT60 norm is used as the evaluation function instead of the L2 norm of the impulse response.
  • FIG. 12 is a table of updates by Gaussian processes, evaluation functions, and estimated parameter values.
  • iter represents the number of iterations
  • target represents the evaluation function
  • var_abs represents the reflectance.
  • a true value of 0.15 maximizes the evaluation function at 13 iterations.
  • the evaluation function squared error
  • the evaluation function is multiplied by a negative value to search for a variable that maximizes the evaluation function.
  • the present invention is not limited to the above embodiments and modifications.
  • the various types of processing described above may not only be executed in chronological order according to the description, but may also be executed in parallel or individually according to the processing capacity of the device that executes the processing or as necessary.
  • appropriate modifications are possible without departing from the gist of the present invention.
  • a program that describes this process can be recorded on a computer-readable recording medium.
  • Any computer-readable recording medium may be used, for example, a magnetic recording device, an optical disk, a magneto-optical recording medium, a semiconductor memory, or the like.
  • this program is carried out, for example, by selling, assigning, lending, etc. portable recording media such as DVDs and CD-ROMs on which the program is recorded.
  • the program may be distributed by storing the program in the storage device of the server computer and transferring the program from the server computer to other computers via the network.
  • a computer that executes such a program for example, first stores the program recorded on a portable recording medium or the program transferred from the server computer once in its own storage device. Then, when executing the process, this computer reads the program stored in its own recording medium and executes the process according to the read program. Also, as another execution form of this program, the computer may read the program directly from a portable recording medium and execute processing according to the program, and the program is transferred from the server computer to this computer. Each time, the processing according to the received program may be executed sequentially. In addition, the above-mentioned processing is executed by a so-called ASP (Application Service Provider) type service, which does not transfer the program from the server computer to this computer, and realizes the processing function only by its execution instruction and result acquisition. may be It should be noted that the program in this embodiment includes information that is used for processing by a computer and that conforms to the program (data that is not a direct instruction to the computer but has the property of prescribing the processing of the computer, etc.).
  • ASP
  • the device is configured by executing a predetermined program on a computer, but at least part of these processing contents may be implemented by hardware.

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)

Abstract

Provided is a technology that uses a simulation space and a real space and makes it possible to estimate reflectance in a short amount of time. According to the present invention, an estimation device estimates the reflectance of a wall surface in a prescribed space. The estimation device includes a simulation acquisition unit that acquires a value based on the reverberation of an acoustic signal emitted from a prescribed location in a simulation space that simulates the prescribed space, a real space acquisition unit that acquires a value based on the reverberation of an acoustic signal emitted from a location that corresponds to the prescribed location in a real space that is the prescribed space, and a parameter estimation unit that estimates the reflectance of the wall surface using an evaluation value obtained using the difference between the value based on the reverberation in the simulation space and the value based on the reverberation in the real space.

Description

推定装置、推定方法、およびプログラムEstimation device, estimation method, and program
 本発明は、壁などの反射率を推定する推定技術に関する。 The present invention relates to an estimation technique for estimating the reflectance of walls and the like.
 LiDAR(Light Detection and RangingまたはLaser Imaging Detection and Ranging)に代表される点群センサを用いた機器は数多く存在する。例えば、非特許文献1では建物の内外の点群データを取得可能である機器が記載されている。ここで、点群データとは現実の所定の空間内を3次元座標空間とし、座標ごとに点の有無を表現するデータのことをいう。点群データには座標ごとの点の有無に加え付加情報を加えることも可能であるが、付加情報は一般に当該点の色に係る情報である。 There are many devices that use point cloud sensors, represented by LiDAR (Light Detection and Ranging or Laser Imaging Detection and Ranging). For example, Non-Patent Document 1 describes a device capable of acquiring point cloud data inside and outside a building. Here, the point cloud data refers to data representing the presence or absence of a point for each coordinate, with a predetermined real space as a three-dimensional coordinate space. It is possible to add additional information to the point cloud data in addition to the presence or absence of a point for each coordinate, but the additional information is generally information relating to the color of the point.
 点群データを利用して当該空間における音響信号をシミュレーションしようとすると、点の有無だけではなく点の反射率を考慮する必要がある。なお、反射率とは、音が反射した際の波形の波の減衰値を表す。例えば、音が反射率0.5の壁にぶつかった場合、反射前の波の半分の振幅の波が壁から反射する。このような反射率を現実空間で測定しようとする場合、インパルス応答が用いられることがあるが、熟練者が時間をかけて行う必要がある。 When trying to simulate acoustic signals in the space using point cloud data, it is necessary to consider not only the presence or absence of points but also the reflectance of points. Note that the reflectance represents the attenuation value of the wave of the waveform when the sound is reflected. For example, if a sound hits a wall with a reflectance of 0.5, a wave with half the amplitude of the wave before reflection is reflected from the wall. When trying to measure such reflectance in the real space, impulse response is sometimes used, but it needs to be done by a skilled person taking a long time.
 本発明は、シミュレーション空間と現実空間を利用し、短時間での反射率の推定を可能とする技術を提供することを目的とする。 The purpose of the present invention is to provide a technology that enables estimation of reflectance in a short period of time using simulation space and real space.
 上記の課題を解決するために、本発明の一態様によれば、推定装置は、所定の空間内における壁面の反射率を推定する。推定装置は、所定の空間内を模擬したシミュレーション空間における所定の位置から発せられた音響信号の残響に基づく値を取得するシミュレーション取得部と、現実空間である所定の空間において、所定の位置に対応する位置から発せられた音響信号の残響に基づく値を取得する現実空間取得部と、シミュレーション空間における残響に基づく値と、現実空間における残響に基づく値と、の差を用いた評価値を利用して壁面の反射率を推定するパラメータ推定部とを含む。 In order to solve the above problems, according to one aspect of the present invention, an estimating device estimates the reflectance of a wall surface within a predetermined space. The estimating device includes a simulation acquisition unit that acquires a value based on the reverberation of an acoustic signal emitted from a predetermined position in a simulation space that simulates a predetermined space, and a simulation acquisition unit that corresponds to the predetermined position in the predetermined space that is the real space. A real space acquisition unit that acquires a value based on the reverberation of an acoustic signal emitted from a position where the object is located, and an evaluation value that uses the difference between the value based on the reverberation in the simulation space and the value based on the reverberation in the real space. and a parameter estimator for estimating the reflectance of the wall surface.
 本発明によれば、短時間で反射率を推定することができるという効果を奏する。 According to the present invention, it is possible to estimate the reflectance in a short time.
第一実施形態に係る推定システムの機能ブロック図。The functional block diagram of the estimation system which concerns on 1st embodiment. 第一実施形態に係る推定システムの処理フローの例を示す図。The figure which shows the example of the processing flow of the estimation system which concerns on 1st embodiment. 取得部の機能ブロック図。FIG. 3 is a functional block diagram of an acquisition unit; 取得部の処理フローの例を示す図。The figure which shows the example of the processing flow of an acquisition part. ノイズ除去後の点群データの例を表す図。The figure showing the example of the point cloud data after noise removal. ノイズ除去後の点群データの例を表す図。The figure showing the example of the point cloud data after noise removal. 部屋形状作成部の機能ブロック図。The functional block diagram of the room shape creation part. 部屋形状作成部の処理フローの例を示す図。The figure which shows the example of the processing flow of a room shape preparation part. 欠損補完部の処理を説明するための図。FIG. 4 is a diagram for explaining processing of a loss complementing unit; 欠損補完部の処理を説明するための図。FIG. 4 is a diagram for explaining processing of a loss complementing unit; ガウス過程によるパラメータ推定を説明するための図。FIG. 4 is a diagram for explaining parameter estimation by a Gaussian process; ガウス過程による更新と評価関数および推定したパラメータ値の表を示す図。The figure which shows the table|surface of update by a Gaussian process, an evaluation function, and the estimated parameter value. 本手法を適用するコンピュータの構成例を示す図。The figure which shows the structural example of the computer which applies this method.
 以下、本発明の実施形態について、説明する。なお、以下の説明に用いる図面では、同じ機能を持つ構成部や同じ処理を行うステップには同一の符号を記し、重複説明を省略する。以下の説明において、ベクトルや行列の各要素単位で行われる処理は、特に断りが無い限り、そのベクトルやその行列の全ての要素に対して適用されるものとする。 Embodiments of the present invention will be described below. It should be noted that in the drawings used for the following description, the same reference numerals are given to components having the same functions and steps that perform the same processing, and redundant description will be omitted. In the following description, processing performed for each element of a vector or matrix applies to all elements of the vector or matrix unless otherwise specified.
<第一実施形態のポイント>
 (1)所定の空間内を模擬したシミュレーション空間における所定の位置から発せられた音響信号の残響に基づく値と、現実空間である所定の空間において、シミュレーション空間における所定の位置に対応する位置から発せられた音響信号の残響に基づく値との差分が最小となるように、壁面の反射率を推定する。
<Points of the first embodiment>
(1) A value based on the reverberation of an acoustic signal emitted from a predetermined position in a simulation space that simulates a predetermined space, and a value based on the reverberation of an acoustic signal emitted from a position corresponding to the predetermined position in the simulation space in a predetermined space that is the real space. The reflectivity of the wall surface is estimated so that the difference from the value based on the reverberation of the received acoustic signal is minimized.
 (2)音響信号の残響に基づく値として、RT60を用いることで、インパルス応答に比べて音源の位置やマイクロホンの位置による影響を減らし、計算量の削減を可能とする。 (2) By using RT60 as a value based on the reverberation of the acoustic signal, it is possible to reduce the amount of calculation by reducing the influence of the position of the sound source and the position of the microphone compared to the impulse response.
 <第一実施形態に係る推定システム>
 図1は第一実施形態に係る推定システムの機能ブロック図を、図2はその処理フローを示す。
<Estimation system according to the first embodiment>
FIG. 1 is a functional block diagram of an estimation system according to the first embodiment, and FIG. 2 shows its processing flow.
 推定システムは、取得部110と、推定装置100と、インパルス応答出力部190とを含む。 The estimation system includes an acquisition unit 110, an estimation device 100, and an impulse response output unit 190.
 推定システムは、取得部110を介して収音を行う空間の点群データを取得し、点群データを用いて所定の空間の反射率を推定し、反射率の推定値iを用いて、所定の空間における、任意の音源位置と任意のマイクロホン位置とのインパルス応答を推定し、出力する。 The estimation system acquires point cloud data of the space in which sound is to be collected via the acquisition unit 110, estimates the reflectance of a predetermined space using the point cloud data, and uses the estimated value i of the reflectance to obtain a predetermined It estimates and outputs the impulse response of an arbitrary sound source position and an arbitrary microphone position in the space of .
 推定装置100は、例えば、中央演算処理装置(CPU: Central Processing Unit)、主記憶装置(RAM: Random Access Memory)などを有する公知又は専用のコンピュータに特別なプログラムが読み込まれて構成された特別な装置である。推定装置100は、例えば、中央演算処理装置の制御のもとで各処理を実行する。推定装置100に入力されたデータや各処理で得られたデータは、例えば、主記憶装置に格納され、主記憶装置に格納されたデータは必要に応じて中央演算処理装置へ読み出されて他の処理に利用される。推定装置100の各処理部は、少なくとも一部が集積回路等のハードウェアによって構成されていてもよい。推定装置100が備える各記憶部は、例えば、RAM(Random Access Memory)などの主記憶装置、またはリレーショナルデータベースやキーバリューストアなどのミドルウェアにより構成することができる。ただし、各記憶部は、必ずしも推定装置100がその内部に備える必要はなく、ハードディスクや光ディスクもしくはフラッシュメモリ(Flash Memory)のような半導体メモリ素子により構成される補助記憶装置により構成し、推定装置100の外部に備える構成としてもよい。 The estimating device 100 is, for example, a special computer configured by reading a special program into a known or dedicated computer having a central processing unit (CPU: Central Processing Unit), a main memory (RAM: Random Access Memory), etc. It is a device. The estimating apparatus 100 executes each process under the control of a central processing unit, for example. Data input to the estimating device 100 and data obtained in each process are stored, for example, in a main memory device, and the data stored in the main memory device are read out to the central processing unit as necessary and used for other purposes. used to process At least a part of each processing unit of the estimation device 100 may be configured by hardware such as an integrated circuit. Each storage unit included in the estimation device 100 can be configured by, for example, a main storage device such as RAM (Random Access Memory), or middleware such as a relational database or a key-value store. However, each storage unit does not necessarily have to be provided inside the estimating apparatus 100, and is configured by an auxiliary storage device configured by a semiconductor memory device such as a hard disk, an optical disk, or a flash memory, and the estimating apparatus 100 may be provided outside.
 以下、各部について説明する。 Each part will be explained below.
<取得部110>
 取得部110は、収音を行う空間の点群データRを取得し(S110)、出力する。図3は取得部110の機能ブロック図を、図4はその処理フローの例を示す。例えば、取得部110は、空間センシング部111とノイズ除去部113と空間モデル結合部115とを含む。
<Acquisition unit 110>
The acquiring unit 110 acquires the point cloud data R of the space in which sound is to be collected (S110) and outputs it. FIG. 3 is a functional block diagram of the acquisition unit 110, and FIG. 4 shows an example of its processing flow. For example, the acquisition unit 110 includes a spatial sensing unit 111 , a noise removing unit 113 and a spatial model combining unit 115 .
<空間センシング部111>
 空間センシング部111は、収音を行う空間の点群データを取得し(S111)、出力する。例えば、空間センシング部111は、収音を行う空間内に設置されたLiDARからなり、光を対象物に向けて発射し、発光後反射光を受光するまでの時間の差により対象物までの距離を求め、発射方向から対象物の方向を求める。この距離や方向が点群データで表される。空間センシング技術としては、様々な従来技術を用いることができる。例えば、既存の空間センシング技術として参考文献1が知られている。
<Spatial Sensing Unit 111>
The space sensing unit 111 acquires point cloud data of the space in which sound is to be collected (S111) and outputs it. For example, the space sensing unit 111 consists of a LiDAR installed in a space where sound is collected. and find the direction of the object from the launch direction. This distance and direction are represented by point cloud data. Various conventional techniques can be used as spatial sensing techniques. For example, Reference 1 is known as an existing space sensing technology.
(参考文献1) 伊東 敏夫、「自動運転のためのLiDAR技術の原理と活用法」、科学情報出版株式会社、2020年
<ノイズ除去部113>
 ノイズ除去部113は、点群データを受け取り、点群データに含まれるノイズを除去し(S113)、除去後の点群データを出力する。ノイズ除去技術としては、様々な従来技術を用いることができる。例えば、既存のノイズ除去技術として参考文献2が知られている。
(Reference 1) Toshio Ito, "Principles and Applications of LiDAR Technology for Autonomous Driving," Science Information Publishing Co., Ltd., 2020 <Noise Removal Unit 113>
The noise removal unit 113 receives the point cloud data, removes noise included in the point cloud data (S113), and outputs the removed point cloud data. Various conventional techniques can be used as the noise removal technique. For example, Reference 2 is known as an existing noise removal technique.
(参考文献2)Rusu, R. B., Z. C. Marton, N. Blodow, M. Dolha, and M. Beetz, "Towards 3D Point Cloud Based Object Maps for Household Environments", Robotics and Autonomous Systems Journal, 2008.
 なお、図5および図6は、ノイズ除去後の点群データの例を表す。図5は空間センシング部111(例えばLiDAR)を仰角0度、図6は仰角-90度として取得した点群データである。
(Reference 2) Rusu, R. B., Z. C. Marton, N. Blodow, M. Dolha, and M. Beetz, "Towards 3D Point Cloud Based Object Maps for Household Environments", Robotics and Autonomous Systems Journal, 2008.
5 and 6 show examples of point cloud data after noise removal. FIG. 5 shows point cloud data acquired with the spatial sensing unit 111 (for example, LiDAR) at an elevation angle of 0 degrees, and FIG. 6 shows point cloud data obtained with an elevation angle of −90 degrees.
<空間モデル結合部115>
 空間モデル結合部115は、複数の点群データを受け取り、複数の点群データを結合して、空間の形状を復元し(S115)、空間の形状を示す点群データRを出力する。例えば、LiDARを回転させながら、仰角の異なる複数の点群データを得、反復最近接点(ICP: Iterative Closest Point)アルゴリズムにより複数の点群データを組み合わせて3次元シーンを再構成する。空間モデル結合技術としては、様々な従来技術を用いることができる。例えば、既存の空間モデル結合技術として参考文献3が知られている。
<Spatial model combining unit 115>
The space model combining unit 115 receives a plurality of point cloud data, combines the plurality of point cloud data, restores the shape of the space (S115), and outputs point cloud data R representing the shape of the space. For example, while rotating the LiDAR, multiple point cloud data with different elevation angles are obtained, and the multiple point cloud data are combined using the Iterative Closest Point (ICP) algorithm to reconstruct a 3D scene. Various conventional techniques can be used as the spatial model combination technique. For example, Reference 3 is known as an existing spatial model combining technique.
(参考文献3) Szymon.R and Marc.L, "Efficient Variants of the ICP Algorithm", Proceedings Third International Conference on 3-D Digital Imaging and Modeling, 2001, pp. 145-152.
<推定装置100>
 推定装置100は、空間の形状を示す点群データRを受け取り、点群データRを用いて所定の空間の反射率を推定し、反射率の推定値iを出力する。
(Reference 3) Szymon.R and Marc.L, "Efficient Variants of the ICP Algorithm", Proceedings Third International Conference on 3-D Digital Imaging and Modeling, 2001, pp. 145-152.
<Estimation device 100>
The estimating apparatus 100 receives point cloud data R indicating the shape of a space, estimates the reflectance of a given space using the point cloud data R, and outputs an estimated value i of the reflectance.
 例えば、推定装置100は、部屋形状作成部120、パラメータ入力部130、インパルス応答推定部140、残響時間計算部160、実環境インパルス応答取得部150、評価関数計算部170およびパラメータ推定部180を含む(図1参照)。 For example, the estimation device 100 includes a room shape generator 120, a parameter input unit 130, an impulse response estimator 140, a reverberation time calculator 160, a real environment impulse response acquirer 150, an evaluation function calculator 170, and a parameter estimator 180. (See Figure 1).
 推定装置100は、推定装置100を構成するコンピュータ内に取得部110を含んでもよいし、含まなくともよい。取得部110を含まない場合には、推定装置100は図示しない点群データを入力するための入力インターフェースを含む。取得部110を含む場合には、推定装置100を構成するコンピュータ内で、処理S113および処理S115、または、処理S115を行ってもよい。この場合、コンピュータ内で行う処理に対応する部分の前段に図示しない入力部が設けられ、入力部は収音を行う空間の点群データに基づく情報(空間センシング部111の出力する点群データそのもの、または、ノイズ除去部113の出力するノイズを除去した点群データ)を受け取る。 The estimating device 100 may or may not include the obtaining unit 110 in the computer that constitutes the estimating device 100 . When the acquisition unit 110 is not included, the estimation device 100 includes an input interface for inputting point cloud data (not shown). When the acquisition unit 110 is included, the processing S113 and the processing S115, or the processing S115 may be performed within the computer that constitutes the estimation device 100 . In this case, an input unit (not shown) is provided before the part corresponding to the processing performed in the computer. , or noise-removed point cloud data output from the noise removal unit 113).
<部屋形状作成部120>
 部屋形状作成部120は、空間の形状を示す点群データRを受け取り、点群データRに現れない空間の形状を補完し(S120)、補完した空間の形状Uを出力する。
<Room shape creation unit 120>
The room shape generator 120 receives the point cloud data R indicating the shape of the space, complements the shape of the space that does not appear in the point cloud data R (S120), and outputs the shape U of the complemented space.
 図7は部屋形状作成部120の機能ブロック図を、図8はその処理フローの例を示す。 FIG. 7 is a functional block diagram of the room shape creation unit 120, and FIG. 8 shows an example of its processing flow.
 例えば、部屋形状作成部120は、壁面抽出部121と欠損補完部123とを含む。 For example, the room shape creation unit 120 includes a wall surface extraction unit 121 and a loss complement unit 123.
<壁面抽出部121>
 壁面抽出部121は、空間の形状を示す点群データRを受け取り、空間の形状を用いて、空間を形成する壁面を抽出し(S121)、出力する。
<Wall surface extraction unit 121>
The wall surface extracting unit 121 receives the point cloud data R indicating the shape of the space, uses the shape of the space to extract the wall surfaces forming the space (S121), and outputs them.
 例えば、壁面抽出部121は、RANSAC(Random Sample Consensus)を壁面抽出用に改善したアルゴリズムを利用して、空間を形成する壁面のパラメータを取得することで、壁面を抽出する。以下にその例を示す。 For example, the wall surface extraction unit 121 extracts the wall surface by acquiring the parameters of the wall surface forming the space using an algorithm that is an improvement of RANSAC (Random Sample Consensus) for wall surface extraction. Examples are shown below.
 (1)空間の形状を示す点群データRを、観測角度に基づき、分割する。例えば、水平面において、M個の方向に分割する。 (1) Divide the point cloud data R, which indicates the shape of the space, based on the observation angle. For example, in the horizontal plane, it is divided into M directions.
 (2)M分割した方向の中からm番目の方向(ROI(region of interest))を指定する。ただし、m=1,2,…,Mとする。 (2) Specify the m-th direction (ROI (region of interest)) from among the M-divided directions. However, m=1,2,...,M.
 (3)ROIに属する点群から3点a,b,cをランダムに取得する。 (3) Randomly acquire 3 points a, b, and c from the point cloud belonging to the ROI.
 (4)3点a,b,cから3点a,b,cを含む平面の方程式ax+by+cz+d=0の係数を計算する。例えば、ベクトルab,acの外積から係数を計算する。 (4) Calculate the coefficients of the plane equation ax+by+cz+d=0 from 3 points a, b, c to 3 points a, b, c. For example, coefficients are calculated from the outer product of vectors ab and ac.
 (5)ROIに属する点群の各点と平面との距離を計算し、距離が所定の閾値以下となる点の数nqを求める。ただし、q=1,2,…,Qとする。例えば、d=n(p-a)により距離dを計算する。nは3点a,b,cを含む平面の法線ベクトルであり、pはROIに属する点群の各点である。 (5) Calculate the distance between each point of the point cloud belonging to the ROI and the plane, and obtain the number n q of points whose distance is equal to or less than a predetermined threshold. However, q=1,2,...,Q. For example, calculate the distance d by d=n(pa). n is the normal vector of the plane containing the three points a, b, c, and p is each point of the point cloud belonging to the ROI.
 (6)上述の(3)~(5)をQ回繰り返し、Q個の数nqの中から最大の点の数nmとその数に対応する係数をその方向mのパラメータとする。 (6) The above (3) to (5) are repeated Q times, and the maximum number n m of points out of the Q numbers n q and the coefficient corresponding to that number are taken as parameters for the direction m.
 (7)上述の(2)~(6)をM回繰り返し、M個の数nmの中から最大の点の数の方向を選択する。 (7) The above (2) to (6) are repeated M times, and the direction with the maximum number of points is selected from the M numbers nm .
 (8)上述の(7)で選択した方向に対応する係数で示される平面との距離が閾値以下の点を集め、集めた点群から法線ベクトルを求める。例えば、SVD(Singular Value Decomposition)や共分散行列のPCA(Principal Component Analysis)で法線ベクトルを求めることができる。 (8) Collect points whose distance from the plane indicated by the coefficient corresponding to the direction selected in (7) above is equal to or less than the threshold, and obtain the normal vector from the collected point group. For example, the normal vector can be obtained by SVD (Singular Value Decomposition) or PCA (Principal Component Analysis) of the covariance matrix.
 (9)法線ベクトルを点群データの中で最大の壁面として記録する。 (9) Record the normal vector as the largest wall surface in the point cloud data.
 (10)最大の壁面に所属する点群データをデータから除き、(1)へ戻る。なお、最大の壁面に所属する点群データが、抽出された壁面を表す。 (10) Remove the point cloud data belonging to the largest wall surface from the data and return to (1). Note that the point cloud data belonging to the largest wall surface represents the extracted wall surface.
 なお、最大の壁面に所属する点群データが全体の点群データRに占める割合がある所定の値以下になる場合、そこには壁面がないものとする。つまり、ある方向において、上述の(5)の閾値以下となる点の数nqの割合が所定の値以下になる場合には、そこに壁面がないものとする。さらに、上述の(10)で最大の壁面に所属する点群データをデータから除くため、抽出できる壁面が徐々に減っていき、自動的に収束する。なお、このような構成により、壁や天井だけでなく、反射がありそうなある程度の点群の平面形状を抽出し、結果として家具や機材もある空間として模擬する。例えば、最大の壁面に所属する点群データが為す平面が所定の閾値(例えば20cm四方)以下になるまで処理を繰り返すように制御してもよい。所定の閾値は処理量に合わせて推定すればよい。 In addition, when the ratio of the point cloud data belonging to the largest wall surface to the entire point cloud data R is less than or equal to a predetermined value, it is assumed that there is no wall surface there. That is, if the ratio of the number nq of points below the threshold in (5) above in a certain direction is below a predetermined value, it is assumed that there is no wall there. Furthermore, in (10) above, since the point cloud data belonging to the largest wall surface is removed from the data, the number of wall surfaces that can be extracted gradually decreases and converges automatically. With such a configuration, not only the walls and ceiling but also the plane shape of the point cloud that is likely to reflect are extracted to some extent, and as a result, furniture and equipment are simulated as a space. For example, the process may be controlled to be repeated until the plane formed by the point cloud data belonging to the largest wall surface becomes equal to or less than a predetermined threshold value (for example, 20 cm square). The predetermined threshold may be estimated according to the amount of processing.
<欠損補完部123>
 欠損補完部123は、空間を形成する壁面を受け取り、空間に壁面以外の物体が存在し、壁面の一部の形状を推定できない場合、推定できた壁面の他部の形状から壁面の一部の形状を推定し、推定した壁面の一部を用いて空間の形状を補完し(S123)、補完後の空間の形状Uを出力する。
<Deficit Compensation Unit 123>
The loss complementing unit 123 receives the wall surface forming the space, and if there is an object other than the wall surface in the space and the shape of the part of the wall surface cannot be estimated, the shape of the part of the wall surface can be estimated from the estimated shape of the other part of the wall surface. The shape is estimated, the shape of the space is complemented using the estimated part of the wall surface (S123), and the shape U of the space after complementation is output.
 上述の通り、LiDARはレーザの反射を利用して距離を計測しているため、遮蔽物がありレーザが届かない領域ではデータの欠損が生じる。そこで、欠損データを推定できた壁面の情報から補完し、頂点(部屋の角)を求める。 As described above, since LiDAR uses laser reflection to measure distance, data loss occurs in areas where the laser cannot reach due to obstructions. Therefore, the missing data is supplemented from the information of the estimated wall surface, and the vertex (corner of the room) is obtained.
 図9の例では、ディスプレイDが遮蔽物となり、図中、破線で囲まれている部分に存在している壁面が観測できない。なお、図9は、仰角0度の水平面の点群データであり、点OはLiDARの設置位置である。  In the example of Fig. 9, the display D becomes a shield, and the walls that exist in the part surrounded by the dashed line in the figure cannot be observed. Note that FIG. 9 shows point cloud data on a horizontal plane with an elevation angle of 0 degrees, and point O is the installation position of the LiDAR.
 例えば、平面近似によって壁面(a,b,c,d)のパラメータを抽出し、直線から壁の頂点座標を求める。つまり、図10のように、直線の交点Pを壁の頂点座標として求め、図10の破線部分に壁面があるものとして扱う。 For example, the parameters of the wall surface (a, b, c, d) are extracted by plane approximation, and the vertex coordinates of the wall are obtained from the straight line. That is, as shown in FIG. 10, the point of intersection P of the straight lines is determined as the vertex coordinate of the wall, and the broken line portion in FIG. 10 is assumed to be the wall surface.
<パラメータ入力部130>
 パラメータ入力部130は、音源位置Sとマイクロホン位置Mとを入力として受け付け(S130)、インパルス応答推定部140に出力する。所定の空間内を模擬したシミュレーション空間における音源位置Sおよびマイクロホン位置Mに対応する現実空間における位置にスピーカおよびマイクロホンが配置される。そのため、音源位置Sおよびマイクロホン位置Mは、現実空間において、スピーカおよびマイクロホンが配置可能な位置である。音源位置Sおよびマイクロホン位置Mとしてそれぞれ1つ以上の位置を想定することができる。パラメータ入力部130は例えばキーボードやマウス等の入力装置であり、音源位置はパラメータ入力部130を介して利用者によって入力される。例えば、パラメータ入力部130は、欠損補完部123の出力である補完後の空間の形状を受け取り(図1中、破線で示す)、補完後の空間の形状をディスプレイ等の表示装置を介して利用者に提示し、利用者が、マウス等を利用して補完後の空間のどこに音源を配置するかを指定する構成としてもよい。または、LiDARから取得した点群データRから参考文献4のような物体認識を利用し、自動で入力する構成としてもよいし、認識した物体の位置から人手に選択する構成としてもよい。
<Parameter input unit 130>
The parameter input unit 130 receives the sound source position S and the microphone position M as inputs (S 130 ) and outputs them to the impulse response estimation unit 140 . A speaker and a microphone are arranged at positions in the real space corresponding to the sound source position S and the microphone position M in the simulation space simulating the inside of a predetermined space. Therefore, the sound source position S and the microphone position M are positions where the speaker and the microphone can be arranged in the real space. One or more positions can be envisaged as the sound source position S and the microphone position M respectively. A parameter input unit 130 is an input device such as a keyboard or a mouse, and the user inputs the sound source position via the parameter input unit 130 . For example, the parameter input unit 130 receives the shape of the space after complementing, which is the output of the loss complementing unit 123 (indicated by the dashed line in FIG. 1), and uses the shape of the space after complementing via a display device such as a display. A configuration may also be adopted in which the sound source is presented to the user, and the user uses a mouse or the like to specify where in the space after complementation the sound source is to be placed. Alternatively, it may be configured to automatically input using object recognition such as Reference 4 from point cloud data R acquired from LiDAR, or to manually select from the position of the recognized object.
(参考文献4) D. Maturana and S. Scherer, "VoxNet: A 3D Convolutional Neural Network for real-time object recognition", 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, 2015, pp. 922-928, doi: 10.1109/IROS.2015.7353481.
 また、パラメータ入力部130は記憶媒体の読み取り装置であり、推定装置100はパラメータ入力部130を介して音源位置とマイクロホン位置を記憶した記憶媒体を読み取ることで音源位置とマイクロホン位置を入力として受け付けてもよい。
(Reference 4) D. Maturana and S. Scherer, "VoxNet: A 3D Convolutional Neural Network for real-time object recognition", 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, 2015, pp. 922-928, doi: 10.1109/IROS.2015.7353481.
The parameter input unit 130 is a storage medium reading device, and the estimation apparatus 100 reads the storage medium storing the sound source positions and the microphone positions via the parameter input unit 130, thereby accepting the sound source positions and the microphone positions as inputs. good too.
 あるいは、人手によって事前に音源位置とマイクロホン位置を推定装置100に入力しておいてもよい。 Alternatively, the sound source position and the microphone position may be manually input to the estimation device 100 in advance.
<インパルス応答推定部140>
 インパルス応答推定部140は、音源位置Sと空間の形状Uと受聴位置Mと反射率の推定値Rまたは初期値Riniとを入力とし、これらの値を用いて、空間の形状Uにおける音源位置Sと受聴位置Mのインパルス応答を推定し(S140)、インパルス応答の推定値ieを出力する。初期値Riniは事前に人手により入力されるものとする。例えば、Rini=0.5とする。反射率の推定値Rは、後述するパラメータ推定部180の出力である。
<Impulse response estimator 140>
The impulse response estimation unit 140 receives the sound source position S, the space shape U, the listening position M, and the estimated value R or the initial value R ini of the reflectance, and uses these values to calculate the sound source position in the space shape U. The impulse response of S and the listening position M is estimated (S140), and the estimated value i e of the impulse response is output. Assume that the initial value R ini is manually input in advance. For example, R ini =0.5. The reflectance estimate value R is the output of the parameter estimator 180, which will be described later.
 例えば、インパルス応答推定部140は、音源位置Sと空間の形状Uと反射率の推定値Rまたは初期値Riniと受聴位置Mとを用いて、音源位置Sでの空間における受聴位置Mの空間伝達関数を算出する。空間伝達関数算出技術としては、様々な従来技術を用いることができる。例えば、空間の形状Uと反射率の推定値Rまたは初期値Riniとから音波伝搬をシミュレーションし、受聴位置Mでの到来音を予測するためFDTD法(finite-difference time-domain method)によるシミュレーションを行い、空間伝達関数を算出する。さらに、インパルス応答推定部140は、空間伝達関数を用いてインパルス応答を推定する。 For example, the impulse response estimating unit 140 uses the sound source position S, the space shape U, the estimated value R or initial value R ini of the reflectance, and the listening position M to calculate the space of the listening position M in the space at the sound source position S. Calculate the transfer function. Various conventional techniques can be used as the spatial transfer function calculation technique. For example, the FDTD method (finite-difference time-domain method) is used to simulate the sound wave propagation from the space shape U and the estimated value R or initial value R ini of the reflectance, and to predict the incoming sound at the listening position M. to calculate the spatial transfer function. Furthermore, the impulse response estimator 140 estimates the impulse response using the spatial transfer function.
<実環境インパルス応答取得部150>
 実環境インパルス応答取得部150は、現実空間である所定の空間において、音源位置Sに対応する位置から発せられる音響信号を受聴位置Mに対応する位置で収音して、取得し、音源位置Sに対応する位置と受聴位置Mに対応する位置のインパルス応答を測定し(S150)、インパルス応答の測定値imを出力する。例えば、現実空間である所定の空間の音源位置Sに対応する位置に配置したスピーカからTSP(Time Stretched Pulse)信号を再生し、現実空間である所定の空間の受聴位置に対応する位置に配置したマイクロホンで収音し、インパルス応答を測定する。マイクとスピーカは、別装置であってもよいし、スマートスピーカなどの一体型の装置であってもよい。
<Real Environment Impulse Response Acquisition Unit 150>
The real environment impulse response acquisition unit 150 picks up and acquires an acoustic signal emitted from a position corresponding to the sound source position S at a position corresponding to the listening position M in a predetermined space that is the real space, and obtains the sound source position S and the position corresponding to the listening position M are measured (S150), and the measured value im of the impulse response is output. For example, a TSP (Time Stretched Pulse) signal is reproduced from a speaker placed at a position corresponding to the sound source position S in a given space, which is the real space, and placed at a position corresponding to the listening position in the given space, which is the real space. Pick up sound with a microphone and measure the impulse response. The microphone and speaker may be separate devices, or may be an integrated device such as a smart speaker.
<残響時間計算部160>
 残響時間計算部160は、インパルス応答の推定値ieとインパルス応答の測定値imとを入力とし、インパルス応答の推定値ieの残響時間reと、インパルス応答の測定値imの残響時間riとを計算し(S160)、出力する。例えば、残響時間としてRT60を用いる。RT60とは、部屋の残響が60dB下がるまでにかかる時間である。RT60を利用することで、インパルス応答のように場所や音源位置で過敏に変化せず、部屋の構造や材質による数値として頑健に得られることから利用している。
<Reverberation time calculator 160>
The reverberation time calculator 160 receives the impulse response estimated value i e and the impulse response measured value im as inputs, and calculates the reverberation time r e of the impulse response estimated value i e and the reverberation time of the impulse response measured value im Calculate time r i (S160) and output. For example, RT60 is used as the reverberation time. RT60 is the time it takes for the reverberation in the room to drop by 60dB. By using the RT60, unlike the impulse response, it does not change sensitively depending on the location or sound source position.
 インパルス応答推定部140と残響時間計算部160の組合せを、所定の空間内を模擬したシミュレーション空間における所定の位置(音源位置)から発せられた音響信号の残響に基づく値(RT60)を取得するため、シミュレーション取得部ともいう。 A combination of the impulse response estimator 140 and the reverberation time calculator 160 is used to obtain a value (RT60) based on reverberation of an acoustic signal emitted from a predetermined position (sound source position) in a simulation space simulating a predetermined space. , is also called a simulation acquisition unit.
 また、実環境インパルス応答取得部150と残響時間計算部160の組合せを、現実空間である所定の空間において、所定の位置に対応する位置から発せられた音響信号の残響に基づく値(RT60)を取得するため、現実空間取得部ともいう。 Also, the combination of the real environment impulse response acquisition unit 150 and the reverberation time calculation unit 160 is used to obtain a value (RT60) based on the reverberation of an acoustic signal emitted from a position corresponding to a given position in a given space that is the real space. Since it acquires, it is also called a real space acquisition unit.
<評価関数計算部170>
 評価関数計算部170は、残響時間reと残響時間rmとを入力とし、残響時間reと残響時間rmの差を用いて評価値Eを計算し(S170)、出力する。評価値としては、例えば、インパルス応答の推定値ieのRT60と、インパルス応答の測定値imのRT60との二乗誤差が考えられ、次式で表される。
<Evaluation function calculator 170>
The evaluation function calculator 170 receives the reverberation time r e and the reverberation time r m as inputs, calculates the evaluation value E using the difference between the reverberation time r e and the reverberation time r m (S170), and outputs it. As an evaluation value, for example, the square error between RT60 of the estimated value i e of the impulse response and RT60 of the measured value im of the impulse response can be considered, and is represented by the following equation.
E=(re-rm)2
<パラメータ推定部180>
 パラメータ推定部180は、評価値Eを入力とし、評価値を利用して壁面の反射率を推定し(S180)、推定値Rを出力する。
E=(r e -r m ) 2
<Parameter estimation unit 180>
The parameter estimator 180 receives the evaluation value E, estimates the reflectance of the wall surface using the evaluation value (S180), and outputs the estimated value R.
 本実施形態では、パラメータ推定部180は、ガウス過程によるパラメータ推定を行う。 In this embodiment, the parameter estimation unit 180 performs parameter estimation using a Gaussian process.
 ガウス過程によるパラメータ推定に必要な項目として、評価値Eと操作するパラメータを用いる。例えば、操作するパラメータは各壁面の反射率である。 The evaluation value E and the parameter to be manipulated are used as items necessary for parameter estimation by Gaussian process. For example, the manipulated parameter is the reflectance of each wall.
 観測点を利用してガウス分布に従う回帰曲線を真値の範囲に絞る。 Use the observation points to narrow down the regression curve that follows the Gaussian distribution to the range of true values.
 どの点を取得していけば最大値が求まるかをガウス過程に従い効率よく探索する。 Efficiently search for which points to obtain to obtain the maximum value according to the Gaussian process.
 図11は、ガウス過程によるパラメータ推定を説明するための図である。図の縦軸は評価値を表し、横軸は反射率のインデックスを表す。 FIG. 11 is a diagram for explaining parameter estimation by a Gaussian process. The vertical axis of the figure represents the evaluation value, and the horizontal axis represents the reflectance index.
 パラメータ推定部180は、各パラメータによって評価関数が一意に求まり、ガウス分布によって回帰曲線を予測し、評価値Eが最小となる予測されるパラメータ(反射率)に逐次更新し、高速な評価関数の最低値の探索が可能となっている。 The parameter estimating unit 180 uniquely obtains an evaluation function from each parameter, predicts a regression curve using a Gaussian distribution, sequentially updates the predicted parameter (reflectance) that minimizes the evaluation value E, and uses a high-speed evaluation function. It is possible to search for the lowest value.
 パラメータ推定部180は、繰り返し回数が所定の回数以下の場合には(S185のNO)、更新したパラメータをインパルス応答推定部140に出力し、処理S140~S180を行い、パラメータ推定を繰り返す。 When the number of iterations is equal to or less than the predetermined number (NO in S185), the parameter estimation section 180 outputs the updated parameters to the impulse response estimation section 140, performs processes S140 to S180, and repeats parameter estimation.
 パラメータ推定部180は、繰り返し回数が所定の回数を超えた場合には(S185のYES)、最小の評価関数となったパラメータを、最終的な反射率の推定値Rfとして出力する。 When the number of iterations exceeds a predetermined number (YES in S185), the parameter estimator 180 outputs the parameter with the minimum evaluation function as the final reflectance estimated value Rf .
<インパルス応答出力部190>
 インパルス応答出力部190は、音源位置S'と空間の形状Uと受聴位置M'と最終的な反射率の推定値Rfとを入力とし、これらの値を用いて、音源位置S'での空間における受聴位置M'のインパルス応答を推定し(S190)、インパルス応答の推定値iを出力する。例えば、インパルス応答推定部140と同様の方法によりインパルス応答を推定する。音源位置S'と受聴位置M'とを変更しながら様々な位置におけるインパルス応答を推定することができる。
<Impulse Response Output Unit 190>
The impulse response output unit 190 receives the sound source position S′, the spatial shape U, the listening position M′, and the final reflectance estimate value R f as inputs, and uses these values to calculate the response at the sound source position S′. The impulse response of the listening position M' in space is estimated (S190), and the estimated value i of the impulse response is output. For example, the impulse response is estimated by the same method as the impulse response estimator 140 . Impulse responses at various positions can be estimated while changing the sound source position S' and the listening position M'.
 推定装置100は、推定装置100を構成するコンピュータ内にインパルス応答出力部190を含んでもよいし、含まなくともよい。 The estimating device 100 may or may not include the impulse response output unit 190 in the computer configuring the estimating device 100 .
<効果>
 以上の構成により、短時間で反射率を推定することができる。また、RT60を利用することでインパルス応答に比べて音源の位置やマイクロホンの位置による影響を減らし、計算量の削減を可能とする。
<effect>
With the configuration described above, the reflectance can be estimated in a short time. In addition, by using RT60, it is possible to reduce the amount of calculation by reducing the influence of the position of the sound source and the position of the microphone compared to the impulse response.
 スマートスピーカなどに含まれるマイクロホンとスピーカ、および、LiDARなどの測位機能を持つ機器を用いて、本実施形態を適用することで、周囲の正確な構造と構造物の反射率などのパラメータを推定することができる。さらに、利用場所の推定や、正確なシミュレーション情報から音響処理への応用(雑音抑圧、残響抑圧)が可能となる。 By applying this embodiment using microphones and speakers included in smart speakers and equipment with positioning functions such as LiDAR, accurate parameters such as the surrounding structures and the reflectance of structures can be estimated. be able to. Furthermore, it is possible to estimate the place of use and apply the accurate simulation information to acoustic processing (noise suppression, reverberation suppression).
<変形例>
 本実施形態では、取得部110の出力値である空間の形状Uをインパルス応答推定部140の入力として用いているが、建物の設計時に作成したBIMデータ(建築CADデータ)をインパルス応答推定部140の入力として用いてもよい。この場合、取得部110および部屋形状作成部120を含まなくともよい。
<Modification>
In this embodiment, the spatial shape U, which is the output value of the acquisition unit 110, is used as the input of the impulse response estimation unit 140. may be used as an input for In this case, acquisition unit 110 and room shape creation unit 120 may not be included.
 本実施形態では、音響信号の残響に基づく値としてRT60を利用しているが、音響信号の残響に基づく値としてインパルス応答を利用してもよい。この場合、推定装置100は、残響時間計算部160を含まなくともよい。ただし、前述の通り、RT60を利用することで、インパルス応答に比べて音源の位置やマイクロホンの位置による影響を減らし、計算量の削減を可能とする。 Although RT60 is used as the value based on the reverberation of the acoustic signal in this embodiment, the impulse response may be used as the value based on the reverberation of the acoustic signal. In this case, estimation device 100 does not need to include reverberation time calculation section 160 . However, as mentioned above, by using RT60, the influence of the position of the sound source and the position of the microphone can be reduced compared to the impulse response, and the amount of calculation can be reduced.
 本実施形態では、評価値としては、E=(re-rm)2を用いているが、他の評価値を用いてもよい。例えば、E=-(re-rm)2を用いてもよい。この場合、パラメータ推定部180では、最大の評価関数となったパラメータを、最終的な反射率の推定値Rfとして出力する。 Although E=(r e −r m ) 2 is used as the evaluation value in this embodiment, other evaluation values may be used. For example, E=-(r e -r m ) 2 may be used. In this case, the parameter estimator 180 outputs the parameter with the maximum evaluation function as the final reflectance estimated value R f .
要は、シミュレーション空間における残響に基づく値と、現実空間における残響に基づく値と、の差を用い、(i)この差が大きくなるほど大きくなる値を用いて、最小の評価関数となったパラメータを、最終的な反射率の推定値Rfとするか、(ii)この差が大きくなるほど小さくなる値を用いて、最大の評価関数となったパラメータを、最終的な反射率の推定値Rfとすればよい。 In short, the difference between the value based on the reverberation in the simulation space and the value based on the reverberation in the real space is used. , the final estimated reflectance value R f And it is sufficient.
<シミュレーション結果>
 FDTD法で推定すると非常に時間がかかるため、本シミュレーションでは鏡像法を利用する。鏡像法にはPyroomacousticsを利用する。反射率の真値を0.15とし、評価関数には、インパルス応答のL2ノルムではなく、RT60のノルムを利用する。
<Simulation result>
Estimation by the FDTD method takes a very long time, so the mirror image method is used in this simulation. Pyroomacoustics is used for the mirror image method. The true value of the reflectance is set to 0.15, and the RT60 norm is used as the evaluation function instead of the L2 norm of the impulse response.
 1変数の推定の結果、RT60の差を用いることで、インパルス応答の差を直接計算するより精度は落ちるが、真値に近づくことが確認された。実測値との差を埋める場合はRT60がインパルス応答に比べて誤差が少ないことが分かった。 As a result of estimating one variable, it was confirmed that using the RT60 difference approached the true value, although the accuracy was lower than directly calculating the impulse response difference. It was found that RT60 has less error than the impulse response when filling the difference with the actual measurement value.
 図12は、ガウス過程による更新と評価関数および推定したパラメータ値の表である。なお、図12中、iterは繰り返し回数を、targetは評価関数を、var_absは反射率を表す。真値0.15が繰り返し回数13回で評価関数を最大としている。本シミュレーションでは、評価関数(二乗誤差)にマイナスを掛けることで、評価関数が最大になる変数を探索している。 FIG. 12 is a table of updates by Gaussian processes, evaluation functions, and estimated parameter values. In FIG. 12, iter represents the number of iterations, target represents the evaluation function, and var_abs represents the reflectance. A true value of 0.15 maximizes the evaluation function at 13 iterations. In this simulation, the evaluation function (squared error) is multiplied by a negative value to search for a variable that maximizes the evaluation function.
<その他の変形例>
 本発明は上記の実施形態及び変形例に限定されるものではない。例えば、上述の各種の処理は、記載に従って時系列に実行されるのみならず、処理を実行する装置の処理能力あるいは必要に応じて並列的にあるいは個別に実行されてもよい。その他、本発明の趣旨を逸脱しない範囲で適宜変更が可能である。
<Other Modifications>
The present invention is not limited to the above embodiments and modifications. For example, the various types of processing described above may not only be executed in chronological order according to the description, but may also be executed in parallel or individually according to the processing capacity of the device that executes the processing or as necessary. In addition, appropriate modifications are possible without departing from the gist of the present invention.
<プログラム及び記録媒体>
 上述の各種の処理は、図13に示すコンピュータの記憶部2020に、上記方法の各ステップを実行させるプログラムを読み込ませ、制御部2010、入力部2030、出力部2040などに動作させることで実施できる。
<Program and recording medium>
The various processes described above can be performed by loading a program for executing each step of the above method into the storage unit 2020 of the computer shown in FIG. .
 この処理内容を記述したプログラムは、コンピュータで読み取り可能な記録媒体に記録しておくことができる。コンピュータで読み取り可能な記録媒体としては、例えば、磁気記録装置、光ディスク、光磁気記録媒体、半導体メモリ等どのようなものでもよい。 A program that describes this process can be recorded on a computer-readable recording medium. Any computer-readable recording medium may be used, for example, a magnetic recording device, an optical disk, a magneto-optical recording medium, a semiconductor memory, or the like.
 また、このプログラムの流通は、例えば、そのプログラムを記録したDVD、CD-ROM等の可搬型記録媒体を販売、譲渡、貸与等することによって行う。さらに、このプログラムをサーバコンピュータの記憶装置に格納しておき、ネットワークを介して、サーバコンピュータから他のコンピュータにそのプログラムを転送することにより、このプログラムを流通させる構成としてもよい。 In addition, the distribution of this program is carried out, for example, by selling, assigning, lending, etc. portable recording media such as DVDs and CD-ROMs on which the program is recorded. Further, the program may be distributed by storing the program in the storage device of the server computer and transferring the program from the server computer to other computers via the network.
 このようなプログラムを実行するコンピュータは、例えば、まず、可搬型記録媒体に記録されたプログラムもしくはサーバコンピュータから転送されたプログラムを、一旦、自己の記憶装置に格納する。そして、処理の実行時、このコンピュータは、自己の記録媒体に格納されたプログラムを読み取り、読み取ったプログラムに従った処理を実行する。また、このプログラムの別の実行形態として、コンピュータが可搬型記録媒体から直接プログラムを読み取り、そのプログラムに従った処理を実行することとしてもよく、さらに、このコンピュータにサーバコンピュータからプログラムが転送されるたびに、逐次、受け取ったプログラムに従った処理を実行することとしてもよい。また、サーバコンピュータから、このコンピュータへのプログラムの転送は行わず、その実行指示と結果取得のみによって処理機能を実現する、いわゆるASP(Application Service Provider)型のサービスによって、上述の処理を実行する構成としてもよい。なお、本形態におけるプログラムには、電子計算機による処理の用に供する情報であってプログラムに準ずるもの(コンピュータに対する直接の指令ではないがコンピュータの処理を規定する性質を有するデータ等)を含むものとする。 A computer that executes such a program, for example, first stores the program recorded on a portable recording medium or the program transferred from the server computer once in its own storage device. Then, when executing the process, this computer reads the program stored in its own recording medium and executes the process according to the read program. Also, as another execution form of this program, the computer may read the program directly from a portable recording medium and execute processing according to the program, and the program is transferred from the server computer to this computer. Each time, the processing according to the received program may be executed sequentially. In addition, the above-mentioned processing is executed by a so-called ASP (Application Service Provider) type service, which does not transfer the program from the server computer to this computer, and realizes the processing function only by its execution instruction and result acquisition. may be It should be noted that the program in this embodiment includes information that is used for processing by a computer and that conforms to the program (data that is not a direct instruction to the computer but has the property of prescribing the processing of the computer, etc.).
 また、この形態では、コンピュータ上で所定のプログラムを実行させることにより、本装置を構成することとしたが、これらの処理内容の少なくとも一部をハードウェア的に実現することとしてもよい。 In addition, in this embodiment, the device is configured by executing a predetermined program on a computer, but at least part of these processing contents may be implemented by hardware.

Claims (6)

  1.  所定の空間内における壁面の反射率を推定する推定装置であって、
     前記所定の空間内を模擬したシミュレーション空間における所定の位置から発せられた音響信号の残響に基づく値を取得するシミュレーション取得部と、
     現実空間である前記所定の空間において、前記所定の位置に対応する位置から発せられた音響信号の残響に基づく値を取得する現実空間取得部と、
     前記シミュレーション空間における残響に基づく値と、前記現実空間における残響に基づく値と、の差を用いた評価値を利用して壁面の反射率を推定するパラメータ推定部と、
     を含む推定装置。
    An estimating device for estimating the reflectance of a wall surface in a predetermined space,
    a simulation acquisition unit that acquires a value based on reverberation of an acoustic signal emitted from a predetermined position in the simulation space that simulates the predetermined space;
    a real space acquisition unit that acquires a value based on reverberation of an acoustic signal emitted from a position corresponding to the prescribed position in the prescribed space that is the physical space;
    a parameter estimation unit that estimates the reflectance of a wall surface using an evaluation value using a difference between a value based on reverberation in the simulation space and a value based on reverberation in the real space;
    estimator including
  2.  請求項1の推定装置であって、
     前記前記シミュレーション空間における残響に基づく値と、前記現実空間における残響に基づく値と、の二乗誤差に基づく値を前記評価値として求める評価関数計算部を含む、
     推定装置。
    The estimating device of claim 1,
    an evaluation function calculation unit that obtains, as the evaluation value, a value based on a squared error between a value based on the reverberation in the simulation space and a value based on the reverberation in the real space;
    estimation device.
  3.  請求項1または請求項2の推定装置であって、
     前記パラメータ推定部は、ガウス分布によって回帰曲線を予測し、前記評価値が最小または最大となる反射率を推定する、
     推定装置。
    The estimating device according to claim 1 or claim 2,
    The parameter estimation unit predicts a regression curve using a Gaussian distribution, and estimates the reflectance at which the evaluation value is the minimum or maximum.
    estimation device.
  4.  請求項1から請求項3の何れかの推定装置であって、
     前記シミュレーション取得部は、
     前記所定の空間内を模擬したシミュレーション空間における音源位置と受聴位置のインパルス応答を推定するンパルス応答推定部と、
     インパルス応答の推定値の残響時間を計算する残響時間計算部と、を含み、
     現実空間取得部は、
     現実空間である前記所定の空間において、前記音源位置に対応する位置から発せられる音響信号を前記受聴位置に対応する位置で収音して、取得し、前記音源位置に対応する位置と前記受聴位置に対応する位置のインパルス応答を測定する実環境インパルス応答取得部と、
     インパルス応答の測定値の残響時間を計算する残響時間計算部と、を含み、
     前記シミュレーション空間における残響に基づく値は、インパルス応答の推定値の残響時間であり、
     前記現実空間における残響に基づく値は、インパルス応答の測定値の残響時間である、
     推定装置。
    The estimation device according to any one of claims 1 to 3,
    The simulation acquisition unit
    an impulse response estimator for estimating impulse responses of a sound source position and a listening position in a simulation space simulating the predetermined space;
    a reverberation time calculator that calculates the reverberation time of the impulse response estimate;
    The real space acquisition unit
    Acoustic signals emitted from a position corresponding to the sound source position are picked up at a position corresponding to the listening position in the predetermined space that is the physical space, and the position corresponding to the sound source position and the listening position are acquired. a real-environment impulse response acquisition unit that measures an impulse response at a position corresponding to
    a reverberation time calculator that calculates the reverberation time of the impulse response measurement;
    The reverberation-based value in the simulation space is the reverberation time of the impulse response estimate,
    the value based on reverberation in the real space is the reverberation time of the impulse response measurement;
    estimation device.
  5.  推定装置を用いて所定の空間内における壁面の反射率を推定する推定方法であって、
     前記推定装置が、前記所定の空間内を模擬したシミュレーション空間における所定の位置から発せられた音響信号の残響に基づく値を取得するシミュレーション取得ステップと、
     前記推定装置が、現実空間である前記所定の空間において、前記所定の位置に対応する位置から発せられた音響信号の残響に基づく値を取得する現実空間取得ステップと、
     前記推定装置が、前記シミュレーション空間における残響に基づく値と、前記現実空間における残響に基づく値と、の差を用いた評価値を利用して壁面の反射率を推定するパラメータ推定ステップと、
     を含む推定方法。
    An estimation method for estimating the reflectance of a wall surface in a predetermined space using an estimation device,
    a simulation acquisition step in which the estimation device acquires a value based on reverberation of an acoustic signal emitted from a predetermined position in a simulation space that simulates the predetermined space;
    a real space obtaining step in which the estimation device obtains a value based on reverberation of an acoustic signal emitted from a position corresponding to the predetermined position in the predetermined space that is the physical space;
    a parameter estimation step in which the estimation device estimates the reflectance of the wall surface using an evaluation value using a difference between a value based on the reverberation in the simulation space and a value based on the reverberation in the real space;
    Estimation method including .
  6.  請求項1から請求項4の何れかの推定装置としてコンピュータを機能させるためのプログラム。 A program for causing a computer to function as the estimation device according to any one of claims 1 to 4.
PCT/JP2021/028897 2021-08-04 2021-08-04 Estimation device, estimation method, and program WO2023012920A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2023539439A JPWO2023012920A1 (en) 2021-08-04 2021-08-04
PCT/JP2021/028897 WO2023012920A1 (en) 2021-08-04 2021-08-04 Estimation device, estimation method, and program

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2021/028897 WO2023012920A1 (en) 2021-08-04 2021-08-04 Estimation device, estimation method, and program

Publications (1)

Publication Number Publication Date
WO2023012920A1 true WO2023012920A1 (en) 2023-02-09

Family

ID=85154411

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/028897 WO2023012920A1 (en) 2021-08-04 2021-08-04 Estimation device, estimation method, and program

Country Status (2)

Country Link
JP (1) JPWO2023012920A1 (en)
WO (1) WO2023012920A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20050048106A (en) * 2003-11-19 2005-05-24 학교법인 한양학원 Impulse response considering the material quality of wall of rectangular parallelepiped room
EP2838084A1 (en) * 2013-08-13 2015-02-18 Thomson Licensing Method and Apparatus for determining acoustic wave propagation within a modelled 3D room
JP2017085265A (en) * 2015-10-26 2017-05-18 日本放送協会 Impulse response generation device and program
JP2020153906A (en) * 2019-03-22 2020-09-24 アルパイン株式会社 Acoustic characteristic measurement system, and acoustic characteristic measurement method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20050048106A (en) * 2003-11-19 2005-05-24 학교법인 한양학원 Impulse response considering the material quality of wall of rectangular parallelepiped room
EP2838084A1 (en) * 2013-08-13 2015-02-18 Thomson Licensing Method and Apparatus for determining acoustic wave propagation within a modelled 3D room
JP2017085265A (en) * 2015-10-26 2017-05-18 日本放送協会 Impulse response generation device and program
JP2020153906A (en) * 2019-03-22 2020-09-24 アルパイン株式会社 Acoustic characteristic measurement system, and acoustic characteristic measurement method

Also Published As

Publication number Publication date
JPWO2023012920A1 (en) 2023-02-09

Similar Documents

Publication Publication Date Title
US10366534B2 (en) Selective surface mesh regeneration for 3-dimensional renderings
US10248744B2 (en) Methods, systems, and computer readable media for acoustic classification and optimization for multi-modal rendering of real-world scenes
KR102174598B1 (en) System and method for localization for non-line of sight sound source using diffraction aware
An et al. Reflection-aware sound source localization
JP6837099B2 (en) Estimating the room impulse response for acoustic echo cancellation
CN103988226A (en) Method for estimating camera motion and for determining three-dimensional model of real environment
CN109697728A (en) Data processing method, device, system and storage medium
Gautier et al. Fpga architectures for real-time dense slam
JP7447796B2 (en) Audio signal processing device, noise suppression method
KR20180071928A (en) Method and system for updating occupancy map based on super ray
Saqib et al. Sound-based distance estimation for indoor navigation in the presence of ego noise
WO2023012920A1 (en) Estimation device, estimation method, and program
JP4877744B2 (en) Building sound simulation system
Potokar et al. HoloOcean: Realistic sonar simulation
US20230184949A1 (en) Learning-based system and method for estimating semantic maps from 2d lidar scans
US20200218941A1 (en) Methods and apparatus to simulate sensor data
WO2022153359A1 (en) Microphone position presentation method, device therefor, and program
WO2022153360A1 (en) Filter design method, device therefor, and program
JP7197003B2 (en) Depth estimation device, depth estimation method, and depth estimation program
Song et al. Inferring 3d shapes of unknown rigid objects in clutter through inverse physics reasoning
Aernouts et al. MapFuse: complete and realistic 3D modelling
WO2022003976A1 (en) Ranging error calculation device, ranging error calculation method, and recording medium
Osmanković et al. Reconstructing the 3D thermal model of indoor environment from unorganized data set acquired by 3D laser scans and thermal imaging camera
Dogaru et al. An efficient sound propagation software simulator based on cellular automata
KR20190108236A (en) System and method for sound source localization using reflection aware

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21952747

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2023539439

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE