WO2022259555A1 - Procédé de reconnaissance vocale, dispositif de reconnaissance vocale et programme de reconnaissance vocale - Google Patents

Procédé de reconnaissance vocale, dispositif de reconnaissance vocale et programme de reconnaissance vocale Download PDF

Info

Publication number
WO2022259555A1
WO2022259555A1 PCT/JP2021/022414 JP2021022414W WO2022259555A1 WO 2022259555 A1 WO2022259555 A1 WO 2022259555A1 JP 2021022414 W JP2021022414 W JP 2021022414W WO 2022259555 A1 WO2022259555 A1 WO 2022259555A1
Authority
WO
WIPO (PCT)
Prior art keywords
speech recognition
model
spike
learning
sequence
Prior art date
Application number
PCT/JP2021/022414
Other languages
English (en)
Japanese (ja)
Inventor
孝典 芦原
太一 浅見
Original Assignee
日本電信電話株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電信電話株式会社 filed Critical 日本電信電話株式会社
Priority to PCT/JP2021/022414 priority Critical patent/WO2022259555A1/fr
Publication of WO2022259555A1 publication Critical patent/WO2022259555A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/16Speech classification or search using artificial neural networks

Definitions

  • the present invention relates to a speech recognition method, a speech recognition device, and a speech recognition program.
  • the feature values are extracted after windowing the input speech waveform to a frame with a certain width, and the feature values are sequentially shifted while shifting the frames by a certain width.
  • a sequence is generated and used as an input for a speech recognition model (see Patent Document 1).
  • Non-Patent Documents 1 and 2 a technique for speeding up speech recognition by reducing the number of frames of feature amounts to be input to the speech recognition model is expected (see Non-Patent Documents 1 and 2).
  • Non-Patent Document 3 For example, in a speech recognition model with a loss function of CTC (Connectionist Temporal Classification), corresponding symbols are output in synchronization with input frames (see Non-Patent Document 3).
  • CTC Connectionist Temporal Classification
  • Non-Patent Document 3 describes a hybrid model such as DNN-HMM (Deep Neural Network-Hidden Markov Model). Also, Non-Patent Document 4 describes simultaneous learning of a speech recognition task and a speaker recognition task.
  • DNN-HMM Deep Neural Network-Hidden Markov Model
  • Bengio "DYNAMIC FRAME SKIPPING FOR FAST SPEECH RECOGNITION IN RECURRENT NEURAL NETWORK BASED ACOUSTIC MODELS", 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary AB, Canada, 2018, pp.4984-4988, doi: 10.1109/ICASSP.2018.8462615 Alex Graves, Santiago Fernandez, Faustino Gomez, and Jurgen Schmidhuber, "Connectionist Temporal Classification: Labeling Unsegmented Sequence Data with Recurrent Neural Networks", In Proceedings of the 23rd international conference on Machine learning (ICML '06). York, NY, USA, 2006, pp.369-376.
  • the present invention has been made in view of the above, and aims to speed up speech recognition by a speech recognition model using CTC.
  • a speech recognition method is a speech recognition method executed by a speech recognition device, which extracts a series of feature amounts for each frame of a speech signal. a learning step of learning a speech recognition model using CTC (Connectionist Temporal Classification) using the series of extracted feature values; and a speech recognition model spike using the learned speech recognition model and a prediction learning step of learning a spike point prediction model that predicts when the spike will be output using the generated spike sequence. characterized by containing
  • FIG. 1 is a schematic diagram illustrating a schematic configuration of a speech recognition device.
  • FIG. 2 is a flow chart showing a speech recognition processing procedure.
  • FIG. 3 is a flow chart showing a speech recognition processing procedure.
  • FIG. 4 is a diagram illustrating a computer that executes a speech recognition program.
  • the speech recognition apparatus of this embodiment focuses on the posterior probability sequence, which is the spike sequence for each frame in the speech recognition model using CTC.
  • the spike series is a series of posterior probabilities, in which the target recognition label symbol instantaneously appears as a spike with an extremely high posterior probability in a short frame of about 1 to 3 frames.
  • other frames are mostly occupied by blank symbols having no recognition label with high posterior probability. For example, for 10 frames of input speech "patent”, if the output label is two characters “patent”, the blank symbol _ is output together with "____special________".
  • the speech recognition device limits the input frames for speech recognition to frames in which spikes appear. This enables the speech recognition device to speed up speech recognition.
  • the speech recognition apparatus learns in advance a model for predicting a spike point, and uses it in the preceding stage of the speech recognition model.
  • a model for predicting this spike point is learned as a binary classification model using a teacher label that assigns 1 to the label of the spike point and 0 to the blank label of the other points.
  • the speech recognition apparatus performs high-speed speech recognition by inputting only spike point frames having a label of 1 predicted from such a binary classification model into a speech recognition model using CTC. Realize.
  • speech recognition models using CTC words and characters are used as output label symbols. In other words, the input required for the output of is extremely narrowed down.
  • the speech recognition device applies multi-task learning, in which one model learns multiple tasks at the same time and multiple tasks are expressed in one model.
  • the speech recognition apparatus integrates the binary classification model and the speech recognition model, and simultaneously solves the speech recognition task and the binary classification task.
  • multi-task learning is realized by sharing an encoder for a speech recognition task and a binary classification task, and learning a decoder for the binary classification task.
  • the speech recognition apparatus achieves an improvement in accuracy and a reduction in size of the model.
  • FIG. 1 is a schematic diagram illustrating a schematic configuration of a speech recognition device.
  • a speech recognition apparatus 10 of this embodiment is realized by a general-purpose computer such as a personal computer, and includes an input unit 11, an output unit 12, a communication control unit 13, a storage unit 14, and a control unit 15. .
  • the input unit 11 is implemented using input devices such as a keyboard and a mouse, and inputs various instruction information such as processing start to the control unit 15 in response to input operations by the practitioner.
  • the output unit 12 is implemented by a display device such as a liquid crystal display, a printing device such as a printer, an information communication device, or the like.
  • the communication control unit 13 is implemented by a NIC (Network Interface Card) or the like, and controls communication between the control unit 15 and an external device such as a server or a device that acquires an acoustic signal via a network.
  • NIC Network Interface Card
  • the storage unit 14 is implemented by semiconductor memory devices such as RAM (Random Access Memory) and flash memory, or storage devices such as hard disks and optical disks. Note that the storage unit 14 may be configured to communicate with the control unit 15 via the communication control unit 13 . In the present embodiment, the storage unit 14 stores, for example, a speech recognition model 14a, a spike point prediction model 14b, and the like used for speech recognition processing, which will be described later.
  • RAM Random Access Memory
  • flash memory or storage devices such as hard disks and optical disks.
  • the storage unit 14 may be configured to communicate with the control unit 15 via the communication control unit 13 .
  • the storage unit 14 stores, for example, a speech recognition model 14a, a spike point prediction model 14b, and the like used for speech recognition processing, which will be described later.
  • the control unit 15 is implemented using a CPU (Central Processing Unit), NP (Network Processor), FPGA (Field Programmable Gate Array), etc., and executes a processing program stored in memory. Thereby, the control unit 15 functions as an acquisition unit 15a, an extraction unit 15b, a learning unit 15c, a generation unit 15d, a prediction learning unit 15e, and a recognition unit 15f, as illustrated in FIG. Note that these functional units may be implemented in different hardware.
  • the learning unit 15c may be implemented as a learning device
  • the recognition unit 15f may be implemented as a recognition device.
  • the control unit 15 may include other functional units.
  • the acquisition unit 15a acquires an audio signal. Specifically, the acquisition unit 15a acquires an analog audio signal and performs A/D conversion to obtain an audio digital signal.
  • the acquisition unit may cause the storage unit 14 to store an audio digital signal obtained by A/D converting the acquired audio signal. Alternatively, the acquiring unit 15a may immediately transfer the audio digital signal obtained by A/D converting the acquired audio signal to the extracting unit 15b described below without storing it in the storage unit 14.
  • the extraction unit 15b extracts a series of feature amounts for each frame of the audio signal. Specifically, the extraction unit 15b extracts the acoustic feature quantity from the speech digital signal, and obtains a feature quantity sequence for each utterance. The extraction unit 15b extracts, for example, a feature amount using MFCC or power based on short-time frame analysis of an audio signal. For example, the extracting unit 15b extracts 1 to 12 dimensions of MFCC (Mel-Frequency Cepstrum Coefficients), dynamic parameters such as ⁇ MFCC, ⁇ MFCC, power, ⁇ power, ⁇ power, etc., as feature amounts. use.
  • MFCC Fel-Frequency Cepstrum Coefficients
  • the extraction unit 15b may perform Cepstral Mean Normalization (CMN) processing on the MFCC.
  • CNN Cepstral Mean Normalization
  • the feature amount is not limited to those using MFCC or power, and may be parameters such as autocorrelation peak value and group delay, which are used for identifying special utterances.
  • the extraction unit 15b may store the extracted feature series in the storage unit 14. Alternatively, the extracting unit 15b may immediately transfer the extracted feature series to the learning unit 15c described below without storing them in the storage unit 14. FIG.
  • the learning unit 15c learns the speech recognition model 14a using CTC using the series of extracted feature amounts. Specifically, the learning unit 15c learns the end-to-end speech recognition model 14a using the feature amount series of the teacher data.
  • This speech recognition model 14a uses a CTC loss function and outputs a sequence of labels that is a spike sequence.
  • the generation unit 15d uses the learned speech recognition model 14a to generate a spike sequence, which is a sequence of labels output by the speech recognition model 14a as spikes. Specifically, the generating unit 15d uses the trained speech recognition model 14a learned by the learning unit 15c to cause the speech recognition model 14a to speech-recognize the feature quantity series used for learning, thereby generating the Generate a spike sequence that is a stochastic sequence.
  • the generation unit 15d may store the generated spike series in the storage unit 14. Alternatively, the generating unit 15d may immediately transfer the generated spike sequence to the prediction learning unit 15e described below without storing it in the storage unit 14. FIG.
  • the prediction learning unit 15e uses the generated spike series to learn the spike point prediction model 14b that predicts when the spike will be output.
  • the spike point prediction model 14b is a binary classification model.
  • the prediction learning unit 15e learns the spike point prediction model 14b using teacher data labeled as 1 for spike points in the spike series and 0 for other non-spike points.
  • the prediction learning unit 15e uses a predetermined threshold as a parameter, and sets 1 when it is equal to or greater than the threshold and 0 when it is less than the threshold. The larger the threshold, the fewer spike points are predicted, and the smaller the threshold, the more spike points are predicted.
  • the prediction learning unit 15e learns a multitask learning model that integrates the spike point prediction model 14b and the speech recognition model 14a using the generated spike series and the extracted feature quantity series.
  • the multitask learning model is, but not limited to, an encoder-decoder model, for example.
  • the prediction learning unit 15e realizes multitask learning by sharing the encoder for the speech recognition task and the spike point prediction task and learning the decoder for the spike point prediction task. do.
  • the speech recognition apparatus 10 can improve the accuracy and reduce the size of the model.
  • the output of the spike point prediction model 14b which is a binary classification model, is the prediction result at time t+1 for the input time t. If the prediction result is 1, run the decoder of the speech recognition model with the input at time t+1.
  • the prediction learning unit 15e predicts whether or not there will be a spike point at time t+1 for the input at time t. Therefore, the target label at the time (+1) next to the input is used for learning of the speech recognition model 14a.
  • the method is not limited to this method. For example, if the output label is 5 for the input time t, the next spike point appears at time t+5. You can predict. In this case, the prediction learning unit 15e inputs the feature amount of the frame at time t+5 to the speech recognition model 14a and the spike point prediction model 14b without using the intermediate frame as an input to the model, and inputs it to the spike point prediction model 14b so that the next spike Used for point prediction.
  • the recognition unit 15f performs speech recognition by inputting into the learned speech recognition model 14a the feature quantity of the speech signal at the spike points predicted using the spike point prediction model 14b.
  • the recognition unit 15f uses the multitask learning model learned by the prediction learning unit 15e to simultaneously perform spike point prediction and speech recognition.
  • the recognition unit 15f inputs, to the speech recognition model 14a, the feature quantity of the speech signal frame at the predicted time of the next spike point with respect to the time at which the speech signal frame was input, as described above. to perform speech recognition.
  • the speech recognition apparatus 10 can speed up speech recognition by performing speech recognition using the speech recognition model 14a limited to the feature quantity of the spark point.
  • FIG. 2 shows the learning processing procedure.
  • the flowchart of FIG. 2 is started, for example, when an input instructing the start of learning processing is received.
  • the acquisition unit 15a acquires an audio signal (step S1). Further, the extracting unit 15b extracts a series of feature amounts for each frame of the audio signal (step S2). Specifically, the extraction unit 15b extracts the acoustic feature quantity from the speech digital signal, and obtains a feature quantity sequence for each utterance.
  • the learning unit 15c learns the end-to-end speech recognition model 14a using CTC using the series of extracted feature amounts (step S3).
  • the generation unit 15d uses the learned speech recognition model 14a to generate a spike sequence, which is a posterior probability sequence of labels output as spikes by the speech recognition model 14a (step S4).
  • the prediction learning unit 15e uses the generated spike series to learn the spike point prediction model 14b that predicts when the spike will be output (step S5).
  • the prediction learning unit 15e learns a multitask learning model that integrates the spike point prediction model 14b and the speech recognition model 14a using the generated spike series and the extracted feature quantity series. This completes a series of learning processes.
  • FIG. 3 shows the recognition processing procedure.
  • the flowchart of FIG. 3 is started, for example, at the timing when an input instructing the start of recognition processing is received.
  • the acquisition unit 15a acquires a speech signal to be speech-recognized (step S11).
  • the recognition unit 15f performs speech recognition by inputting the feature amount of the speech signal at the spike points predicted using the spike point prediction model 14b to the learned speech recognition model 14a (step S12).
  • the recognition unit 15f uses the multitask learning model learned by the prediction learning unit 15e to simultaneously perform spike point prediction and speech recognition. This completes a series of recognition processes.
  • the extracting unit 15b extracts a series of feature quantities for each frame of the speech signal.
  • the learning unit 15c learns the speech recognition model 14a using CTC using the series of extracted feature amounts.
  • the generation unit 15d uses the learned speech recognition model 14a to generate a spike sequence, which is a sequence of labels output as spikes by the speech recognition model.
  • the prediction learning unit 15e uses the generated spike series to learn the spike point prediction model 14b that predicts the point in time when the spike will be output.
  • the spike point prediction model 14b is a binary classification model.
  • the speech recognition apparatus 10 predicts frames in which sparks of speech signals appear, and limits input frames for speech recognition to frames in which spikes appear. As a result, the speech recognition apparatus 10 can speed up speech recognition using a speech recognition model using CTC.
  • the prediction learning unit 15e learns a multitask learning model that integrates the spike point prediction model 14b and the speech recognition model 14a using the generated spike series and the extracted feature quantity series.
  • a multi-task learning model is an encoder-decoder model.
  • the recognition unit 15f performs speech recognition by inputting the feature amount of the speech signal at the spike points predicted using the spike point prediction model 14b to the learned speech recognition model 14a. This makes it possible to suppress the processing load and perform speech recognition at high speed.
  • the speech recognition apparatus 10 can be implemented by installing a speech recognition program for executing the above speech recognition processing as package software or online software on a desired computer.
  • the information processing apparatus can function as the speech recognition apparatus 10 by causing the information processing apparatus to execute the above speech recognition program.
  • information processing devices include mobile communication terminals such as smartphones, mobile phones and PHS (Personal Handyphone Systems), and slate terminals such as PDAs (Personal Digital Assistants).
  • the functions of the speech recognition device 10 may be implemented in a cloud server.
  • FIG. 4 is a diagram showing an example of a computer that executes a speech recognition program.
  • Computer 1000 includes, for example, memory 1010 , CPU 1020 , hard disk drive interface 1030 , disk drive interface 1040 , serial port interface 1050 , video adapter 1060 and network interface 1070 . These units are connected by a bus 1080 .
  • the memory 1010 includes a ROM (Read Only Memory) 1011 and a RAM 1012 .
  • the ROM 1011 stores a boot program such as BIOS (Basic Input Output System).
  • BIOS Basic Input Output System
  • Hard disk drive interface 1030 is connected to hard disk drive 1031 .
  • Disk drive interface 1040 is connected to disk drive 1041 .
  • a removable storage medium such as a magnetic disk or an optical disk is inserted into the disk drive 1041, for example.
  • a mouse 1051 and a keyboard 1052 are connected to the serial port interface 1050, for example.
  • a display 1061 is connected to the video adapter 1060 .
  • the hard disk drive 1031 stores an OS 1091, application programs 1092, program modules 1093 and program data 1094, for example. Each piece of information described in the above embodiment is stored in the hard disk drive 1031 or the memory 1010, for example.
  • the speech recognition program is stored in the hard disk drive 1031 as a program module 1093 in which commands to be executed by the computer 1000 are described, for example.
  • the hard disk drive 1031 stores a program module 1093 that describes each process executed by the speech recognition apparatus 10 described in the above embodiment.
  • Data used for information processing by the voice recognition program is stored as program data 1094 in the hard disk drive 1031, for example. Then, the CPU 1020 reads out the program module 1093 and the program data 1094 stored in the hard disk drive 1031 to the RAM 1012 as necessary, and executes each procedure described above.
  • program module 1093 and the program data 1094 related to the speech recognition program are not limited to being stored in the hard disk drive 1031, but may be stored in a removable storage medium and read by the CPU 1020 via the disk drive 1041 or the like. may be issued.
  • the program module 1093 and program data 1094 related to the speech recognition program are stored in another computer connected via a network such as LAN (Local Area Network) or WAN (Wide Area Network), and are stored via network interface 1070. may be read by the CPU 1020 at the same time.
  • LAN Local Area Network
  • WAN Wide Area Network

Landscapes

  • Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

Unité d'extraction (15b) extrayant une séquence de caractéristiques pour chaque trame de signal vocal. Une unité d'apprentissage (15c) forme un modèle de reconnaissance vocale (14a) qui utilise une CTC, à l'aide de la séquence de caractéristiques extraite. Une unité de génération (15d) génère une séquence de pointes, qui est une séquence d'étiquettes délivrées par le modèle de reconnaissance vocale en tant que pointes, à l'aide du modèle de reconnaissance vocale (14a) formé. Une unité d'apprentissage par prédiction (15e) forme un modèle de prédiction de moment de pointe (14b) pour prédire le moment auquel ladite pointe sera délivrée en utilisant la séquence de pointes générée.
PCT/JP2021/022414 2021-06-11 2021-06-11 Procédé de reconnaissance vocale, dispositif de reconnaissance vocale et programme de reconnaissance vocale WO2022259555A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2021/022414 WO2022259555A1 (fr) 2021-06-11 2021-06-11 Procédé de reconnaissance vocale, dispositif de reconnaissance vocale et programme de reconnaissance vocale

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2021/022414 WO2022259555A1 (fr) 2021-06-11 2021-06-11 Procédé de reconnaissance vocale, dispositif de reconnaissance vocale et programme de reconnaissance vocale

Publications (1)

Publication Number Publication Date
WO2022259555A1 true WO2022259555A1 (fr) 2022-12-15

Family

ID=84424561

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/022414 WO2022259555A1 (fr) 2021-06-11 2021-06-11 Procédé de reconnaissance vocale, dispositif de reconnaissance vocale et programme de reconnaissance vocale

Country Status (1)

Country Link
WO (1) WO2022259555A1 (fr)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2020112787A (ja) * 2019-01-08 2020-07-27 バイドゥ オンライン ネットワーク テクノロジー (ベイジン) カンパニー リミテッド 切断アテンションに基づくリアルタイム音声認識方法、装置、機器及びコンピュータ読み取り可能な記憶媒体
JP2021039219A (ja) * 2019-09-02 2021-03-11 日本電信電話株式会社 音声信号処理装置、音声信号処理方法、音声信号処理プログラム、学習装置、学習方法及び学習プログラム

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2020112787A (ja) * 2019-01-08 2020-07-27 バイドゥ オンライン ネットワーク テクノロジー (ベイジン) カンパニー リミテッド 切断アテンションに基づくリアルタイム音声認識方法、装置、機器及びコンピュータ読み取り可能な記憶媒体
JP2021039219A (ja) * 2019-09-02 2021-03-11 日本電信電話株式会社 音声信号処理装置、音声信号処理方法、音声信号処理プログラム、学習装置、学習方法及び学習プログラム

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
HIROFUMI INAKUMA; MASATO MIMURA; TATSUYA KAWAHARA: "Speech recognition by streaming attention mechanism type sequence-to-sequence model", IPSJ SIG TECHNICAL REPORT, vol. 2020-SLP-131, no. 9, 6 February 2020 (2020-02-06), JP, pages 1 - 7, XP009535113 *

Similar Documents

Publication Publication Date Title
CN111613212B (zh) 语音识别方法、系统、电子设备和存储介质
US10672380B2 (en) Dynamic enrollment of user-defined wake-up key-phrase for speech enabled computer system
US20180357998A1 (en) Wake-on-voice keyword detection with integrated language identification
CN113327609B (zh) 用于语音识别的方法和装置
TW201503106A (zh) 聲學模型訓練方法和裝置
KR20110038474A (ko) 문장경계 인식 장치 및 방법
EP2941768A1 (fr) Procédés et systèmes pour fournir des systèmes de reconnaissance vocale sur la base de journaux d'enregistrements de voix
JPH09127978A (ja) 音声認識方法及び装置及びコンピュータ制御装置
CN110648691A (zh) 基于语音的能量值的情绪识别方法、装置和系统
CN112151003A (zh) 并行语音合成方法、装置、设备以及计算机可读存储介质
CN114495904B (zh) 语音识别方法以及装置
CN112331207A (zh) 服务内容监控方法、装置、电子设备和存储介质
EP4367663A1 (fr) Amélioration de la reconnaissance de la parole à l'aide d'une adaptation de modèle basée sur la synthèse de la parole
US11587567B2 (en) User utterance generation for counterfactual analysis and improved conversation flow
US20230410794A1 (en) Audio recognition method, method of training audio recognition model, and electronic device
CN116601648A (zh) 备选软标签生成
WO2022259555A1 (fr) Procédé de reconnaissance vocale, dispositif de reconnaissance vocale et programme de reconnaissance vocale
CN116580698A (zh) 基于人工智能的语音合成方法、装置、计算机设备及介质
WO2023137175A1 (fr) Système et procédé de génération d'informations de résumé
CN115171724A (zh) 一种语速分析方法及系统
CN113516964A (zh) 语音合成方法、可读存储介质及计算机程序产品
CN113763939B (zh) 基于端到端模型的混合语音识别系统及方法
US20220310061A1 (en) Regularizing Word Segmentation
JP7505584B2 (ja) 話者ダイアライゼーション方法、話者ダイアライゼーション装置および話者ダイアライゼーションプログラム
US20240185839A1 (en) Modular Training for Flexible Attention Based End-to-End ASR

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21945224

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21945224

Country of ref document: EP

Kind code of ref document: A1