WO2023055938A2 - Contrôleur de stabilisation basé sur l'apprentissage automatique qui peut apprendre sur un système instable - Google Patents

Contrôleur de stabilisation basé sur l'apprentissage automatique qui peut apprendre sur un système instable Download PDF

Info

Publication number
WO2023055938A2
WO2023055938A2 PCT/US2022/045236 US2022045236W WO2023055938A2 WO 2023055938 A2 WO2023055938 A2 WO 2023055938A2 US 2022045236 W US2022045236 W US 2022045236W WO 2023055938 A2 WO2023055938 A2 WO 2023055938A2
Authority
WO
WIPO (PCT)
Prior art keywords
neural network
controller
drift
control variables
error
Prior art date
Application number
PCT/US2022/045236
Other languages
English (en)
Other versions
WO2023055938A3 (fr
Inventor
Dan Wang
Qiang DU
Russell Wilcox
Tong Zhou
Christos BAKALIS
Derun Li
Original Assignee
The Regents Of The University Of California
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by The Regents Of The University Of California filed Critical The Regents Of The University Of California
Publication of WO2023055938A2 publication Critical patent/WO2023055938A2/fr
Publication of WO2023055938A3 publication Critical patent/WO2023055938A3/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S3/00Lasers, i.e. devices using stimulated emission of electromagnetic radiation in the infrared, visible or ultraviolet wave range
    • H01S3/05Construction or shape of optical resonators; Accommodation of active medium therein; Shape of active medium
    • H01S3/06Construction or shape of active medium
    • H01S3/063Waveguide lasers, i.e. whereby the dimensions of the waveguide are of the order of the light wavelength
    • H01S3/067Fibre lasers
    • H01S3/06754Fibre amplifiers
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S3/00Lasers, i.e. devices using stimulated emission of electromagnetic radiation in the infrared, visible or ultraviolet wave range
    • H01S3/10Controlling the intensity, frequency, phase, polarisation or direction of the emitted radiation, e.g. switching, gating, modulating or demodulating
    • H01S3/13Stabilisation of laser output parameters, e.g. frequency or amplitude
    • H01S3/1305Feedback control systems
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S3/00Lasers, i.e. devices using stimulated emission of electromagnetic radiation in the infrared, visible or ultraviolet wave range
    • H01S3/10Controlling the intensity, frequency, phase, polarisation or direction of the emitted radiation, e.g. switching, gating, modulating or demodulating
    • H01S3/13Stabilisation of laser output parameters, e.g. frequency or amplitude
    • H01S3/1307Stabilisation of the phase
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01SDEVICES USING THE PROCESS OF LIGHT AMPLIFICATION BY STIMULATED EMISSION OF RADIATION [LASER] TO AMPLIFY OR GENERATE LIGHT; DEVICES USING STIMULATED EMISSION OF ELECTROMAGNETIC RADIATION IN WAVE RANGES OTHER THAN OPTICAL
    • H01S3/00Lasers, i.e. devices using stimulated emission of electromagnetic radiation in the infrared, visible or ultraviolet wave range
    • H01S3/23Arrangements of two or more lasers not provided for in groups H01S3/02 - H01S3/22, e.g. tandem arrangements of separate active media
    • H01S3/2383Parallel arrangements

Definitions

  • the technology of this disclosure pertains generally to coherent beam combining, and more particularly to implementing machine learning for coherent beam combining in an unstable system. [0006] 2. Background Discussion
  • Machine Learning is a powerful tool, with neural networks that can map the complex I non-linearity function from measurement to a system error array, and then use feedback to quickly correct the system input.
  • Training a machine learning model requires a dataset consisting of the measured observation data and the corresponding sets of controller actions that have an influence on the system.
  • the accuracy of the training dataset is critical for the precision of the ML model. To obtain accurate training samples, it requires the system to be at stable and reproducible states during training, which is impractical in real experiments due to system drift.
  • model-free controllers require a random dithering and searching process to map the multi-dimensional parameter space of the system under control, such as using stochastic parallel gradient descent (SPGD), which randomly dithers all the inputs and searches for an optimum set of values. Since the controller doesn’t remember its experiences, the dithering process has to be continuous, causing additional perturbations to the system, and is very inefficient, and difficult to scale.
  • SPGD stochastic parallel gradient descent
  • Machine learning also treats the system as a black box, it learns to map the function from observation to action in a deterministic way. Thus, it is able to provide rapid feedback and better performance in regard to stability.
  • the previous ML method trained with absolute values of input and output, requires training on a stable system, which is impractical in real experiments due to system drift; while it also requires periodic re-training.
  • This disclosure describes a machine learning controller that can learn to stabilize systems based on measurements in an unstable system. This allows for training on a system not yet controlled and for continuous learning as the stabilizer operates.
  • the disclosed approach is based on obtaining two measurements close together in time, separated by a known differential interval at the input.
  • the controller has improved performance on unstable systems compared to similar technologies, especially complex ones with many inputs and outputs. Furthermore, there is no need for modelling the physics, and the controller can adapt to un-analyzed or partially analyzed systems.
  • the machine learning (ML) controller learns differentially. For example, a known action is input, and the result are seen before and after (which is a multi-state in observation space), and this is performed quickly compared with parameter drifts. All of the information is present.
  • the trained ML model is capable of building a map between the differential observation space and the controller action space.
  • the system of the present disclosure feeds the trained neural network a current measurement (possibly unseen in the training dataset), together with a desired pattern in the observation space, and the neural network predicts the action needed to move the system between the two states in a deterministic way.
  • the system has numerous benefits and characteristics, briefly summarized below.
  • the system is robust against drift during training.
  • the system is capable of continuous learning while operating.
  • the system automatically updates as the system changes so that there is no need to retrain. There is no need to otherwise stabilize the system while training.
  • differential states training enables system identification on a free-drifting many-in-many-out system, without a knowledge of a mathematical model.
  • the neural network can be continuously retrained using the locking data, to capture slow transfer function change of the system, without introducing additional exploration actions or dithering.
  • the neural network prediction is deterministic and fast. Inference speed of tens of nanoseconds can be achieved on an FPGA device for coherent beam combining.
  • FIG. 1 is a system flow diagram of an embodiment of an NN-based interference pattern recognition algorithm, trained with multi-state dither, in a stabilized control loop robust against phase drift, according to at least one embodiment of the present disclosure.
  • FIG. 2A is a plot of RMS error during training of the NN, according to testing of at least one embodiment of the present disclosure.
  • FIG. 2B is a plot of feedback for 50 cases using random sampling, with a drift rate of 6 degrees, according to testing of at least one embodiment of the present disclosure.
  • FIG. 2C is a plot of feedback for 50 cases using random sampling, with a drift rate of 8 degrees, according to testing of at least one embodiment of the present disclosure.
  • FIG. 2D is a plot of feedback for 50 cases using random sampling, with a drift rate of 14 degrees, according to testing of at least one embodiment of the present disclosure.
  • FIG. 3 is a block diagram showing an experimental optical setup for stabilized beam combination, according to at least one testing embodiment of the present disclosure.
  • FIG. 4A are plots of time history of output beam powers and DAC outputs before and after activating CALIPR, according to testing of at least one embodiment of the present disclosure.
  • FIG. 4B are output diffraction patterns, according to testing of at least one embodiment of the present disclosure.
  • FIG. 4C are plots of out-of-loop measurements of combined and side beam power, according to testing of at least one embodiment of the present disclosure.
  • FIG. 5A is a training diagram for a neural network that stabilizes drift in a beam combiner, according to at least one embodiment of the present disclosure.
  • FIG. 5B is an feedback operation diagram for drift stabilization using the neural network trained according to FIG. 5A.
  • FIG. 6A is a flow diagram showing a neural network being trained with two patterns and a fast phase dither in a coherent beam combiner according to an embodiment of the presented technology.
  • FIG. 6B is a flow diagram showing feedback by correcting a predicted phase error using the neural network trained according to FIG. 6A.
  • FIG. 7 is a flow diagram showing am embodiment of a generalized neural network feedback loop for stabilizing drift in a controlled system according to the presented technology.
  • Coherent beam combining is a promising technique for laser power scaling, and is key to a broad range of applications.
  • Laser energy can be combined in different ways, including temporal pulse stacking and spatial beam addition using schemes such as binary tree, tiled aperture and filled aperture diffractive combining.
  • schemes such as binary tree, tiled aperture and filled aperture diffractive combining.
  • there are challenges to identify errors and build a deterministic error detector due to the large number of degrees of freedom, incomplete diagnostic information, non-linearity, measurement noise, and most importantly, the spontaneous system phase state drift.
  • CALIPR reads interference patterns at the combiner output, derives a phase error array, and feeds back to quickly correct the input beams or pulses.
  • CALIPR also employs a unique multi-state training scheme to capture the system characteristic into a neural network, and to detect errors for correction from measured interference pattern recognition in a deterministic way.
  • FIG. 1 is a schematic illustrating an embodiment 100 of a neural network (NN) based method of stabilizing a system against drift shown in the context of a diffractive laser combining system.
  • NN neural network
  • the NN 102 comprises an interference pattern recognition algorithm that is trained with multi-state dither 104 (e.g., at least a double-state pair), and positioned in a stabilized control loop robust against phase drift.
  • multi-state dither 104 e.g., at least a double-state pair
  • a controller 106 uses control variables based on error recognition algorithms (mapping function between measurement and system errors) to correct system errors in order to stabilize the output of a plant/system 108.
  • the system output is measured using a diagnostic sensor (e.g., camera) 110 to provide an input stream to the NN.
  • a diagnostic sensor e.g., camera
  • an 8- beam individual input laser beams (in a 3 x 3 array) 112 are combined by a diffractive optics element (combiner 114, 116), into 5 x 5 (25-beam) interference diffractive patterns 118.
  • the optimal pattern has a very bright power 120 in the center that combines most of the power from all the individual input beams.
  • the center beam 120 in the diffractive pattern is the brightest, with maximum power and maximum combining efficiency. Otherwise, if the input (before the combiner) is changed, the diffractive pattern is off optimal (not optimal), with lower center beam power and extra power loss into the side beams.
  • the notation d(t) 122 indicates different drift I noise I perturbation from the environment that can change the diffractive pattern. It should be noted that d(t) is unknown and uncontrollable.
  • the diagnostic sensor 110 measures the real-time diffractive pattern y(t) 124, which contains the information of system error, but it is incomplete and non-unique due to the loss of phase information.
  • the phase controller 106 corrects the input laser beam phase with control signal(s) u(t) 126, to compensate for the system errors based on the measured intensity pattern y(t) from the diagnostic sensor.
  • the controller is implemented with a NN-based interference pattern recognition algorithm 102 that is trained with a multi-state dither scheme 104 such as a double-state pair.
  • CALIPR is model-free, calibration-free, and can be implemented on various types of processors and process-controlled devices, including a standard computer, or can be implemented on a GPU/FPGA platform for fast real time performance.
  • the technology provides a method that enables a simple neural network that can be ported to a FPGA device for real time performance.
  • a neural network can be trained to recognize diffraction intensity patterns and translate them into an absolute beam phase error array, despite non-unique interference patterns in 2TT phase, nonlinearity and noisy measurements.
  • the training range only needs to be a small fraction of the full multidimensional phase space.
  • absolute phase recognition requires the combiner to be stable and have reproducible phase states during training, which is impractical in real experiments due to phase drift.
  • the problem of deceptive absolute phase recognition due to phase drift can be solved by CALIPR using a multi-state training scheme.
  • the multi-state training can be a double-state, or more states, used for training.
  • the neural network is trained using a large number of samples, each comprising a set of two or more interference patterns and a random phase dither. If the random dither is faster than phase drift, the trained neural network is capable of building a map between the differential phase space and a double-frame diffraction pattern space. In feedback, if we input the measured pattern and the target diffraction pattern with the highest combining efficiency, the neural network can predict the phase error in between, within an accuracy depending on the training parameters.
  • Samples used to train the NN are required to be within a limited phase range ([-TT/2, TT/2]) around the optimal state, to avoid the ambiguity of pattern recognition in the full 2TT range. Randomly chosen samples can be used when drift is slow (drift rate OAd ⁇ 6 degrees) as we start from the optimal state. It fails for fast drift rates (OAd > 8 degrees), as even for a small number of samples the system drift would bring phase over the IT range.
  • the present disclosure uses the selected sampling method which only uses the patterns near optimal as samples to train the NN and get rid of patterns with low combining efficiency to avoid ambiguity.
  • FIG. 2A and FIG. 2B show training of the NN and CALIPR performance on feedback at different drift rates with random sampling or selected sampling methods, in FIG. 2A the RMS error between the predicted phase and the known phase drops as the NN is trained. It is expected to be large as shown in FIG. 2A, because our known phase neglects system drift, while the measurement includes system drift.
  • a diffractive beam combiner includes side beams (interference patterns) which contain information giving the phase of the input beams, and these can be analyzed to yield phase error correction signals. Otherwise, these interference patterns can be “learned” by a machine without actual analysis, which is useful when the patterns are complex.
  • a CALIPR implemented machine is illustrated with many examples of mistuned beams and the consequent efficiency loss.
  • the machine learns to recognize phase errors from their side beam signature, providing for single-step correction.
  • Training the machine to recognize patterns and corresponding input phase states implies the patterns are stable enough to be measured and correlated with controlled phases. This may not be the case in a realistic fiber laser system where thermal drifts are causing phase to continuously vary. It is possible to stabilize the phases using a noisier, less robust algorithm, such as Stochastic Parallel Gradient Descent (SPGD), while the machine learns. Alternatively, one can make the training process robust against drift. This can be done, for example, by training using the difference between two observations as the input phases are changed in a controlled way, basically finding a difference rather than an absolute value. This two-state dither scheme allows the phases to drift during training, because the two measurements are acquired in a short time interval compared with the drift rate.
  • SPGD Stochastic Parallel Gradient Descent
  • the dither intervals are randomly chosen from a set of orthogonal vectors in 8-dimensional space.
  • FIG. 3 illustrates an example optical setup 200 using the stabilized beam combination.
  • the configuration shown in FIG. 3 is designed to equalize delays and combine an array of high-energy fiber amplifiers seeded by signals successively picked off from a central beam.
  • a plurality of piezo-actuated mirrors (piezo mirror) 202 were used for phase control, using a single-frequency CW laser as a source (input beam 204), split eight ways using a separate diffractive optic.
  • An array of mirrors (motorized mirrors 206) forms the eight beams into a 3x3 matrix with no center beam, which is then incident on the dispersion compensating diffractive optic (dispersion compensator 208).
  • a fraction of the output power is passed through a lens 212 to a camera 214 that images the far-field spots, showing interference patterns associated with the input beam array phase state.
  • This information is processed by the neural network 216 to yield error signals applied to the high voltage amplifiers (HV amplifiers 218), which then drive the piezo mirrors 202 to shift optical phases and stabilize the system against variable perturbations.
  • Optical signals from the piezo mirrors can be amplified if needed using an amplifier array 220.
  • the test measured stability and combining efficiency 224 when the control loop is closed.
  • the test measures the intensity of central combined beam power with a power meter 226 after a mirror with a hole 228, which is independent of the control loop.
  • the combining efficiency defined as the ratio of the central combined beam power to the total power from all twenty-five diffracted beams, is measured with another power meter 230 after the concave mirror 232 and mirror with hole 228 (output through the hole directed to power meter 226 coupled to 224).
  • FIG. 4A through FIG. 4C illustrate different measurements taken in the example system of FIG. 3.
  • Training takes approximately 8,000 measurements over a period of 3.3 minutes, resulting in an RMS prediction error of about 11 degrees.
  • the trained neural network is presented with a measured diffraction pattern and the best pattern it has seen in the training dataset (the target), so that the predicted phase error array can be used for feedback and maintain the system phase state against perturbations, as shown in FIG. 4A.
  • FIG. 4A In the upper plot of FIG. 4A is seen a plot of pattern beam power with respect to time. Before locking occurs, the free running beams are seen, then after locking a saturated combined beam is shown with 24 side beams at the bottom of this plot.
  • the lower plot of FIG. 4A depicts the 8 DAC outputs with respect to time, and showing the locking takes place in less than 20 steps.
  • FIG. 4B In FIG. 4B is shown interference patterns before (free-running) and after (stabilized pattern) the beams are combined.
  • FIG. 4C In FIG. 4C is seen out-of-loop power meter measurements (in microwatts) of the combined and total side beam powers, where stability of the combined beam is RMS 0.4% after locking. For comparison, when implementing an SPGD algorithm the stability was observed to be 8.8% RMS. In addition to adding noise from random dither, the SPGD recovers slowly from perturbations. The pattern recognition scheme of the present disclosure recovers in one cycle, and this is true for any number of beams, whereas the latency of SPGD scales with the number of beams. The measured combining efficiency (center beam power divided by side beams power) is 74%. This is lower than our previous result of 86%, possibly due to a bad beam.
  • the retrained NN In the testing we purposely blocked one beam, and the retrained NN is still capable of locking seven beams with similar performance. Additionally, in testing the present disclosure we quantized the NN and implemented it on an FPGA which has a direct camera interface, and demonstrated a frame rate of 800 frames per second.
  • Deterministic stabilization in a beam combiner can be achieved by pattern recognition, after characterizing the transmission function of the combiner optic, where the pattern recognition process used for system control recognizes the control variables of input beam phases based on the measured intensity patterns.
  • a machine learning controller is an effective solution and mapping information can be learned from experimental data.
  • the NN Since drift/disturbances exist in most systems, the NN needs to be trained on a drifting system in order to be effective. If the system could be completely trained before any significant drift, that would work, but with the current sample rates, this is not possible. Several thousand samples are required and, with a sample time of 1 kHz for example, this would require several seconds during which time the drift will be unacceptable. Furthermore, parameters that are not controlled by the phase actuator (such as relative beam power) change during long-term operation and cause the phase/pattern correlations to change.
  • phase actuator such as relative beam power
  • this solution trains the NN to correlate pattern differences with phase differences so that the NN can be trained on a drifting system, and then retrain the NN during operation in order to track changes.
  • the NN becomes a device that learns which differences in interference patterns are correlated with which vectors in phase space, so that given an ideal pattern, it can find the error vector for feedback.
  • FIG. 5A and FIG. 5B illustrate an embodiment of DDRM applied to a coherent laser combining system.
  • FIG. 5A shows training 300 and FIG. 5B shows feedback 400.
  • DDRM captures the system characteristic into a neural network, and detects errors for correction from measured interference pattern recognition in a deterministic way.
  • DDRM maps between a multi-state diffraction pattern space and the corresponding differential phase space.
  • the NN becomes a device that learns which differences in interference patterns are correlated with which vectors in phase space, so that given an ideal pattern, it can find the error vector for feedback.
  • the neural network is trained with two patterns and corresponding phase dither therebetween. While two patterns and dither are shown as an example (e.g., double-state dither), the approach is not limited to two states and should be considered a multi-state approach.
  • the beam combiner 302 serves as a mapping function from laser phase to intensity pattern. Intensity pattern A (304) and intensity pattern B (306) can be measured with a sensor such as a camera and the measurements serve as input 1 (308) and input 2 (310) to the neural network (NN) pattern recognizer 312.
  • the neural network pattern recognizer 312 used for control is the reverse of a combiner, and performs a mapping function from intensity patterns to the phase space.
  • the measured intensity patterns 304, 306 (pattern A and pattern B, respectively) on the right side of FIG. 5A are associated with the phase states A (314), A (316) and B (318) shown in circles on the left side of FIG. 5A.
  • Phase states 314 and 318 are measured states 320
  • phase state 316 is phase state 314 plus noise 322.
  • the absolute beam phase state is unknown due to drift and noise.
  • the only known quantity in the phase space is the active phase dither 324 that is injected between pattern A and pattern B. Accordingly, differential space is used rather than absolute values.
  • phase dither 324 is injected and intensity patterns are measured before and after injection (pattern A and pattern B). Then correlated data samples of [pattern A, phase dither, pattern B] are used to train the NN pattern recognizer 312. The trained NN pattern recognizer then becomes a device that learns which differences in interference patterns are correlated with which vectors in phase space, so that given an ideal pattern, it can find the error vector for feedback.
  • FIG. 5B illustrates feedback 400 for correcting a predicted phase error between a current (measured) pattern and a target pattern.
  • the trained NN becomes a device 312' that learns which differences in interference patterns are correlated with which vectors exist in phase space. For example, with a given measured interference pattern C (402), together with a target pattern D (404) as input 1 (406) and input 2 (408), respectively, the NN pattern recognizer 312 can find the error vector, i.e., phase correction 410, for feedback correction.
  • phase state diagram on the left side of FIG. 5B shows the phase states A (314) and B (318) as the measured states 320
  • Pattern C (412) is a new state 414 for the pattern recognizer 312', which is typically unseen in the training dataset.
  • Pattern C (416) is pattern C (412) plus noise 418.
  • the neural network acts like a multidimensional interpolation to predict the distance vector (correction 410) from pattern C (416) to the near target state D' (420) from its experience of 314, 318.
  • FIG. 6A is a flow diagram for an embodiment 500 of a process for obtaining a training dataset 502 with conventional Stochastic parallel gradient descent (SPGD) algorithms 504 for a phase estimation/error recognition in the feedback loop for stabilizing against drift 506 in a coherent beam combiner 508.
  • SPGD Stochastic parallel gradient descent
  • This flow diagram can serve as the basis for creating code for executable instructions for processor-based implementations.
  • ( j) is the horizontal and vertical coordinate of both the input beam array ), and the far-field diffracted beam array from the incident direction, with the zero-order beam located at (0,0).
  • the output pattern is a 5 x 5 array.
  • the diffractive combining system serves as a mapping function f 518, from phase space to the intensity pattern space 516.
  • f 518 mapping function
  • the SPGD algorithm optimizes and stabilizes the system by dithering and searching.
  • a phase dithering routine 526 generates a random dither ⁇ p ( ) in time series] 528, a gradient is calculated based on the intensity of the center combined beam in the pattern after dithers, is calculated as the error signal, and the error signal is compared with the phase setpoint 0 (534) and sent to the controller 536.
  • PID proportional-integral-derivative
  • the SPGD loop stabilizes the phase state near-optimal.
  • Either the SPGD phase dithering action and pattern pairs, or the controller correction action and pattern pairs, can be used.
  • the training samples can be from the recorded observation and action pair of any existing controller that can roughly maintain the optimal combining state.
  • a controller can be the popular SPGD process.
  • the controller can also be a neural network-based controller itself, which in turn becomes a continuous relearning process that can capture and track the system variants.
  • FIG. 6B is a flow diagram showing an embodiment 600 of a process for stabilizing the system against drift using the trained NN pattern recognizer in a feedback loop with an incremental learning I relearning loop 602. As with FIG. 6A, this flow diagram can serve as the basis for creating code for executable instructions for processor-based implementations. A number of the same elements from FIG. 6A are shown in this figure.
  • the NN pattern recognizer 604 becomes a mapping function ( -1 ), which is opposite to the combining system (f) 518.
  • the NN pattern recognizer 604 maps from the intensity space (/) 516 to the control variable/phase space ( ⁇ p). Accordingly, the NN pattern recognizer 604 becomes a device that learns which differences in interference patterns are correlated with which vectors in phase space (control variable space), so that given a measured diffractive interference pattern t)
  • Eq. 1 2D convolution
  • FIG. 7 schematically illustrates a generalized embodiment 700 of a neural network (NN) 702 trained and configured for stabilizing drift in a system 704.
  • the notation d(t) 706 indicates drift I noise I perturbation in the system that can affect the system output y(t) 708.
  • the NN is trained with multiple samples of system output y(t) and other information if desired.
  • the sample that is closest to the desired output is selected as a target 716.
  • the NN 702 After the NN 702 is trained, the sensor 710 measures the real-time system output y(t) 708 and the controller 712 corrects the system for drift with control signals u(t) 714. The NN 702 generates control variables for the controller 712 which in turn generates the control signals u(t) 714.
  • the NN 720 maps s(t) 718 to system error.
  • the apparatus can be embodied in other equivalent ways.
  • the NN 702 and sensor 710 could be implemented as a single device, such as a smart sensor, where the NN serves as an algorithm to detect distance vectors to the target in the control variables' space, based on, for example, a conventional sensor output (e.g., camera images).
  • a conventional sensor output e.g., camera images.
  • Embodiments of the present technology may be described herein with reference to flowchart illustrations of methods and systems according to embodiments of the technology, and/or procedures, algorithms, steps, operations, formulae, or other computational depictions, which may also be implemented as computer program products.
  • each block or step of a flowchart, and combinations of blocks (and/or steps) in a flowchart, as well as any procedure, algorithm, step, operation, formula, or computational depiction can be implemented by various means, such as hardware, firmware, and/or software including one or more computer program instructions embodied in computer-readable program code.
  • any such computer program instructions may be executed by one or more computer processors, including without limitation a general purpose computer or special purpose computer, or other programmable processing apparatus to produce a machine, such that the computer program instructions which execute on the computer processor(s) or other programmable processing apparatus create means for implementing the function(s) specified.
  • blocks of the flowcharts, and procedures, algorithms, steps, operations, formulae, or computational depictions described herein support combinations of means for performing the specified function(s), combinations of steps for performing the specified function(s), and computer program instructions, such as embodied in computer-readable program code logic means, for performing the specified function(s).
  • each block of the flowchart illustrations, as well as any procedures, algorithms, steps, operations, formulae, or computational depictions and combinations thereof described herein can be implemented by special purpose hardware-based computer systems which perform the specified function(s) or step(s), or combinations of special purpose hardware and computer-readable program code.
  • these computer program instructions may also be stored in one or more computer-readable memory or memory devices that can direct a computer processor or other programmable processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory or memory devices produce an article of manufacture including instruction means which implement the function specified in the block(s) of the flowchart(s).
  • the computer program instructions may also be executed by a computer processor or other programmable processing apparatus to cause a series of operational steps to be performed on the computer processor or other programmable processing apparatus to produce a computer-implemented process such that the instructions which execute on the computer processor or other programmable processing apparatus provide steps for implementing the functions specified in the block(s) of the flowchart(s), procedure (s) algorithm(s), step(s), operation(s), formula(e), or computational depiction(s).
  • programming or “program executable” as used herein refer to one or more instructions that can be executed by one or more computer processors to perform one or more functions as described herein.
  • the instructions can be embodied in software, in firmware, or in a combination of software and firmware.
  • the instructions can be stored local to the device in non-transitory media, or can be stored remotely such as on a server, or all or a portion of the instructions can be stored locally and remotely. Instructions stored remotely can be downloaded (pushed) to the device by user initiation, or automatically based on one or more factors.
  • processor hardware processor, computer processor, central processing unit (CPU), and computer are used synonymously to denote a device capable of executing the instructions and communicating with input/output interfaces and/or peripheral devices, and that the terms processor, hardware processor, computer processor, CPU, and computer are intended to encompass single or multiple devices, single core and multicore devices, and variations thereof.
  • a machine-learning based stabilized beam combining apparatus comprising: an optical phase controller; an optical beam combining system; a neural network configured to be trained with multi-state dither information from the optical beam combining system; wherein the neural network is configured to, after being trained with labelled data as said multi-state dither, to map (i) a target and (ii) interference diffractive patterns measured from the optical beam combining system, to error in the interference diffractive patterns measured from the optical beam combining system, and compare the error to a reference; wherein, based on said comparison, the neural network is configured to generate phase control variables as feedback on error correction for the optical phase controller to compensate for drift and noise in the optical beam combining system and adjust system output to near target; and wherein the neural network is configured to send the generated phase control variables to the optical phase controller, whereby the optical phase controller can use the generated phase control variables to stabilize the optical beam combining system against drift and noise.
  • the feedback is configured to feed the neural network, after training, a current measurement, which need not be contained in a training dataset, together with a desired pattern in the observation space, from which the neural network predicts the action needed to move apparatus output between these two states in a deterministic way.
  • periodical non-uniqueness mapping comprises interferometric control on coherent beam combining.
  • a neural network for pattern recognition that has been trained with multi-state dither information.
  • a neural network positioned in a feedback loop to stabilize drift in a system wherein the neural network has been trained with multi-state dither information and corrects system output error by mapping system output to a target.
  • An apparatus for stabilizing drift in a system comprising: a neural network that is trained with output signals from the system, wherein the trained neural network maps a target and output signals from the system to system error, compares the system error to a reference, and generates control variables for a controller coupled to the system to adjust system output to near target whereby the system is stabilized against drift.
  • a method for stabilizing drift in a system comprising: training a neural network with output signals from the system, wherein the trained neural network maps a target and output signals from the system to system error, compares the system error to a reference, and generates control variables for a controller coupled to the system to adjust system output to near target whereby the system is stabilized against drift.
  • a machine-learning based apparatus for stabilizing drift in a system comprising: a neural network configured to be trained with measured output signals from the system, the measured output signals including system drift; wherein the neural network is configured to, after being trained, map the output signals and a target to system error and compare the system error to a reference; wherein, based on said comparison, the neural network is configured to generate control variables for a controller to compensate for the system drift and adjust output signals from the system to near target; and wherein the neural network is configured to send the generated control variables to the controller, whereby the controller can use the generated control variables to stabilize the system against drift.
  • a neural network positioned in a feedback loop between the controller input and the system output; the neural network configured to be trained with measured output signals from the system, the measured output signals including system drift; wherein the neural network is configured to, after being trained, map the measured output signals and a target to system error and compare the system error to a reference; wherein, based on said comparison, the neural network is configured to generate control variables for the controller to compensate for the system drift and adjust the system output near target; and wherein the neural network is configured to send the generated control variables to the controller, whereby the controller can use the generated control variables to stabilize the system against drift.
  • a method for stabilizing a system against drift the system having a system input, a system output, a controller coupled to the system input, a controller input, and a controller output, the method comprising: positioning a neural network in a feedback loop between the controller input and the system output; training the neural network with output signals from the system, the measured output signals including system drift; wherein after being trained the neural network maps the system output signals and a target to system error and compares the system error to a reference; wherein, based on said comparison, the neural network generates control variables for the controller to compensate for the system drift and adjust output signals from the system to near target; and sending the generated control variables to the controller input, whereby the controller can use the generated control variables to stabilize the system against drift.
  • a machine-learning based stabilized beam combining apparatus comprising: an optical phase controller; an optical beam combining system; a neural network configured to be trained with multi-state dither information from the optical beam combining system; wherein the neural network is configured to, after being trained, map (i) a target and (ii) interference diffractive patterns measured from the optical beam combining system, to error in the interference diffractive patterns measured from the optical beam combining system, and compare the error to a reference; wherein, based on said comparison, the neural network is configured to generate phase control variables for the optical phase controller to compensate for drift in the optical beam combining system and adjust system output to near target; and wherein the neural network is configured to send the generated phase control variables to the optical phase controller, whereby the optical phase controller can use the generated phase control variables to stabilize the optical beam combining system against drift.
  • a machine-learning based apparatus for stabilizing an optical beam combining system against drift comprising: a neural network configured to be trained with multi-state dither information from an optical beam combining system; wherein the neural network is configured to, after being trained, map (i) a target and (ii) interference diffractive patterns measured from the optical beam combining system, to error in the measured interference diffractive, and compare the error to a reference; wherein, based on said comparison, the neural network is configured to determine phase control variables for the optical phase controller to compensate for drift in the optical beam combining system and adjust system output to near target; and wherein the neural network is configured to send the generated phase control variables to the optical phase controller, whereby the optical phase controller can use the generated phase control variables to stabilize the optical beam combining system against drift.
  • an optical beam combining system having a system input, a system output, an optical phase controller coupled to the system input, the optical phase controller having an input and an output
  • the improvement comprising: a neural network positioned in a feedback loop between the controller input and the system output; the neural network configured to be trained with multi-state dither information from the optical beam combining system; wherein the neural network is configured to, after being trained, map (i) a target and (ii) interference diffractive patterns measured from the optical beam combining system, to error in the measured interference diffractive patterns, and compare the error to a reference; wherein, based on said comparison, the neural network is configured to determine phase control variables for the optical phase controller to compensate for the system drift and adjust the system output to near target; and wherein the neural network is configured to send the generated phase control variables to the optical phase controller, whereby the optical phase controller can use the generated phase control variables to stabilize the system against drift.
  • a method for stabilizing an optical beam combining system against drift comprising: positioning a neural network positioned in a feedback loop between the controller input and the system output; training the neural network with multi-state dither information from the optical beam combining system; wherein after the neural network is trained, the neural network maps (i) a target and (ii) interference diffractive patterns measured from the optical beam combining system, to error in the measured interference diffractive patterns, and compares the error to a reference; wherein, based on said comparison, the neural network determines phase control variables for the optical phase controller to compensate for the system drift and adjust the system output to near target; and sending the generated phase control variables to the optical phase controller, whereby the optical phase controller can use the generated control variables to stabilize the system against drift.
  • Phrasing constructs such as “A, B and/or C”, within the present disclosure describe where either A, B, or C can be present, or any combination of items A, B and C.
  • references in this disclosure referring to “an embodiment”, “at least one embodiment” or similar embodiment wording indicates that a particular feature, structure, or characteristic described in connection with a described embodiment is included in at least one embodiment of the present disclosure. Thus, these various embodiment phrases are not necessarily all referring to the same embodiment, or to a specific embodiment which differs from all the other embodiments being described.
  • the embodiment phrasing should be construed to mean that the particular features, structures, or characteristics of a given embodiment may be combined in any suitable manner in one or more embodiments of the disclosed apparatus, system or method.
  • a set refers to a collection of one or more objects.
  • a set of objects can include a single object or multiple objects.
  • Relational terms such as first and second, top and bottom, upper and lower, left and right, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions.
  • the terms can refer to a range of variation of less than or equal to ⁇ 10% of that numerical value, such as less than or equal to ⁇ 5%, less than or equal to ⁇ 4%, less than or equal to ⁇ 3%, less than or equal to ⁇ 2%, less than or equal to ⁇ 1 %, less than or equal to ⁇ 0.5%, less than or equal to ⁇ 0.1 %, or less than or equal to ⁇ 0.05%.
  • substantially aligned can refer to a range of angular variation of less than or equal to ⁇ 10°, such as less than or equal to ⁇ 5°, less than or equal to ⁇ 4°, less than or equal to ⁇ 3°, less than or equal to ⁇ 2°, less than or equal to ⁇ 1 °, less than or equal to ⁇ 0.5°, less than or equal to ⁇ 0.1 °, or less than or equal to ⁇ 0.05°.
  • Coupled as used herein is defined as connected, although not necessarily directly and not necessarily mechanically.
  • a device or structure that is “configured” in a certain way is configured in at least that way, but may also be configured in ways that are not listed.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Feedback Control In General (AREA)

Abstract

L'invention concerne un contrôleur d'apprentissage automatique (ML) et un procédé pour des systèmes qui peuvent apprendre à les stabiliser sur la base de mesures d'un système instable. Ceci permet d'assurer un apprentissage sur un système non encore contrôlé et d'assurer un apprentissage continu au fur et à mesure que le stabilisateur fonctionne.
PCT/US2022/045236 2021-10-01 2022-09-29 Contrôleur de stabilisation basé sur l'apprentissage automatique qui peut apprendre sur un système instable WO2023055938A2 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US202163251346P 2021-10-01 2021-10-01
US63/251,346 2021-10-01

Publications (2)

Publication Number Publication Date
WO2023055938A2 true WO2023055938A2 (fr) 2023-04-06
WO2023055938A3 WO2023055938A3 (fr) 2023-05-11

Family

ID=85783505

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2022/045236 WO2023055938A2 (fr) 2021-10-01 2022-09-29 Contrôleur de stabilisation basé sur l'apprentissage automatique qui peut apprendre sur un système instable

Country Status (1)

Country Link
WO (1) WO2023055938A2 (fr)

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101973660B1 (ko) * 2010-10-22 2019-04-30 일렉트로 싸이언티픽 인더스트리이즈 인코포레이티드 빔 디더링 및 스카이빙을 위한 레이저 처리 시스템 및 방법
US11354586B2 (en) * 2019-02-15 2022-06-07 Q Bio, Inc. Model parameter determination using a predictive model
FI129608B (en) * 2020-01-08 2022-05-31 Nokia Technologies Oy MACHINE LEARNING SIGNAL DETECTION AND RECONSTRUCTION OF LOW RESOLUTION SAMPLES

Also Published As

Publication number Publication date
WO2023055938A3 (fr) 2023-05-11

Similar Documents

Publication Publication Date Title
US10596698B2 (en) Machine learning device, robot control system, and machine learning method
Wang et al. Stabilization of the 81-channel coherent beam combination using machine learning
KR102582515B1 (ko) 양자 컴퓨팅 장치의 비선형 교정
Yu et al. Design of data-driven PID controllers with adaptive updating rules
CN103814387A (zh) 用于尖峰神经网络中的结构化延迟可塑性的方法和装置
Sun et al. Identification and adaptive control of a high-contrast focal plane wavefront correction system
Wilson An Adaptive Phase-Amplitude Reduction Framework without O(ϵ) Constraints on Inputs
Lai et al. Model-free optimal control of discrete-time systems with additive and multiplicative noises
Wang et al. Machine learning pattern recognition algorithm with applications to coherent laser combination
Nakamura-Zimmerer et al. A causality-free neural network method for high-dimensional Hamilton-Jacobi-Bellman equations
Annaswamy Adaptive control and intersections with reinforcement learning
Du et al. Experimental beam combining stabilization using machine learning trained while phases drift
WO2023055938A2 (fr) Contrôleur de stabilisation basé sur l'apprentissage automatique qui peut apprendre sur un système instable
Verhoek et al. Direct data-driven state-feedback control of general nonlinear systems
Zhang et al. Data-driven learning control algorithms for unachievable tracking problems
Preti et al. Continuous quantum gate sets and pulse-class meta-optimization
Kucukdemiral et al. Data-driven H∞ control of constrained systems: An application to bilateral teleoperation system
Zioui et al. Design of a new hybrid linearising-backstepping controller using quantum state equations and quantum spins
Lin et al. Resilient H∞ dynamic output feedback controller design for USJSs with time-varying delays
JP7360162B2 (ja) 制御系設計方法及び制御装置
Romano et al. Matchable-observable linear models for multivariable identification: Structure selection and experimental results
Lin et al. Machine learning-based error compensation for high precision laser arbitrary beam splitting
Pratheep et al. Genetic algorithm–based robust controller for an inverted pendulum using model order reduction
Valluru et al. Evaluation of Moth-Flame Optimization, Genetic and Simulated Annealing tuned PID controller for Steering Control of Autonomous Underwater Vehicle
JP4809268B2 (ja) 電力系統縮約モデル作成装置、電力系統縮約モデル作成方法および電力系統縮約モデル作成プログラム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22877332

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase

Ref country code: DE