CN110196642B - Navigation type virtual microscope based on intention understanding model - Google Patents
Navigation type virtual microscope based on intention understanding model Download PDFInfo
- Publication number
- CN110196642B CN110196642B CN201910543153.3A CN201910543153A CN110196642B CN 110196642 B CN110196642 B CN 110196642B CN 201910543153 A CN201910543153 A CN 201910543153A CN 110196642 B CN110196642 B CN 110196642B
- Authority
- CN
- China
- Prior art keywords
- image
- sample
- information
- user
- microscope
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B21/00—Microscopes
- G02B21/24—Base structure
- G02B21/241—Devices for focusing
- G02B21/242—Devices for focusing with coarse and fine adjustment mechanism
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/73—Deblurring; Sharpening
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Chemical & Material Sciences (AREA)
- Analytical Chemistry (AREA)
- Optics & Photonics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
本发明提出了一种基于意图理解模型的导航式虚拟显微镜,包括多模态输入与感知模块、多模态信息整合模块和交互应用模块;多模态输入与感知模块用于通过传声器获得用户的语音信息,以及获得用户的操作行为;多模态信息整合模块用于将语音信息通过视觉通道信息处理和将操作行为通过触觉通道信息处理,然后将处理后的语音信息和操作行为通过多通道信息整合,完成显微镜和用户之间的交互。本发明通过多模态信息获取和整合,利用简单的传感元件,加入了多种模态的信号输入和智能化感应技术,在保证数字显微镜优点的基础上,让广大普通和贫困中学学生也能有条件进行显微镜的学习,增加对微观世界的认知感受,和体验智能显微镜。
The invention provides a navigational virtual microscope based on an intention understanding model, which includes a multi-modal input and perception module, a multi-modal information integration module and an interactive application module; the multi-modal input and perception module is used to obtain the user's information through a microphone. Voice information, and obtain the user's operation behavior; the multimodal information integration module is used to process the voice information through the visual channel information and the operation behavior through the tactile channel information, and then process the processed voice information and operation behavior through the multi-channel information. Integration, completes the interaction between the microscope and the user. The invention acquires and integrates multi-modal information, utilizes simple sensing elements, adds multi-modal signal input and intelligent sensing technology, and on the basis of ensuring the advantages of digital microscopes, allows the majority of ordinary and poor middle school students to also Ability to learn microscopes, increase cognitive experience of the microscopic world, and experience intelligent microscopes.
Description
技术领域technical field
本发明属于智能显微镜设计的技术领域,特别涉及一种基于意图理解模型的导航式虚拟显微镜。The invention belongs to the technical field of intelligent microscope design, and particularly relates to a navigational virtual microscope based on an intention understanding model.
背景技术Background technique
显微镜的使用,是中学生物科学学习的一个重点内容,也是考试过程中,学生经常犯错的内容,在学习过程中,掌握好显微镜的使用规则,对学好理论知识和对微观世界的探究都大有裨益。然而,中学经常使用的和教材介绍的显微镜,都是十分传统的单目显微镜,学生在使用显微镜过程中,常常会因为调节焦距和合适的亮度花费大量的时间,而且成像效果不好,一堂实验课下来,学生情绪往往从初始的兴奋转为疲惫不堪。传统的教学实验课,一般是老师在讲台演示,学生在下面观察,即使学生操作过程比较规范,也有可能得不到好的结果,教师往往奔走在不同学生之间,导致师生之间交流不够有效,导致学生感性层面上认知生物会有偏差。因此,如何使学生正确观察实际的生物结构和现象变得尤为重要。更局限的是,每一个显微镜同一时间只能是一人观察,老师想要看到学生的实验结果,就必须逐一检查学生显微镜下的像,不同学生之间也无法达到共享实验结果的目的,因此,若能将成功的实验结果和错误的案例结果展示给学生和教师,对学生的学习和和激励引导都会产良好的引导作用,同时也为老师评估学生的实验情况带来极大的便利。The use of microscopes is a key content in the study of biological sciences in middle school, and it is also a content that students often make mistakes during the examination process. In the process of learning, mastering the rules of using microscopes is of great benefit to learning theoretical knowledge and exploring the microscopic world. . However, the microscopes often used in middle schools and introduced in the textbooks are very traditional monocular microscopes. Students often spend a lot of time adjusting the focus and proper brightness in the process of using the microscope, and the imaging effect is not good. One experiment After class, students' mood often changes from initial excitement to exhaustion. In traditional teaching experimental classes, teachers usually demonstrate on the podium and students observe below. Even if the students' operation process is relatively standardized, they may not get good results. Teachers often run between different students, resulting in insufficient communication between teachers and students. Effective, causing students to have biases in cognitive biology at the perceptual level. Therefore, how to enable students to correctly observe the actual biological structures and phenomena becomes particularly important. What is more limited is that each microscope can only be observed by one person at the same time. If the teacher wants to see the experimental results of the students, he must check the images under the microscopes of the students one by one, and the purpose of sharing the experimental results between different students cannot be achieved. Therefore, , if successful experimental results and wrong case results can be displayed to students and teachers, it will have a good guiding effect on students' learning and motivation, and it will also bring great convenience for teachers to evaluate students' experimental conditions.
目前正在普及的是数字显微镜。所谓数字显微镜,就是能与电脑连接的,将在显微镜视野内观察到的结构或现象通过软件显示在电脑的显示屏上,使得观察到的结构或现象更加清晰,而且可以让多人同时观看。数字显微镜有以下优点:第一,物像呈现在电脑显示屏上,降低寻找物像难度;第二,教师可实时监控学生电脑显示屏,迅速判断学生操作问题,及时进行调整;第三,如果学生实验失败,教师可以电脑共享其他小组成功的物像截图,督促学生分析失败的原因。利用数字显微镜,可以使实验的评价过程简化,可行性提高,覆盖面扩大。但是,数字显微镜的造价还是比较昂贵的,最普通的显微镜价格要在2000元左右,对于一般的或者较贫困的中学来说,大量购买显微镜设备成为困难。What is currently gaining popularity is the digital microscope. The so-called digital microscope can be connected to a computer, and the structures or phenomena observed in the microscope field of view are displayed on the computer display screen through software, so that the observed structures or phenomena are clearer and can be viewed by multiple people at the same time. The digital microscope has the following advantages: first, the object image is displayed on the computer screen, which reduces the difficulty of finding the object image; second, the teacher can monitor the student's computer screen in real time, quickly judge the students' operation problems, and make adjustments in time; third, if When students fail in experiments, teachers can share screenshots of other groups' successful images on the computer, urging students to analyze the reasons for failure. Using a digital microscope can simplify the evaluation process of the experiment, improve the feasibility and expand the coverage. However, the cost of digital microscopes is still relatively expensive. The price of the most common microscopes is about 2,000 yuan. For ordinary or poor middle schools, it becomes difficult to purchase microscope equipment in large quantities.
发明内容SUMMARY OF THE INVENTION
本发明提出了一种基于意图理解模型的导航式虚拟显微镜,让广大普通和贫困中学学生也能有条件进行显微镜的学习,增加对微观世界的认知感受,和体验智能显微镜。The invention proposes a navigational virtual microscope based on an intention understanding model, so that ordinary and poor middle school students can also have the conditions to study microscopes, increase their cognition and experience of the microscopic world, and experience intelligent microscopes.
为了实现上述目的,本发明提出的一种基于意图理解模型的导航式虚拟显微镜,包括多模态输入与感知模块、多模态信息整合模块和交互应用模块;In order to achieve the above purpose, a navigational virtual microscope based on an intention understanding model proposed by the present invention includes a multimodal input and perception module, a multimodal information integration module and an interactive application module;
所述多模态输入与感知模块用于通过传声器获得用户的语音信息,以及获得用户的操作行为;所述操作行为包括通过转动传感器获得用户调节显微镜粗细准焦螺旋的方向和通过图像采集设备识别待观察样本以及检测待观察样本运动的图像信息;The multimodal input and perception module is used to obtain the user's voice information through the microphone, and obtain the user's operation behavior; the operation behavior includes obtaining the direction of the user's adjustment of the microscope thickness and quasi-focus spiral by rotating the sensor and identifying it through the image acquisition device. The sample to be observed and the image information to detect the movement of the sample to be observed;
所述多模态信息整合模块用于将语音信息通过视觉通道信息处理和将操作行为通过触觉通道信息处理,然后将处理后的语音信息和操作行为通过多通道信息整合,完成显微镜和用户之间的交互;The multimodal information integration module is used to process the voice information through the visual channel information and the operation behavior through the tactile channel information, and then integrate the processed voice information and operation behavior through the multi-channel information to complete the communication between the microscope and the user. interaction;
所述交互应用模块通过视觉显示和听觉指引完成对用户意图进行预测以及给出操作引导。The interactive application module predicts the user's intention and provides operation guidance through visual display and auditory guidance.
进一步的,所述通过图像采集设备识别待观察样本的方法为,载玻片的上表面设置有表示所述待观察样本的二维码图片,图像采集设备识别不同的二维码,进而从数据库中调取样本图像。Further, the method for identifying the sample to be observed by the image acquisition device is that the upper surface of the glass slide is provided with a two-dimensional code picture representing the sample to be observed, and the image acquisition device recognizes different two-dimensional codes, and then retrieves the data from the database. to retrieve a sample image.
进一步的,所述视觉通道信息处理的方法为:识别样本的原始RGB图像,其中D为样本图像合集;观察样本的变化,当前观察到的样本的图像为P,其中移动变换函数为PW(),新生成的图像为P’,则位移变换过程公式为P’(X,Y)=PM(X-△X,Y-△Y);Further, the method for processing the visual channel information is: identifying the original RGB image of the sample, where D is a collection of sample images; observing the change of the sample, the image of the currently observed sample is P, and the moving transformation function is PW() , the newly generated image is P', then the displacement transformation process formula is P'(X, Y)=PM(X-△X, Y-△Y);
所述P∈D,X,Y)∈P,△X、△Y分别为图像P中像素点的偏移量。The P∈D, X, Y)∈P, ΔX and ΔY are the offsets of the pixels in the image P, respectively.
进一步的,所述检测待观察样本运动的图像信息的方法为载玻片的下表面设置有带有检测标记的圆面,在操作过程时待检测样本发生位置移动,利用带有检测标记的圆面和所述位移变换过程的公式,计算出样本位置移动,根据偏移量△X和△Y观察到样本图像。Further, the method for detecting the image information of the movement of the sample to be observed is that the lower surface of the slide glass is provided with a circular surface with detection marks, the position of the sample to be detected moves during the operation process, and the circular surface with detection marks is used. According to the formula of the surface and the displacement transformation process, the sample position movement is calculated, and the sample image is observed according to the offsets ΔX and ΔY.
进一步的,所述触觉通道信息处理的方法为:通过调节粗准焦螺旋或细准焦螺旋来调节图像清晰度,其中粗准焦螺旋或细准焦螺旋的转动变化函数为PV(),调节粗准焦螺旋,得到离散值t∈T{11,12,13,14,15,16},Sxy表示中心点在(x,y)处、滤波窗口的大小为mxn、均值滤波函数为PV,另外所以根据算术均值滤波函数可以得到新生成的图像P″Further, the method for processing the tactile channel information is: adjusting the image sharpness by adjusting the coarse focal-focus screw or the fine focal-focus screw, wherein the rotation change function of the coarse quasi-focus screw or the fine quasi-focus screw is PV( ), and the adjustment Coarse quasi-focus spiral, get discrete values t∈T{11,12,13,14,15,16}, S xy means the center point is at (x, y), the size of the filter window is mxn, and the mean filter function is PV ,in addition Therefore, the newly generated image P" can be obtained according to the arithmetic mean filter function.
调节细准焦螺旋,得到离散值s,s∈S{1,2,3,4,5,6},Sxy表示中心点在(x,y)处、滤波窗口的大小为mxn、均值滤波函数为PV,另外 Adjust the fine quasi-focus spiral to obtain discrete values s, s∈S{1,2,3,4,5,6}, S xy indicates that the center point is at (x, y), the size of the filter window is mxn, and the mean filter The function is PV, and in addition
所以根据算术均值滤波函数可以得到新生成的图像P″Therefore, the newly generated image P" can be obtained according to the arithmetic mean filter function.
所述P(x,y)为当前未经粗准焦螺旋调节或者未经细准焦螺旋调节的图像。The P(x, y) is the current image that has not been adjusted by the coarse focal-focus screw or not adjusted by the fine-focus screw.
进一步的,所述多通道信息整合包括双通道信息整合和三通道信息整合;Further, the multi-channel information integration includes dual-channel information integration and three-channel information integration;
所述双通道信息整合用于调整粗准焦螺旋或者细准焦螺旋中的任意一个,移动样本并观察样本的数据整合;多通道整合函数为Muf(),未调节前的图像为P,移动变换函数为PM(),转动变换函数为PV(),因此,多通道整合函数可以表示为:The two-channel information integration is used to adjust either the coarse focal spiral or the fine focal spiral, move the sample and observe the data integration of the sample; the multi-channel integration function is Muf(), the unadjusted image is P, and the moving The transformation function is PM(), and the rotation transformation function is PV(). Therefore, the multi-channel integration function can be expressed as:
Muf(P)=ɑPM(P)+(1-ɑ)PV(P);Muf(P)=ɑPM(P)+(1-ɑ)PV(P);
所述ɑ为选择参数,所述ɑ=0,1,转动传感器转动,当完成一次调节后,P″′=Muf(P);P″′作为当前观察图像可继续被调节;The ɑ is the selection parameter, the ɑ=0, 1, the rotation sensor rotates, when one adjustment is completed, P″′=Muf(P); P″′ can be adjusted as the current observation image;
所述三通道信息整合用于在整合策略中输入图像信息、操作行为和语音信息,在当前状态条件下,系统根据不同的用户行为进入不同的系统状态。The three-channel information integration is used to input image information, operation behavior and voice information in the integration strategy. Under the current state conditions, the system enters different system states according to different user behaviors.
进一步的,所述操作指引为在用户使用导航式虚拟显微镜时,语音提示操作步骤或操作方法使待测样本位于显微镜的视野内。Further, the operation guide is that when the user uses the navigation virtual microscope, the voice prompts the operation steps or the operation method so that the sample to be tested is located in the field of view of the microscope.
进一步的,所述图像保存为在转动粗准焦螺旋或者细准焦螺旋时,当视野内的样本清晰度达到设定的阈值,语音提示保存图像,待用户确认后,虚拟显微镜系统将当前视野内的图像保存到指定文件夹中。Further, the image is saved as when the coarse focus screw or the fine focus screw is rotated, when the clarity of the sample in the field of view reaches the set threshold, a voice prompt will be given to save the image, and after the user confirms, the virtual microscope system will save the current field of view. The images in it are saved to the specified folder.
发明内容中提供的效果仅仅是实施例的效果,而不是发明所有的全部效果,上述技术方案中的一个技术方案具有如下优点或有益效果:The effects provided in the summary of the invention are only the effects of the embodiments, rather than all the effects of the invention. One of the above technical solutions has the following advantages or beneficial effects:
本发明实施例提出了一种基于意图理解模型的导航式虚拟显微镜,包括多模态输入与感知模块、多模态信息整合模块和交互应用模块;多模态输入与感知模块用于通过传声器获得用户的语音信息,以及获得用户的操作行为;其中操作行为包括通过转动传感器获得用户调节显微镜粗细准焦螺旋的方向和通过图像采集设备识别待观察样本以及检测待观察样本运动的图像信息;多模态信息整合模块用于将语音信息通过视觉通道信息处理和将操作行为通过触觉通道信息处理,然后将处理后的语音信息和操作行为通过多通道信息整合,完成显微镜和用户之间的交互。交互应用模块通过视觉显示和听觉指引完成对用户意图进行预测以及给出操作引导。本发明通过多模态信息获取和整合,在显微镜的调节过程中,通过语音信息、操作行为的获取等调节粗细准焦螺旋,移动载玻片,切换不同的目镜和物镜观察不同倍数下的样本图像。对于学生而言,语音和视觉呈现是自然的交互方式,深度学习技术使得语音识别,语音转化为文本信息和语音合成技术更加准确,也使得人机交互过程更加灵活智能。在图像处理方面,智能显微镜借助图像处理技术能够准确的识别到不同的样本标签,实时计算出跟踪点的位置,按照一定的规则,控制样本图像的移动。本发明利用简单的传感元件,加入了多种模态的信号输入和智能化感应技术,在保证数字显微镜优点的基础上,让广大普通和贫困中学学生也能有条件进行显微镜的学习,增加对微观世界的认知感受,和体验智能显微镜。The embodiment of the present invention proposes a navigable virtual microscope based on an intention understanding model, which includes a multimodal input and perception module, a multimodal information integration module and an interactive application module; the multimodal input and perception module is used to obtain information through a microphone. The user's voice information, and the user's operation behavior; the operation behavior includes obtaining the direction of the user's adjustment of the microscope thickness and quasi-focus spiral by rotating the sensor, and identifying the sample to be observed and detecting the image information of the movement of the sample to be observed through the image acquisition device; multi-mode The state information integration module is used to process the speech information through the visual channel information and the operation behavior through the tactile channel information, and then integrate the processed speech information and operation behavior through the multi-channel information to complete the interaction between the microscope and the user. The interactive application module predicts the user's intention and gives operation guidance through visual display and auditory guidance. Through the acquisition and integration of multimodal information, the present invention adjusts the thickness of the quasi-focus spiral through the acquisition of voice information and operation behaviors, moves the slide, and switches between different eyepieces and objective lenses to observe samples under different magnifications during the adjustment process of the microscope. image. For students, speech and visual presentation are natural interaction methods. Deep learning technology makes speech recognition, speech-to-text information and speech synthesis technology more accurate, and also makes the human-computer interaction process more flexible and intelligent. In terms of image processing, the intelligent microscope can accurately identify different sample labels with the help of image processing technology, calculate the position of the tracking point in real time, and control the movement of the sample image according to certain rules. The invention utilizes a simple sensing element, adds a variety of modal signal input and intelligent sensing technology, on the basis of ensuring the advantages of the digital microscope, so that the majority of ordinary and poor middle school students can also have the conditions to study the microscope, increasing the Cognitive feelings of the microscopic world, and experience the intelligent microscope.
附图说明Description of drawings
附图1是本发明实施例1提出的一种基于意图理解模型的导航式虚拟显微镜多模态交互的总体框架图;1 is an overall frame diagram of a multi-modal interaction of a navigation virtual microscope based on an intention understanding model proposed in
附图2是本发明实施例1提出的一种基于意图理解模型的导航式虚拟显微镜中三通道信息整合策略图;2 is a diagram of a three-channel information integration strategy in a navigational virtual microscope based on an intention understanding model proposed in
附图3是本发明实施例1提出的一种基于意图理解模型的导航式虚拟显微镜实现的结构示意图;3 is a schematic structural diagram of the realization of a navigation virtual microscope based on an intention understanding model proposed in
附图4是本发明实施例1提出的一种基于意图理解模型的导航式虚拟显微镜的目镜结构示意图;4 is a schematic diagram of the eyepiece structure of a navigation virtual microscope based on an intention understanding model proposed in
其中:1-通光孔;2-载物台;3-摄像头;4-载玻片下表面;5-载玻片1上表面;6-载玻片2上表面。Among them: 1- light hole; 2- stage; 3- camera; 4- lower surface of glass slide; 5- upper surface of
具体实施方式Detailed ways
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only a part of the embodiments of the present invention, but not all of the embodiments. Based on the embodiments of the present invention, all other embodiments obtained by those of ordinary skill in the art without creative efforts shall fall within the protection scope of the present invention.
在本发明的描述中,需要理解的是,术语“纵向”、“横向”、“上”、“下”、“前”、“后”、“左”、“右”、“竖直”、“水平”、“顶”、“底”、“内”、“外”等指示的方位或位置关系为基于附图所示的方位或位置关系,仅是为了便于描述本发明,而不是指示或暗示所指的装置或元件必须具有特定的方位、以特定的方位构造和操作,因此不能理解为对本发明的限制。In the description of the present invention, it should be understood that the terms "portrait", "horizontal", "upper", "lower", "front", "rear", "left", "right", "vertical", The orientation or positional relationship indicated by "horizontal", "top", "bottom", "inner", "outer", etc. is based on the orientation or positional relationship shown in the drawings, and is only for the convenience of describing the present invention, rather than indicating or It is implied that the device or element referred to must have a particular orientation, be constructed and operate in a particular orientation, and therefore should not be construed as limiting the invention.
实施例1Example 1
本发明实施例1提出了一种基于意图理解模型的导航式虚拟显微镜,包括括多模态输入与感知模块、多模态信息整合模块和交互应用模块;
多模态输入与感知模块用于通过传声器获得用户的语音信息,以及获得用户的操作行为;操作行为包括通过转动传感器获得用户调节显微镜粗细准焦螺旋的方向和通过图像采集设备识别待观察样本以及检测待观察样本运动的图像信息;The multi-modal input and perception module is used to obtain the user's voice information through the microphone, and obtain the user's operation behavior; the operation behavior includes obtaining the direction of the user's adjustment of the microscope's thickness and quasi-focus spiral by rotating the sensor, and identifying the sample to be observed through the image acquisition device. Detect the image information of the movement of the sample to be observed;
多模态信息整合模块用于将语音信息通过视觉通道信息处理和将操作行为通过触觉通道信息处理,然后将处理后的语音信息和操作行为通过多通道信息整合,完成显微镜和用户之间的交互;The multimodal information integration module is used to process the speech information through the visual channel information and the operation behavior through the tactile channel information, and then integrate the processed speech information and operation behavior through the multi-channel information to complete the interaction between the microscope and the user ;
交互应用模块通过视觉显示和听觉指引完成对用户意图进行预测以及给出操作引导。The interactive application module predicts the user's intention and gives operation guidance through visual display and auditory guidance.
如图1给出了本发明实施例1提出的一种基于意图理解模型的导航式虚拟显微镜多模态交互的总体框架图;Figure 1 shows an overall frame diagram of a multi-modal interaction of a navigation virtual microscope based on an intention understanding model proposed in
多模态输入与感知模块包括多模态输入和交互设备,使用传声器获取用户的语音信息,其中传声器采用麦克风,MIC与语音识别设备相连,用户的操作行通过触觉和视觉传达,即来源于转动传感器和图像采集设备,转动传感器用来捕获用户调节粗细准焦螺旋的方向,用于样本清晰度的调节,图像采集设备用来识别样本和检测样本移动。其中图像采集设备使用相机。The multi-modal input and perception module includes multi-modal input and interactive devices, and uses a microphone to obtain the user's voice information. The microphone uses a microphone, and the MIC is connected to the voice recognition device. The user's operation is conveyed through touch and vision, that is, from rotation Sensor and image acquisition device, the rotation sensor is used to capture the direction of the user to adjust the thickness of the quasi-focus screw, which is used to adjust the clarity of the sample, and the image acquisition device is used to identify the sample and detect the movement of the sample. The image acquisition device uses a camera.
通过图像采集设备识别待观察样本的方法为,载玻片的上表面设置有表示所述待观察样本的二维码图片,图像采集设备识别不同的二维码,进而从数据库中调取样本图像。The method of identifying the sample to be observed by the image acquisition device is as follows: the upper surface of the glass slide is provided with a two-dimensional code picture representing the sample to be observed, the image acquisition device recognizes different two-dimensional codes, and then retrieves the sample image from the database .
多模态信息整合模块用于意图理解,将语音信息通过视觉通道信息处理和将操作行为通过触觉通道信息处理,然后将处理后的语音信息和操作行为通过多通道信息整合。将语音信息通过视觉通道信息处理的方法为:识别样本的原始RGB图像,其中D为样本图像合集;观察样本的变化,当前观察到的样本的图像为P,其中移动变换函数为PW(),新生成的图像为P’,则位移变换过程公式为P’(X,Y)=PM(X-△X,Y-△Y);The multimodal information integration module is used for intention understanding, processing the speech information through the visual channel information and the operation behavior through the tactile channel information, and then integrating the processed speech information and operation behavior through the multi-channel information. The method of processing the speech information through the visual channel information is: identify the original RGB image of the sample, where D is the sample image collection; observe the change of the sample, the image of the currently observed sample is P, and the moving transformation function is PW(), The newly generated image is P', then the displacement transformation process formula is P'(X,Y)=PM(X-△X,Y-△Y);
其中P∈D,(X,Y)∈P,△X、△Y分别为图像P中像素点的偏移量。Where P∈D, (X,Y)∈P, △X, △Y are the offsets of the pixels in the image P, respectively.
将操作行为通过触觉通道信息处理的方法为,通过调节粗准焦螺旋或细准焦螺旋来调节图像清晰度,其中粗准焦螺旋或细准焦螺旋的转动变化函数为PV(),调节粗准焦螺旋,得到离散值t∈T{11,12,13,14,15,16},Sxy表示中心点在(x,y)处、滤波窗口的大小为mxn、均值滤波函数为PV,另外所以根据算术均值滤波函数可以得到新生成的图像P″The method of processing the operation behavior through the tactile channel information is to adjust the image sharpness by adjusting the coarse focal-focus screw or the fine focal-focus screw, wherein the rotation change function of the coarse focal-focus screw or the fine-focus screw is PV(), and the adjustment of the coarse focal-focus screw or the fine-focus screw is PV(). Quasi-focal spiral, the discrete value t∈T{11,12,13,14,15,16} is obtained, S xy indicates that the center point is at (x, y), the size of the filter window is mxn, and the mean filter function is PV, in addition Therefore, the newly generated image P" can be obtained according to the arithmetic mean filter function.
调节细准焦螺旋,得到离散值s,s∈S{1,2,3,4,5,6},Sxy表示中心点在(x,y)处、滤波窗口的大小为mxn、均值滤波函数为PV,另外 Adjust the fine quasi-focus spiral to obtain discrete values s, s∈S{1,2,3,4,5,6}, S xy indicates that the center point is at (x, y), the size of the filter window is mxn, and the mean filter The function is PV, and in addition
所以根据算术均值滤波函数可以得到新生成的图像P″Therefore, the newly generated image P" can be obtained according to the arithmetic mean filter function.
其中,P(X,Y)为当前未经粗准焦螺旋调节或者未经细准焦螺旋调节的图像。Among them, P(X,Y) is the current image that has not been adjusted by the coarse focal-focus screw or not adjusted by the fine-focus screw.
将处理后的语音信息和操作行为通过多通道信息整合,其中多通道信息整合包括双通道信息整合和三通道信息整合;双通道信息整合用于调整粗准焦螺旋或者细准焦螺旋中的任意一个,移动样本并观察样本的数据整合;多通道整合函数为Muf(),未调节前的图像为P,移动变换函数为PM(),转动变换函数为PV(),因此,多通道整合函数可以表示为:The processed voice information and operation behavior are integrated through multi-channel information, wherein the multi-channel information integration includes dual-channel information integration and three-channel information integration; dual-channel information integration is used to adjust any of the coarse focal spiral or fine focal spiral. One, move the sample and observe the data integration of the sample; the multi-channel integration function is Muf(), the unadjusted image is P, the moving transformation function is PM(), and the rotation transformation function is PV(). Therefore, the multi-channel integration function It can be expressed as:
Muf(P)=ɑPM(P)+(1-ɑ)PV(P);Muf(P)=ɑPM(P)+(1-ɑ)PV(P);
其中,ɑ为选择参数,ɑ=0,1,转动传感器转动,当完成一次调节后,P″′=Muf(P);P″′作为当前观察图像可继续被调节;Among them, ɑ is the selection parameter, ɑ=0, 1, the rotation sensor rotates, when one adjustment is completed, P″′=Muf(P); P″′ can continue to be adjusted as the current observation image;
附图2给出了本发明实施例1提出的一种基于意图理解模型的导航式虚拟显微镜中三通道信息整合策略图。其中圆圈表示当前执行的状态,有向边指向可能会转变成得状态,有向边上附带了发生转变需要满足的条件。整合策略中的输入包括,图像信息P0,P1,操作行为,由转动传感器获得),语音信息,输出内容是P,P1,P2,和语音提示。其中,P0是样本标签图,P1是控制样本移动的标记图,P,P1,P2,是观察到的样本图像。系统中,在当前状态条件下,系统根据不同的用户行为进入不同的系统状态。在图3中,例如,样本标签P0识别成功后,显示模糊处理后的样本P,用户可以选择移动样本来观察,也可以选择把当前样本调节清晰,如ɑ=0,系统进入样本移动状态,摄像头检测到当前样本移动,实时计算偏移量(△X、△Y),并显示移动后的样本P2,在该状态下,可以继续与Picture Move状态持续“互动”,发生交替式的状态切换,也可以根据用户行为转变到Recognition adjustment状态。C为意图预测变量,它并不是用户输入,而是系统根据当前P2的清晰度,产生的控制条件,当P2满足清晰度要求后,认为用户已经得到最想要观察的样本图像,即推断出用户会停留在此刻观察样本,所以系统引导用户保存当前图像,留作后期观察。因此系统会进入语音交互状态,提示用户是否要保存当前图像,若用户选择保存图像,则系统进入保存状态,否则,返回上一状态。图像保存成功,则转到图像显示状态,不成功,则需要进入语音输入.根据语音识别的置信度,认为不成功的原因是语音录入不符合要求因此要用户重新进行语音交互状态。FIG. 2 shows a three-channel information integration strategy diagram in a navigation virtual microscope based on an intention understanding model proposed in
本发明提出的一种基于意图理解模型的导航式虚拟显微镜区别于传统的显微镜,不再采用光学元件来构造,而是使用多种传感器和通信模块,对已经设计好的带有标签的简单样本进行识别,这样做的优点是避免了学生将过多精力放到样本制作之上,也避免因样本制作不当,导致难以观察情况的发生,同时也解决了一些样本难以取材和无法重复利用的难题。本发明在传统显微镜功能基础之上,又增加了交互应用模块,其中交互应用模块通过视觉显示和听觉指引完成对用户意图进行预测以及给出操作引导。Different from the traditional microscope, the navigation virtual microscope based on the intention understanding model proposed by the present invention is no longer constructed with optical elements, but uses a variety of sensors and communication modules. The advantage of doing identification is that it avoids students from putting too much energy on sample production, and also avoids the occurrence of difficult observation due to improper sample production, and also solves the problem that some samples are difficult to obtain and cannot be reused. . On the basis of the traditional microscope function, the invention adds an interactive application module, wherein the interactive application module completes the prediction of the user's intention and gives the operation guidance through visual display and auditory guidance.
用户使用传统显微镜时,在上下左右移动样本的过程中,很有容易将要观察的样本细胞移出视野范围,同时,使用智能显微镜时,载玻片下的检测标记,也可能移出相机的视野范围。因此,对这种情况进行提示。例如:用户一直向左移动样本,视野内图像一直右移,当样本图像将要移出视野范围后,语音提示用户:“图像即将超出左侧区域,请向右移动”。When users use a traditional microscope, when moving the sample up, down, left and right, it is easy to move the sample cells to be observed out of the field of view. At the same time, when using a smart microscope, the detection mark under the slide may also move out of the camera's field of view. Therefore, a hint is given for this situation. For example, the user keeps moving the sample to the left, and the image in the field of view keeps moving to the right. When the sample image is about to move out of the field of view, the user is voiced: "The image is about to exceed the left area, please move to the right".
图像的保存是通过人机之间的对话实验的,用户在一定放大倍数下,转动粗准焦螺旋或者细准焦螺旋,调节视野内样本的清晰度,当视野内的样本清晰度达到设定的阈值,系统认为用户已经得到了想要观察的样本,因此,语音提示用户:“样本图像已经十分清晰,您可选择保存图像”。此时,用户可说“保存”、“保存这张图像”等,系统将智能的将当前视野内的图像保存到指定文件夹中。也可以说“不需要”或者继续调节粗细准焦螺旋,继续调整观察的样本。The image is saved through the dialogue experiment between man and machine. The user rotates the coarse focus screw or the fine focus screw under a certain magnification to adjust the clarity of the sample in the field of view. When the clarity of the sample in the field of view reaches the set point The system considers that the user has obtained the sample he wants to observe, so the user is prompted by voice: "The sample image is very clear, you can choose to save the image". At this point, the user can say "save", "save this image", etc., and the system will intelligently save the image in the current field of view to the specified folder. You can also say "no need" or continue to adjust the thickness of the focus screw, and continue to adjust the observed sample.
如图3给出了本发明实施例1提出的一种基于意图理解模型的导航式虚拟显微镜实现的结构示意图。智能显微镜上,用方形立方体替代反光镜和载物台并且对载玻片做了特殊处理。立方体内部安置了微型摄像头,摄像头正对通光孔,用来观察载玻片样本。载玻片的上表面是二维码,不同样本的二维码不同,下表面是带有标记的圆面。采用图像采集设备识别待观察样本,每一个载玻片的上面都带有表示该样本的二维码图片,根据图像识别不同的二维,进而从数据库中调取样本图像。初始化样本图像,可以观察到模糊的样本P。FIG. 3 is a schematic structural diagram of the realization of a navigation virtual microscope based on an intention understanding model proposed in
在操作过程时待检测样本发生位置移动,利用带有检测标记的圆面和位移变换过程的公式P’(X,Y)=PM(X-△X,Y-△Y),计算出样本位置移动,根据偏移量△X和△Y观察到样本图像P2。During the operation process, the position of the sample to be detected moves, and the sample position is calculated by using the circular surface with the detection mark and the formula P'(X,Y)=PM(X-△X,Y-△Y) of the displacement transformation process Moving, the sample image P 2 is observed according to the offset amounts ΔX and ΔY.
首先载玻片上表面的二维码朝向通光孔,相机识别到样本标签,显微镜下看到样本图片,之后在载玻片下表面盖在通光孔上,左右移动,相机根据检测到带有检测标记的圆面位置,沿相反方向控制显微镜下观察的样本。First, the QR code on the upper surface of the glass slide faces the light hole, the camera recognizes the sample label, and the sample picture is seen under the microscope. Detect the position of the circular surface of the marker and control the sample under the microscope in the opposite direction.
粗准焦螺旋或者细准焦螺旋,均由大小不一的两个同心圆柱组成,一端密封,另一端设计了可以遮光的转动轴,转动轴一端,内层圆柱内部均匀安装了6个光敏传感器,转动转轴,经处理后,将依次得到离散值t,s,作为用户转动刻度的度量。根据公式或者以及Coarse focusing spiral or fine focusing spiral are composed of two concentric cylinders of different sizes, one end is sealed, and the other end is designed with a rotating shaft that can block light. At one end of the rotating shaft, six photosensitive sensors are evenly installed inside the inner cylinder. , rotate the shaft, after processing, the discrete values t, s will be obtained in turn, as the measurement of the user's rotation scale. According to the formula or as well as
调节样本图像P3的清晰度。Adjust the sharpness of the sample image P3 .
如图4给出了本发明实施例1提出一种基于意图理解模型的导航式虚拟显微镜的目镜结构示意图,对显微镜的目镜,顶部是智能显示,样本调节的效果会实时显示在屏幕上,镜筒上设计了3个传感按钮,分别代表了5倍,10倍和40倍的放大倍数,选择不同的按钮,可观察到当前样本图像被放大。Figure 4 shows a schematic diagram of the eyepiece structure of a navigational virtual microscope based on an intention understanding model proposed in
以上内容仅仅是对本发明的结构所作的举例和说明,所属本技术领域的技术人员对所描述的具体实施例做各种各样的修改或补充或采用类似的方式替代,只要不偏离发明的结构或者超越本权利要求书所定义的范围,均应属于本发明的保护范围。The above content is only an example and description of the structure of the present invention. Those skilled in the art can make various modifications or supplements to the described specific embodiments or use similar methods to replace them, as long as they do not deviate from the structure of the invention. Or beyond the scope defined by the claims, all belong to the protection scope of the present invention.
Claims (7)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910543153.3A CN110196642B (en) | 2019-06-21 | 2019-06-21 | Navigation type virtual microscope based on intention understanding model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910543153.3A CN110196642B (en) | 2019-06-21 | 2019-06-21 | Navigation type virtual microscope based on intention understanding model |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110196642A CN110196642A (en) | 2019-09-03 |
CN110196642B true CN110196642B (en) | 2022-05-17 |
Family
ID=67755017
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910543153.3A Expired - Fee Related CN110196642B (en) | 2019-06-21 | 2019-06-21 | Navigation type virtual microscope based on intention understanding model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110196642B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111694476B (en) * | 2020-05-15 | 2022-07-08 | 平安科技(深圳)有限公司 | Translation browsing method and device, computer system and readable storage medium |
CN111999880A (en) * | 2020-09-15 | 2020-11-27 | 亚龙智能装备集团股份有限公司 | Virtual microscope system |
CN118411700B (en) * | 2024-06-27 | 2024-09-13 | 浙江恒逸石化有限公司 | Detection method, detection device, detection equipment and storage medium |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
AU2002332910A1 (en) * | 2001-09-07 | 2003-03-24 | Arizona Board Of Regents On Behalf Of The University Of Arizona | Multimodal miniature microscope |
CN102354349A (en) * | 2011-10-26 | 2012-02-15 | 华中师范大学 | Human-machine interaction multi-mode early intervention system for improving social interaction capacity of autistic children |
CN102405463A (en) * | 2009-04-30 | 2012-04-04 | 三星电子株式会社 | User intention inference apparatus and method using multi-modal information |
CN104965592A (en) * | 2015-07-08 | 2015-10-07 | 苏州思必驰信息科技有限公司 | Voice and gesture recognition based multimodal non-touch human-machine interaction method and system |
CN106997236A (en) * | 2016-01-25 | 2017-08-01 | 亮风台(上海)信息科技有限公司 | Based on the multi-modal method and apparatus for inputting and interacting |
CN107239139A (en) * | 2017-05-18 | 2017-10-10 | 刘国华 | Based on the man-machine interaction method and system faced |
CN109495724A (en) * | 2018-12-05 | 2019-03-19 | 济南大学 | A kind of virtual microscopic of view-based access control model perception and its application |
-
2019
- 2019-06-21 CN CN201910543153.3A patent/CN110196642B/en not_active Expired - Fee Related
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
AU2002332910A1 (en) * | 2001-09-07 | 2003-03-24 | Arizona Board Of Regents On Behalf Of The University Of Arizona | Multimodal miniature microscope |
CN102405463A (en) * | 2009-04-30 | 2012-04-04 | 三星电子株式会社 | User intention inference apparatus and method using multi-modal information |
CN102354349A (en) * | 2011-10-26 | 2012-02-15 | 华中师范大学 | Human-machine interaction multi-mode early intervention system for improving social interaction capacity of autistic children |
CN104965592A (en) * | 2015-07-08 | 2015-10-07 | 苏州思必驰信息科技有限公司 | Voice and gesture recognition based multimodal non-touch human-machine interaction method and system |
CN106997236A (en) * | 2016-01-25 | 2017-08-01 | 亮风台(上海)信息科技有限公司 | Based on the multi-modal method and apparatus for inputting and interacting |
CN107239139A (en) * | 2017-05-18 | 2017-10-10 | 刘国华 | Based on the man-machine interaction method and system faced |
CN109495724A (en) * | 2018-12-05 | 2019-03-19 | 济南大学 | A kind of virtual microscopic of view-based access control model perception and its application |
Also Published As
Publication number | Publication date |
---|---|
CN110196642A (en) | 2019-09-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110196642B (en) | Navigation type virtual microscope based on intention understanding model | |
CN114120432B (en) | Online learning attention tracking method based on gaze estimation and its application | |
US20210192181A1 (en) | Microscope system and projection unit | |
Coughlan et al. | AR4VI: AR as an accessibility tool for people with visual impairments | |
KR101563312B1 (en) | System for gaze-based providing education content | |
US11662565B2 (en) | Microscope system, projection unit, and image projection method | |
CN109634407B (en) | A control method based on synchronous acquisition and fusion of multi-mode human-machine sensing information | |
JP7209128B2 (en) | Gaze tracking system and method for smart glasses | |
US11112952B2 (en) | Interface for display of multi-layer images in digital microscopy | |
Chippendale et al. | Personal shopping assistance and navigator system for visually impaired people | |
Sankaran et al. | A survey report on the emerging technologies on assistive device for visually challenged people for analyzing traffic rules | |
US11994664B2 (en) | Augmented reality laser capture microdissection machine | |
US20170363854A1 (en) | Augmented reality visual rendering device | |
Wang et al. | A structural design and interaction algorithm of smart microscope embedded on virtual and real fusion technologies | |
US10043414B2 (en) | Microscope-based learning | |
Paswan et al. | Camera Based Indoor Object Detection and Distance Estimation Framework for Assistive Mobility | |
Divya et al. | Smart Object and Face Detection Assistant for Visually Impaired | |
US20240185534A1 (en) | Mixed reality workflow generation | |
EP3793192B1 (en) | Information processing device, information processing method, and program | |
Shaohan | Assistive System for Visual Impaired Experimenters Based on Object Detection and Apparatus Locating | |
Li | A human-machine interaction system for visual acuity test based on gesture recognition | |
Upadhyay et al. | eyeRoad–An App that Helps Visually Impaired Peoples | |
Varshini et al. | EchoVision: RealTime Object Detection and Voice Assistance for the Visually Impaired | |
Madhu et al. | Vision Assistance System for Visually Impaired | |
Zanjani et al. | The Usability of Augmented Reality Applications for Visually Impaired Individuals: A Systematic Review |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20220517 |