TW202311817A - Method, processing device, and display system for information display - Google Patents

Method, processing device, and display system for information display Download PDF

Info

Publication number
TW202311817A
TW202311817A TW111128679A TW111128679A TW202311817A TW 202311817 A TW202311817 A TW 202311817A TW 111128679 A TW111128679 A TW 111128679A TW 111128679 A TW111128679 A TW 111128679A TW 202311817 A TW202311817 A TW 202311817A
Authority
TW
Taiwan
Prior art keywords
display
information
user
target
virtual
Prior art date
Application number
TW111128679A
Other languages
Chinese (zh)
Other versions
TWI844902B (en
Inventor
羅翌瑋
陳健龍
鄭莛薰
趙玉如
林郁欣
Original Assignee
財團法人工業技術研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 財團法人工業技術研究院 filed Critical 財團法人工業技術研究院
Priority to CN202210997341.5A priority Critical patent/CN115794004A/en
Priority to US17/939,900 priority patent/US20230070721A1/en
Publication of TW202311817A publication Critical patent/TW202311817A/en
Application granted granted Critical
Publication of TWI844902B publication Critical patent/TWI844902B/en

Links

Images

Landscapes

  • Controls And Circuits For Display Device (AREA)

Abstract

A method, a processing device, and a system for information display are proposed, and the system includes a light transmissive display. A first information extraction device extracts spatial position information of a user, and a second information extraction device extracts spatial position information of a target object. The processing device performs the following steps. Display position information of virtual information of the target object on the display is determined according to the spatial position information of the user and the spatial position information of the target object. The display position information includes a first display reference position corresponding to a previous time and a second display position corresponding to a current time. An actual display position of the virtual information on the display corresponding to the current time is determined according to the distance between the first display reference position and the second display reference position. The virtual information is displayed on the display according to the actual display position.

Description

資訊顯示方法及其處理裝置與資訊顯示系統Information display method, processing device and information display system

本發明是有關於一種資訊顯示技術。The invention relates to an information display technology.

隨著影像處理技術與空間定位技術的發展,透明顯示器的應用已逐漸受到重視。此類的技術可讓顯示器搭配實體物件,再輔以虛擬相關資訊,並且依照使用者的需求來產生互動式的體驗,可使資訊以更為直觀的方式呈現。With the development of image processing technology and spatial positioning technology, the application of transparent displays has gradually attracted attention. This type of technology allows the display to be matched with physical objects, supplemented with virtual related information, and generates an interactive experience according to the user's needs, so that the information can be presented in a more intuitive way.

再者,關聯於實體物件的虛擬資訊可顯示於透明顯示器的特定位置上,讓使用者可透過透明顯示器同時觀看到實體物件與疊加於實體物件上或一側的虛擬資訊。然而,當實體物件動態移動時,可能發生虛擬資訊偏離實體物件的情況,進而導致透明顯示器所顯示的虛擬資訊無法即時地跟隨實體物件移動。此外,由於實體物件的移動不一定可被預測或可能是多變的,因此可能造成基於即時物件辨識而顯示的虛擬資訊出現晃動或堆疊的情況,導致觀看者閱讀虛擬資訊的困難度與不適。Furthermore, the virtual information associated with the physical object can be displayed on a specific position of the transparent display, so that the user can watch the physical object and the virtual information superimposed on or on one side of the physical object through the transparent display at the same time. However, when the physical object moves dynamically, the virtual information may deviate from the physical object, so that the virtual information displayed on the transparent display cannot follow the movement of the physical object in real time. In addition, since the movement of physical objects may not be predictable or may be changeable, it may cause shaking or stacking of virtual information displayed based on real-time object recognition, resulting in difficulty and discomfort for viewers to read virtual information.

本揭露提供一種資訊顯示方法及其處理裝置與資訊顯示系統。The disclosure provides an information display method, a processing device thereof, and an information display system.

在本揭露的一範例實施例中,上述的資訊顯示系統包括可透光的顯示器、第一資訊擷取裝置、第二資訊擷取裝置以及處理裝置,其中處理裝置連接於顯示器、第一資訊擷取裝置以及第二資訊擷取裝置。第一資訊擷取裝置用以擷取使用者的空間位置資訊,第二資訊擷取裝置用以擷取目標物的空間位置資訊。處理裝置經配置以執行下列步驟。根據使用者的空間位置資訊以及目標物的空間位置資訊決定目標物的虛擬資訊於顯示器上的顯示位置資訊。顯示位置資訊包括對應於先前時間的第一顯示參考位置與對應至當前時間的第二顯示參考位置。根據第一顯示參考位置與第二顯示參考位置之間的距離決定虛擬資訊於顯示器上對應於當前時間的實際顯示位置。根據實際顯示位置顯示虛擬資訊於顯示器上。In an exemplary embodiment of the present disclosure, the above-mentioned information display system includes a light-transmissible display, a first information capture device, a second information capture device, and a processing device, wherein the processing device is connected to the display, the first information capture device fetching device and the second information fetching device. The first information capturing device is used for capturing the user's spatial position information, and the second information capturing device is used for capturing the target object's spatial position information. The processing device is configured to perform the following steps. The display position information of the virtual information of the target on the display is determined according to the spatial position information of the user and the spatial position information of the target. The display location information includes a first display reference location corresponding to a previous time and a second display reference location corresponding to a current time. The actual display position of the virtual information on the display corresponding to the current time is determined according to the distance between the first display reference position and the second display reference position. Display virtual information on the monitor according to the actual display position.

在本揭露的一範例實施例中,上述的資訊顯示方法適用於具有可透光的顯示器、第一資訊擷取裝置、第二資訊擷取裝置以及處理裝置的資訊顯示系統,並且包括下列步驟。利用第一資訊擷取裝置擷取使用者的空間位置資訊。利用第二資訊擷取裝置擷取目標物的空間位置資訊。根據使用者的空間位置資訊以及目標物的空間位置資訊決定目標物的虛擬資訊於顯示器上的顯示位置資訊。顯示位置資訊包括對應於先前時間的第一顯示參考位置與對應至當前時間的第二顯示參考位置。根據第一顯示參考位置與第二顯示參考位置之間的距離決定虛擬資訊於顯示器上對應於當前時間的實際顯示位置。根據實際顯示位置顯示虛擬資訊於顯示器上。In an exemplary embodiment of the present disclosure, the above information display method is applicable to an information display system having a light-transmissive display, a first information capture device, a second information capture device, and a processing device, and includes the following steps. The spatial position information of the user is retrieved by using the first information retrieval device. The spatial position information of the target is captured by the second information capturing device. The display position information of the virtual information of the target on the display is determined according to the spatial position information of the user and the spatial position information of the target. The display location information includes a first display reference location corresponding to a previous time and a second display reference location corresponding to a current time. The actual display position of the virtual information on the display corresponding to the current time is determined according to the distance between the first display reference position and the second display reference position. Display virtual information on the monitor according to the actual display position.

在本揭露的一範例實施例中,上述的處理裝置連接於可透光的顯示器、第一資訊擷取裝置以及第二資訊擷取裝置。第一資訊擷取裝置用以擷取使用者的空間位置資訊,第二資訊擷取裝置用以擷取目標物的空間位置資訊。處理裝置包括記憶體與連接記憶體的處理器。記憶體用以儲存資料,而處理器經配置以執行下列步驟。利用第一資訊擷取裝置擷取使用者的空間位置資訊。利用第二資訊擷取裝置擷取目標物的空間位置資訊。根據使用者的空間位置資訊以及目標物的空間位置資訊決定目標物的虛擬資訊於顯示器上的顯示位置資訊。顯示位置資訊包括對應於先前時間的第一顯示參考位置與對應至當前時間的第二顯示參考位置。根據第一顯示參考位置與第二顯示參考位置之間的距離決定虛擬資訊於顯示器上對應於當前時間的實際顯示位置。根據實際顯示位置顯示虛擬資訊於顯示器上。In an exemplary embodiment of the present disclosure, the above-mentioned processing device is connected to a light-transmittable display, a first information capturing device, and a second information capturing device. The first information capturing device is used for capturing the user's spatial position information, and the second information capturing device is used for capturing the target object's spatial position information. The processing device includes a memory and a processor connected to the memory. The memory is used to store data and the processor is configured to perform the following steps. The spatial position information of the user is retrieved by using the first information retrieval device. The spatial position information of the target is captured by the second information capturing device. The display position information of the virtual information of the target on the display is determined according to the spatial position information of the user and the spatial position information of the target. The display location information includes a first display reference location corresponding to a previous time and a second display reference location corresponding to a current time. The actual display position of the virtual information on the display corresponding to the current time is determined according to the distance between the first display reference position and the second display reference position. Display virtual information on the monitor according to the actual display position.

為讓本揭露能更明顯易懂,下文特舉實施例,並配合所附圖式作詳細說明如下。In order to make the present disclosure more comprehensible, the following specific embodiments are described in detail together with the accompanying drawings.

本揭露的部份範例實施例接下來將會配合附圖來詳細描述,以下的描述所引用的元件符號,當不同附圖出現相同的元件符號將視為相同或相似的元件。這些範例實施例只是本揭露的一部份,並未揭示所有本揭露的可實施方式。更確切的說,這些範例實施例僅為本揭露的專利申請範圍中的方法、裝置以及系統的範例。Part of the exemplary embodiments of the present disclosure will be described in detail with reference to the accompanying drawings. For the referenced component symbols in the following description, when the same component symbols appear in different drawings, they will be regarded as the same or similar components. These exemplary embodiments are only part of the present disclosure, and do not disclose all possible implementations of the present disclosure. Rather, these exemplary embodiments are merely examples of methods, apparatuses, and systems within the scope of the present disclosure.

圖1A是根據本揭露一範例實施例所繪示的資訊顯示系統的方塊圖。首先圖1A先介紹系統中的各個構件以及配置關係,詳細功能將配合後續範例實施例的流程圖一併揭露。FIG. 1A is a block diagram of an information display system according to an exemplary embodiment of the present disclosure. Firstly, FIG. 1A firstly introduces various components and configuration relationships in the system, and the detailed functions will be disclosed together with the flow charts of subsequent exemplary embodiments.

請參照圖1A,本範例實施例中的資訊顯示系統100可包括顯示器110、第一資訊擷取裝置120、第二資訊擷取裝置130以及處理裝置140,其中處理裝置140可以是以無線、有線或電性連接於顯示器110、第一資訊擷取裝置120以及第二資訊擷取裝置130。Please refer to FIG. 1A, the information display system 100 in this exemplary embodiment may include a display 110, a first information capture device 120, a second information capture device 130, and a processing device 140, wherein the processing device 140 may be wireless, wired Or electrically connected to the display 110 , the first information capturing device 120 and the second information capturing device 130 .

顯示器110可用以顯示資訊,其包括例如是液晶顯示器(Liquid crystal display,LCD)、場色序(Field sequential color)液晶顯示器、發光二極體(Light emitting diode,LED)顯示器、電濕潤顯示器等穿透式可透光顯示器,或者是投影式可透光顯示器。The display 110 can be used to display information, which includes, for example, a liquid crystal display (Liquid crystal display, LCD), a field sequential color (Field sequential color) liquid crystal display, a light emitting diode (Light emitting diode, LED) display, an electrowetting display, etc. Transmissive light-transmissive display, or projection-type light-transmissive display.

第一資訊擷取裝置120用以擷取使用者的空間位置資訊,而第二資訊擷取裝置130用以擷取實體場景中目標物的空間位置資訊。第一資訊擷取裝置120可以例如是至少一個影像感測器或者是至少一個影像感測器搭配至少一個深度感測器來對使用者進行影像辨識定位,其中影像感測器可包括可見光感測器或非可見光感測器如紅外線感測器等。此外,第一資訊擷取裝置120更可以例如是光學定位器來對使用者進行光學空間定位。只要是可以定位出使用者所在位置資訊的裝置或其組合,皆屬於第一資訊擷取裝置120的範疇。The first information capturing device 120 is used for capturing the spatial position information of the user, and the second information capturing device 130 is used for capturing the spatial position information of the object in the physical scene. The first information capture device 120 may be, for example, at least one image sensor or at least one image sensor combined with at least one depth sensor to identify and locate the user's image, wherein the image sensor may include a visible light sensor. Sensors or non-visible light sensors such as infrared sensors, etc. In addition, the first information capturing device 120 can be, for example, an optical locator to perform optical spatial positioning of the user. As long as it is a device capable of locating the user's location information or a combination thereof, it all belongs to the category of the first information capturing device 120 .

第二資訊擷取裝置120則可以例如是至少一個影像感測器或者是至少一個影像感測器搭配至少一個深度感測器來對目標物進行影像辨識定位,其中影像感測器可包括可見光感測器或非可見光感測器如紅外線感測器等。只要是可以定位出目標物所在位置資訊的裝置或其組合,皆屬於第二資訊擷取裝置130的範疇。The second information capture device 120 can be, for example, at least one image sensor or at least one image sensor combined with at least one depth sensor to perform image recognition and positioning of the target object, wherein the image sensor can include a visible light sensor. detector or non-visible light sensor such as infrared sensor, etc. As long as it is a device capable of locating the position information of the target object or a combination thereof, it all belongs to the category of the second information capturing device 130 .

於本揭露實施例中,上述的影像感測器可用以擷取影像並且包括具有透鏡以及感光元件的攝像鏡頭。上述的深度感測器可用以偵測深度資訊,其可以利用主動式深度感測技術以及被動式深度感測技術來實現。主動式深度感測技術可藉由主動發出光源、紅外線、超音波、雷射等作為訊號搭配時差測距技術來計算深度資訊。被動式深度感測技術可以藉由兩個影像感測器以不同視角擷取其前方的兩張影像,以利用兩張影像的視差來計算深度資訊。In the embodiment of the present disclosure, the above-mentioned image sensor can be used to capture images and includes a camera lens having a lens and a photosensitive element. The above-mentioned depth sensor can be used to detect depth information, which can be realized by using active depth sensing technology and passive depth sensing technology. Active depth sensing technology can calculate depth information by actively emitting light sources, infrared rays, ultrasonic waves, lasers, etc. as signals with time-of-flight ranging technology. The passive depth sensing technology can use two image sensors to capture two images in front of it from different angles of view, and use the parallax of the two images to calculate depth information.

處理裝置140用以控制資訊顯示系統100的作動,其可包括記憶體141以及處理器142。記憶體141可以例如是任意型式的固定式或可移動式隨機存取記憶體(random access memory,RAM)、唯讀記憶體(read-only memory,ROM)、快閃記憶體(flash memory)、硬碟或其他類似裝置、積體電路及其組合。處理器142可以例如是中央處理單元(central processing unit,CPU)、應用處理器(application processor,AP),或是其他可程式化之一般用途或特殊用途的微處理器(microprocessor)、數位訊號處理器(digital signal processor,DSP)、影像訊號處理器(image signal processor,ISP)、圖形處理器(graphics processing unit,GPU)或其他類似裝置、積體電路及其組合。以下範例實施例中將搭配資訊顯示系統100的各元件來說明處理裝置140執行資訊顯示方法流程的細節。The processing device 140 is used to control the operation of the information display system 100 , and may include a memory 141 and a processor 142 . The memory 141 can be, for example, any type of fixed or removable random access memory (random access memory, RAM), read-only memory (read-only memory, ROM), flash memory (flash memory), Hard disk or other similar devices, integrated circuits and combinations thereof. The processor 142 can be, for example, a central processing unit (central processing unit, CPU), an application processor (application processor, AP), or other programmable general purpose or special purpose microprocessor (microprocessor), digital signal processing Digital signal processor (DSP), image signal processor (image signal processor, ISP), graphics processing unit (graphics processing unit, GPU) or other similar devices, integrated circuits and combinations thereof. In the following exemplary embodiment, each component of the information display system 100 will be used to illustrate the details of the process of the information display method performed by the processing device 140 .

在本範例實施例中,處理裝置140可以是內建於顯示器110或連接顯示器110的計算機裝置。第一資訊擷取裝置120以及第二資訊擷取裝置130可以是分別設置於資訊顯示系統100所屬場域相對於顯示器110的相對兩側等,用以對使用者以及目標物進行定位,並且透過各自的通訊介面以有線或是無線的方式傳輸資訊至處理裝置140。於一些實施例中,第一資訊擷取裝置120與第二資訊擷取裝置130也可各自具有處理器與記憶體,並具有可根據影像資料進行物件辨識與物件追蹤的計算能力。在另一範例實施例中,資訊顯示系統100可為單一整合系統,其可以實作成頭戴式顯示裝置、智慧型手機、平板電腦等,本揭露不在此設限。以下範例實施例中將搭配資訊顯示系統100的各元件來說明資訊顯示方法的實施細節。In this exemplary embodiment, the processing device 140 may be a computer device built into the display 110 or connected to the display 110 . The first information capture device 120 and the second information capture device 130 can be respectively installed on the opposite sides of the field where the information display system 100 belongs to the display 110, etc., to locate the user and the target, and through The respective communication interfaces transmit information to the processing device 140 in a wired or wireless manner. In some embodiments, the first information capturing device 120 and the second information capturing device 130 may each have a processor and a memory, and have computing capabilities for object recognition and object tracking based on image data. In another exemplary embodiment, the information display system 100 can be a single integrated system, which can be implemented as a head-mounted display device, a smart phone, a tablet computer, etc., and the disclosure is not limited here. In the following exemplary embodiments, the implementation details of the information display method will be illustrated with each component of the information display system 100 .

圖1B是根據本揭露一範例實施例所繪示的資訊顯示系統的示意圖。請參照圖1B,使用者U1與目標物Obj1分別位於顯示器110的不同側。使用者U1可透過顯示器110觀看疊加有目標物Obj1的虛擬資訊Vf1的實體場景。虛擬資訊Vf1可視為基於目標物Obj1而擴增的擴增實境內容。於一些實施例中,第一資訊擷取裝置120可包括用以拍攝使用者U1的第一影像感測器,且第二資訊擷取裝置130包括用以拍攝目標物Obj1的第二影像感測器。第一資訊擷取裝置120可基於所擷取的使用者影像獲取使用者U1的空間位置資訊,而第二資訊擷取裝置130可基於所擷取的目標物影像獲取目標物Obj1的空間位置資訊。FIG. 1B is a schematic diagram of an information display system according to an exemplary embodiment of the present disclosure. Referring to FIG. 1B , the user U1 and the object Obj1 are located on different sides of the display 110 . The user U1 can watch the physical scene superimposed with the virtual information Vf1 of the object Obj1 through the display 110 . The virtual information Vf1 can be regarded as augmented reality content amplified based on the object Obj1. In some embodiments, the first information capture device 120 may include a first image sensor for photographing the user U1, and the second information capture device 130 may include a second image sensor for photographing the target Obj1 device. The first information capturing device 120 can obtain the spatial position information of the user U1 based on the captured user image, and the second information capturing device 130 can obtain the spatial position information of the object Obj1 based on the captured object image. .

處理裝置140可根據使用者U1的空間位置資訊以及目標物Obj1的空間位置資訊來決定虛擬資訊Vf1於顯示器110上的顯示位置資訊。於圖1B的範例中,上述顯示位置資訊實施為顯示平面上的一個參考顯示物件框RF1。於其他實施例中,顯示位置資訊可實施為顯示器110的顯示平面上的一個點。具體來說,顯示位置資訊可視為使用者觀看目標物Obj1時視線投射於顯示平面上的落點或區域。藉此,處理裝置140可利用顯示位置資訊(例如參考顯示物件框RF1)為參考基準來顯示虛擬資訊Vf1。更具體而言,基於各式需求或不同應用,處理裝置140可根據顯示位置資訊決定虛擬資訊Vf1的實際顯示位置,以讓使用者U1可透過顯示器110看到位於目標物Obj1附近的虛擬資訊Vf1或看到疊加於目標物Obj1上的虛擬資訊Vf1。The processing device 140 can determine the display position information of the virtual information Vf1 on the display 110 according to the spatial position information of the user U1 and the spatial position information of the object Obj1. In the example shown in FIG. 1B , the above display location information is implemented as a reference display object frame RF1 on the display plane. In other embodiments, the display location information can be implemented as a point on the display plane of the display 110 . Specifically, the display position information can be regarded as the point or area where the user's line of sight is projected on the display plane when viewing the object Obj1. Thereby, the processing device 140 can display the virtual information Vf1 by using the display position information (for example, referring to the display object frame RF1 ) as a reference. More specifically, based on various needs or different applications, the processing device 140 can determine the actual display position of the virtual information Vf1 according to the display position information, so that the user U1 can see the virtual information Vf1 near the object Obj1 through the display 110 Or see the virtual information Vf1 superimposed on the object Obj1.

值得一提的是,於一些實施例中,使用者U1及/或目標物Obj1都可能會動態移動。舉例而言,目標物Obj1可以是飼養於水族箱裡的魚隻,而虛擬資訊Vf1可以是魚隻的種類名稱或介紹資訊。於是,第一資訊擷取裝置120可對著使用者U1持續進行拍攝而產生包括分別對應至不同時間點的多張使用者影像的視頻序列,第一資訊擷取裝置120可根據包括多張使用者影像的視頻序列來追蹤使用者U1的空間位置資訊。同理,第二資訊擷取裝置130可對著目標物Obj1持續進行拍攝而產生包括分別對應至不同時間點的多張目標物影像的視頻序列,第二資訊擷取裝置130可根據包括多張目標物影像的視頻序列來追蹤目標物Obj1的空間位置資訊。由於處理裝置140可基於目標物Obj1與使用者U1的追蹤結果不斷地更新虛擬資訊Vf1的顯示位置資訊,因此顯示器110所顯示的虛擬資訊Vf1可反應於使用者U1的移動與/或目標物Obj1的移動而對應移動,以達到虛擬資訊Vf1跟隨目標物Obj1的視覺效果。It is worth mentioning that, in some embodiments, both the user U1 and/or the object Obj1 may move dynamically. For example, the target object Obj1 can be a fish kept in an aquarium, and the virtual information Vf1 can be the species name or introduction information of the fish. Therefore, the first information capture device 120 can continuously shoot the user U1 to generate a video sequence including a plurality of user images corresponding to different time points, and the first information capture device 120 can use the The spatial location information of the user U1 is tracked by a video sequence of the user's images. Similarly, the second information capture device 130 can continuously shoot the target object Obj1 to generate a video sequence including a plurality of target object images respectively corresponding to different time points. The video sequence of the object image is used to track the spatial position information of the object Obj1. Since the processing device 140 can continuously update the display position information of the virtual information Vf1 based on the tracking results of the object Obj1 and the user U1, the virtual information Vf1 displayed on the display 110 can reflect the movement of the user U1 and/or the object Obj1. The corresponding movement of the virtual information Vf1 to achieve the visual effect of following the target Obj1.

圖2是根據本揭露一範例實施例所繪示的資訊顯示方法的流程圖,請同時參照圖1A、圖1B以及圖2,而圖2的方法流程可由圖1A與圖1B的資訊顯示系統100來實現。在此,使用者U1可透過資訊顯示系統100的顯示器110來觀看目標物Obj1及其虛擬資訊Vf1。FIG. 2 is a flowchart of an information display method according to an exemplary embodiment of the present disclosure. Please refer to FIG. 1A, FIG. 1B and FIG. 2 at the same time, and the method flow in FIG. to fulfill. Here, the user U1 can view the target object Obj1 and its virtual information Vf1 through the display 110 of the information display system 100 .

於步驟S210,資訊顯示系統100中利用第一資訊擷取裝置120擷取使用者U1的空間位置資訊。於步驟S220,資訊顯示系統100中利用第二資訊擷取裝置130擷取目標物Obj1的空間位置資訊。如同前述,第一資訊擷取裝置120以及第二資訊擷取裝置130例如是可針對使用者U1以及目標物Obj1的所在位置進行定位的影像感測器、深度感測器或其組合。In step S210, the information display system 100 utilizes the first information capturing device 120 to capture the spatial location information of the user U1. In step S220 , the information display system 100 uses the second information capturing device 130 to capture the spatial position information of the object Obj1 . As mentioned above, the first information capturing device 120 and the second information capturing device 130 are, for example, an image sensor, a depth sensor or a combination thereof that can locate the positions of the user U1 and the object Obj1.

於步驟S230,資訊顯示系統100的處理裝置140根據使用者U1的空間位置資訊以及目標物Obj1的空間位置資訊決定目標物Obj1的虛擬資訊Vf1於顯示器110上的顯示位置資訊。顯示位置資訊包括對應於先前時間的第一顯示參考位置與對應至當前時間的第二顯示參考位置。具體而言,處理裝置140可根據使用者U1的空間位置資訊以及目標物Obj1的空間位置資訊持續地計算出分別對應至多個連續時間的多個顯示參考位置。換言之,第一資訊擷取裝置120與第二資訊擷取裝置130可持續更新目標物Obj1與使用者U1的空間位置資訊,而處理裝置140可據以持續更新虛擬資訊Vf1的顯示位置資訊。In step S230 , the processing device 140 of the information display system 100 determines the display position information of the virtual information Vf1 of the object Obj1 on the display 110 according to the spatial position information of the user U1 and the spatial position information of the object Obj1 . The display location information includes a first display reference location corresponding to a previous time and a second display reference location corresponding to a current time. Specifically, the processing device 140 can continuously calculate a plurality of display reference positions respectively corresponding to a plurality of consecutive times according to the spatial position information of the user U1 and the spatial position information of the object Obj1 . In other words, the first information capturing device 120 and the second information capturing device 130 can continuously update the spatial position information of the object Obj1 and the user U1, and the processing device 140 can continuously update the display position information of the virtual information Vf1 accordingly.

於步驟S240,資訊顯示系統100的處理裝置140根據第一顯示參考位置與第二顯示參考位置之間的距離決定虛擬資訊Vf1於顯示器110上對應於當前時間的實際顯示位置。接著,於步驟S250,資訊顯示系統100的處理裝置140根據實際顯示位置顯示虛擬資訊Vf1於顯示器110上。In step S240, the processing device 140 of the information display system 100 determines the actual display position of the virtual information Vf1 corresponding to the current time on the display 110 according to the distance between the first display reference position and the second display reference position. Next, in step S250 , the processing device 140 of the information display system 100 displays the virtual information Vf1 on the display 110 according to the actual display position.

也就是說,於處理裝置140決定虛擬資訊Vf1對應於當前時間的實際顯示位置的過程中,處理裝置140會計算先前時間的第一顯示參考位置與當前時間的第二顯示參考位置之間的距離。接著,處理裝置140再根據上述距離來決定利用當前時間的第二顯示參考位置更新虛擬資訊Vf1於顯示器110上的實際顯示位置,或決定不要更新虛擬資訊Vf1的實際顯示位置。That is to say, when the processing device 140 determines that the virtual information Vf1 corresponds to the actual display position at the current time, the processing device 140 will calculate the distance between the first display reference position at the previous time and the second display reference position at the current time . Next, the processing device 140 decides to use the second display reference position at the current time to update the actual display position of the virtual information Vf1 on the display 110 according to the above distance, or decides not to update the actual display position of the virtual information Vf1.

圖3是根據本揭露一範例實施例所繪示的決定對應於當前時間的實際顯示位置的流程圖。請參照圖3,圖2的步驟S240可實施為步驟S241~步驟S243。於步驟S241,處理裝置140判斷對應於先前時間的第一顯示參考位置與對應於當前時間的第二顯示參考位置之間的距離是否大於預設臨界值。FIG. 3 is a flow chart of determining an actual display position corresponding to the current time according to an exemplary embodiment of the present disclosure. Referring to FIG. 3 , step S240 in FIG. 2 may be implemented as steps S241 to S243 . In step S241 , the processing device 140 determines whether the distance between the first display reference position corresponding to the previous time and the second display reference position corresponding to the current time is greater than a preset threshold.

若第一顯示參考位置與第二顯示參考位置之間的距離大於預設臨界值,代表使用者U1所看到的目標物Obj1的位移是明顯的。因此,反應於第一顯示參考位置與第二顯示參考位置之間的距離大於預設臨界值(步驟S241判斷為是),於步驟S242,處理裝置140根據對應至當前時間的第二顯示參考位置更新虛擬資訊Vf1於顯示器110上對應於當前時間的實際顯示位置。If the distance between the first displayed reference position and the second displayed reference position is greater than a preset threshold, it means that the displacement of the object Obj1 seen by the user U1 is obvious. Therefore, in response to the fact that the distance between the first display reference position and the second display reference position is greater than the preset threshold (step S241 judges yes), in step S242, the processing device 140 uses the second display reference position corresponding to the current time Updating the actual display position of the virtual information Vf1 corresponding to the current time on the display 110 .

另一方面,若第一顯示參考位置與第二顯示參考位置之間的距離未大於預設臨界值,代表使用者U1所看到的目標物Obj1的位移較小。因此,反應於第一顯示參考位置與第二顯示參考位置之間的距離未大於預設臨界值(步驟S241判斷為否),於步驟S243,處理裝置140不更新虛擬資訊Vf1於顯示器110上對應於當前時間的實際顯示位置,亦即對應至當前時間的實際顯示位置會相同於對應至先前時間的實際顯示位置。於一實施例中,虛擬資訊Vf1可根據基於先前時間的第一顯示參考位置所決定的實際顯示位置來進行顯示。也就是說,若第一顯示參考位置與第二顯示參考位置之間的距離未大於預設臨界值,虛擬資訊Vf1於顯示器110上的顯示位置可以不更動。藉此,虛擬資訊Vf1的實際顯示位置不會因為目標物Obj1輕微的晃動或移動而變化,大幅提昇使用者U1觀看虛擬資訊Vf1的舒適度。意即,虛擬資訊Vf1可平穩地顯示於顯示器110上,而減少晃動或殘影堆疊的情況發生。On the other hand, if the distance between the first displayed reference position and the second displayed reference position is not greater than the preset threshold, it means that the displacement of the object Obj1 seen by the user U1 is relatively small. Therefore, in response to the fact that the distance between the first display reference position and the second display reference position is not greater than the preset threshold (step S241 judged as NO), in step S243, the processing device 140 does not update the virtual information Vf1 on the display 110 to correspond to The actual display position at the current time, that is, the actual display position corresponding to the current time will be the same as the actual display position corresponding to the previous time. In one embodiment, the virtual information Vf1 can be displayed according to the actual display position determined based on the first display reference position at a previous time. That is to say, if the distance between the first display reference position and the second display reference position is not greater than the preset threshold, the display position of the virtual information Vf1 on the display 110 may not be changed. In this way, the actual display position of the virtual information Vf1 will not change due to the slight shaking or movement of the object Obj1, which greatly improves the comfort of the user U1 viewing the virtual information Vf1. That is to say, the virtual information Vf1 can be displayed on the display 110 stably without shaking or stacking of afterimages.

為了方便明瞭,以下將搭配顯示系統100列舉實施例以說明本揭露計算顯示位置資訊的實施方式。For the sake of convenience and clarity, the following examples will be combined with the display system 100 to illustrate the implementation of calculating and displaying position information in the present disclosure.

圖4是根據本揭露一範例實施例所繪示的資訊顯示方法的流程圖。請參照圖4,於步驟S402,第一資訊擷取裝置120擷取使用者影像,其中使用者個數可為一個以上。於步驟S404,第一資訊擷取裝置120對多張使用者影像進行人臉偵測而獲取各使用者影像中的人臉位置。第一資訊擷取裝置120可透過各種人臉偵測技術與人臉特徵偵測技術從各張使用者影像中偵測出人臉區塊以及人臉區塊中的臉部特徵。人臉偵測技術例如是利用像素值分析的傳統人臉偵測演算法或應用機器學習的深度學習演算法等等,本揭露對此不限制。上述臉部特徵例如是雙眼。於一些實施例中,第一資訊擷取裝置120可將人臉區塊的中心點作為人臉位置或將雙眼中心點作為人臉位置,但本揭露對此不限制。FIG. 4 is a flowchart of an information display method according to an exemplary embodiment of the present disclosure. Please refer to FIG. 4 , in step S402 , the first information capture device 120 captures images of users, where the number of users may be more than one. In step S404, the first information capture device 120 performs face detection on multiple user images to obtain the face positions in each user image. The first information capturing device 120 can detect face blocks and facial features in the face blocks from each user image through various face detection technologies and face feature detection technologies. The face detection technology is, for example, a traditional face detection algorithm using pixel value analysis or a deep learning algorithm using machine learning, etc., which is not limited in this disclosure. The aforementioned facial features are, for example, both eyes. In some embodiments, the first information capturing device 120 may use the center point of the face block as the face position or the center point of both eyes as the face position, but this disclosure is not limited thereto.

於步驟S406,第一資訊擷取裝置120根據人臉位置進行座標轉換而獲取使用者U1的空間位置資訊。具體來說,基於使用者影像獲取的人臉位置是屬於二維的畫素座標。第一資訊擷取裝置120可根據使用者U1的深度資訊與使用者影像中的人臉位置進行座標轉換,以產生使用者U1的空間位置資訊。於圖4的實施例中,使用者U1的空間位置資訊可包括相機座標系下的使用者相機座標。In step S406, the first information capturing device 120 performs coordinate conversion according to the position of the face to obtain the spatial position information of the user U1. Specifically, the face position acquired based on the user image is a two-dimensional pixel coordinate. The first information capturing device 120 can perform coordinate conversion according to the depth information of the user U1 and the face position in the user's image, so as to generate the spatial position information of the user U1. In the embodiment of FIG. 4 , the spatial position information of the user U1 may include the user's camera coordinates in the camera coordinate system.

接著,於步驟S408,處理裝置140根據第一資訊擷取裝置120的第一影像感測器的拍攝方向將使用者相機座標轉換為相對於顯示器110的使用者位置座標。於一實施例中,第一影像感測器的拍攝方向可包括垂直方向上的拍攝俯仰角。舉例而言,圖5A是根據本揭露一範例實施例所繪示的第一影像感測器與第二影像感測器的拍攝方向的示意圖。請參照圖5A,用以拍攝使用者U1的第一影像感測器具有拍攝俯仰角θ1。拍攝俯仰角θ1為第一影像感測器之鏡頭光軸CX1與顯示器110之顯示平面法線HX1之間的夾角。於一實施例中,假設使用者相機座標為(cx 1, cy 1, cz 1),則處理裝置140可根據下列公式(1)將使用者相機座標(cx 1, cy 1, cz 1)轉換為相對於顯示器110的使用者位置座標(t x1, t y1, t z1)。

Figure 02_image001
公式(1) 於一實施例中,相對於顯示器110的使用者位置座標(t x1, t y1, t z1)所屬的參考座標系的三座標軸分別為顯示器110的顯示X軸、顯示Y軸以及顯示平面法線。拍攝俯仰角θ1為顯示器110的顯示平面法線與第一影像感測器之鏡頭光軸CX1之間的夾角。 Next, in step S408 , the processing device 140 converts the user camera coordinates into user position coordinates relative to the display 110 according to the shooting direction of the first image sensor of the first information capturing device 120 . In one embodiment, the shooting direction of the first image sensor may include a shooting elevation angle in the vertical direction. For example, FIG. 5A is a schematic diagram illustrating shooting directions of the first image sensor and the second image sensor according to an exemplary embodiment of the present disclosure. Referring to FIG. 5A , the first image sensor for photographing the user U1 has a photographing elevation angle θ1 . The shooting pitch angle θ1 is the angle between the lens optical axis CX1 of the first image sensor and the display plane normal HX1 of the display 110 . In one embodiment, assuming that the user camera coordinates are (cx 1 , cy 1 , cz 1 ), the processing device 140 can transform the user camera coordinates (cx 1 , cy 1 , cz 1 ) according to the following formula (1): are the user position coordinates (t x1 , ty1 , t z1 ) relative to the display 110 .
Figure 02_image001
Formula (1) In one embodiment, the three coordinate axes of the reference coordinate system to which the user position coordinates (t x1 , ty1 , t z1 ) of the display 110 belong are respectively the display X axis, the display Y axis and the display 110 of the display 110. Displays plane normals. The shooting pitch angle θ1 is the angle between the normal of the display plane of the display 110 and the lens optical axis CX1 of the first image sensor.

另一方面,於步驟S410,第二資訊擷取裝置130擷取目標物影像,其中目標物個數可為一個以上。於步驟S412,第二資訊擷取裝置130對多張目標物影像進行物件偵測而獲取目標物於各目標物影像中的物件邊界框。第二資訊擷取裝置130可利用深度學習演算法中的卷積神經網路(Convolution Neural Network,CNN)模型來進行物件偵測,上述卷積神經網路模型例如是用以進行物件偵測的R-CNN、Fast R-CNN、Faster R-CNN、YOLO或SSD等等,但本揭露對此不限制。第二資訊擷取裝置130所使用的物件偵測模型可輸出目標物Obj1的物件邊界框(Bonding box)的位置與尺寸以及目標物Obj1的分類結果。於一實施例中,處理裝置140可根據目標物Obj1的分類結果而從資料庫挑選對應的虛擬資訊Vf1進行顯示。On the other hand, in step S410, the second information capturing device 130 captures the image of the target object, wherein the number of the target object may be more than one. In step S412 , the second information capturing device 130 performs object detection on multiple target images to obtain object bounding boxes of the target in each target image. The second information capturing device 130 can utilize a convolutional neural network (Convolution Neural Network, CNN) model in a deep learning algorithm to perform object detection. The above-mentioned convolutional neural network model is for example used for object detection R-CNN, Fast R-CNN, Faster R-CNN, YOLO or SSD, etc., but this disclosure is not limited thereto. The object detection model used by the second information acquisition device 130 can output the position and size of the object's bounding box (Bonding box) of the object Obj1 and the classification result of the object Obj1 . In one embodiment, the processing device 140 can select the corresponding virtual information Vf1 from the database according to the classification result of the object Obj1 for display.

於步驟S414,第二資訊擷取裝置130根據物件邊界框進行座標轉換而獲取目標物Obj1的空間位置資訊。具體來說,基於目標物影像獲取的物件邊界框的位置屬於二維的畫素座標,上述物件邊界框的基準點可以是物件邊界框的頂點位置、中心點位置或框邊界,本揭露對此不限制。第二資訊擷取裝置130可根據目標物Obj1的深度資訊與目標物影像中的物件邊界框進行座標轉換,以產生目標物Obj1的空間位置資訊。於圖4所示的實施例中,目標物Obj1的空間位置資訊包括相機座標系下的目標物相機座標。In step S414, the second information capturing device 130 performs coordinate conversion according to the object bounding box to obtain the spatial position information of the object Obj1. Specifically, the position of the object bounding box acquired based on the target image belongs to two-dimensional pixel coordinates, and the reference point of the above object bounding box can be the vertex position, center point position or frame boundary of the object bounding box. not limited. The second information capturing device 130 can perform coordinate conversion according to the depth information of the object Obj1 and the object bounding box in the image of the object, so as to generate the spatial position information of the object Obj1. In the embodiment shown in FIG. 4 , the spatial position information of the object Obj1 includes the camera coordinates of the object in the camera coordinate system.

接著,於步驟S416,處理裝置140根據第二影像感測器的拍攝方向將目標物相機座標轉換為相對於顯示器110的目標物位置座標。於一實施例中,第二影像感測器的拍攝方向可包括垂直方向上的拍攝俯仰角。舉例而言,請再參照圖5A,用以拍攝目標物Obj1的第二影像感測器具有拍攝俯仰角θ2。拍攝俯仰角θ2為第二影像感測器之鏡頭光軸CX2與顯示平面法線HX1之間的夾角。假設目標物相機座標為(cx 2, cy 2, cz 2),則處理裝置140可根據下列公式(2)將目標物相機座標(cx 2, cy 2, cz 2)轉換為相對於顯示器110的目標物位置座標(t x2, t y2, t z2)。

Figure 02_image003
公式(2) 於一實施例中,相對於顯示器110的目標物位置座標(t x2, t y2, t z2)所屬的參考座標系相同於使用者位置座標(t x1, t y1, t z1)所屬的參考座標系。拍攝俯仰角θ2為顯示器110的顯示平面法線與第二影像感測器之鏡頭光軸CX2之間的夾角。 Next, in step S416 , the processing device 140 converts the object camera coordinates into object position coordinates relative to the display 110 according to the shooting direction of the second image sensor. In one embodiment, the shooting direction of the second image sensor may include a shooting pitch angle in the vertical direction. For example, please refer to FIG. 5A again, the second image sensor for photographing the object Obj1 has a photographing pitch angle θ2. The shooting pitch angle θ2 is the angle between the optical axis CX2 of the lens of the second image sensor and the normal line HX1 of the display plane. Assuming that the object camera coordinates are (cx 2 , cy 2 , cz 2 ), the processing device 140 can convert the object camera coordinates (cx 2 , cy 2 , cz 2 ) into relative to the display 110 according to the following formula (2): Target position coordinates (t x2 , ty2 , t z2 ).
Figure 02_image003
Equation (2) In one embodiment, the reference coordinate system to which the target position coordinates (t x2 , ty2 , t z2 ) relative to the display 110 belongs is the same as the user position coordinates (t x1 , ty1 , t z1 ) The reference coordinate system to which it belongs. The shooting pitch angle θ2 is the angle between the normal of the display plane of the display 110 and the optical axis CX2 of the lens of the second image sensor.

於一些實施例中,第一影像感測器與第二影像感測器的拍攝方向可透過設置於第一影像感測器與第二影像感測器上的慣性感測器而獲取。以圖5A的範例來說,慣性感測器的感測值可用以計算出圖5A中的拍攝俯仰角θ1、θ2。舉例而言,上述慣性感測器例如是加速度感測器等等。In some embodiments, the shooting directions of the first image sensor and the second image sensor can be obtained through inertial sensors disposed on the first image sensor and the second image sensor. Taking the example of FIG. 5A as an example, the sensing value of the inertial sensor can be used to calculate the shooting pitch angles θ1 and θ2 in FIG. 5A . For example, the above-mentioned inertial sensor is, for example, an acceleration sensor or the like.

於步驟S418,處理裝置140根據目標物位置座標與使用者位置座標之間的連線決定虛擬資訊Vf1於顯示器110上的顯示位置資訊。圖5B是根據本揭露一範例實施例所繪示的決定顯示位置資訊的示意圖。請參照圖5B,顯示位置資訊可包括目標物位置座標(t x2, t y2, t z2)與使用者位置座標(t x1, t y1, t z1)之間的連線相交於顯示器110的顯示平面的交點位置IP1。根據公式(1)與公式(2)所計算的使用者位置座標(t x1, t y1, t z1)與目標物位置座標(t x2, t y2, t z2),位於顯示平面(即Z=0)上的交點位置IP1的空間座標(X, Y, Z)可根據下列公式(3)而計算出來。

Figure 02_image005
公式(3) 之後,處理裝置140可根據顯示器110的尺寸與解析度獲取顯示位置資訊,而顯示位置資訊為顯示器110的畫素座標。更詳細來說,處理裝置140可根據顯示器110的尺寸與解析度將交點位置IP1的空間座標轉換為顯示器110上的畫素座標,以獲取虛擬資訊Vf1於顯示器110上的顯示位置資訊,亦即由畫素座標來表示的交點位置IP1。 In step S418 , the processing device 140 determines the display position information of the virtual information Vf1 on the display 110 according to the connection line between the target position coordinates and the user position coordinates. FIG. 5B is a schematic diagram of determining to display location information according to an exemplary embodiment of the present disclosure. Please refer to FIG. 5B , the displayed position information may include the display on the display 110 where the connection line between the target position coordinates (t x2 , ty2 , t z2 ) and the user position coordinates (t x1 , t y1 , t z1 ) intersects. The intersection point of the plane is IP1. The user position coordinates (t x1 , t y1 , t z1 ) and target object position coordinates (t x2 , t y2 , t z2 ) calculated according to formula (1) and formula (2) are located on the display plane (ie Z= The spatial coordinates (X, Y, Z) of the intersection point IP1 on 0) can be calculated according to the following formula (3).
Figure 02_image005
After the formula (3), the processing device 140 can obtain display position information according to the size and resolution of the display 110 , and the display position information is the pixel coordinates of the display 110 . More specifically, the processing device 140 can convert the spatial coordinates of the intersection point IP1 into pixel coordinates on the display 110 according to the size and resolution of the display 110, so as to obtain the display position information of the virtual information Vf1 on the display 110, that is, The intersection position IP1 represented by pixel coordinates.

如同前述,顯示位置資訊包括對應至先前時間的第一顯示參考位置與對應至當前時間的第二顯示參考位置。於步驟S420,處理裝置140根據第一顯示參考位置與第二顯示參考位置之間的距離決定虛擬資訊Vf1於顯示器110上對應於當前時間的實際顯示位置。於步驟S422,處理裝置140根據實際顯示位置顯示虛擬資訊Vf1於顯示器110上。As mentioned above, the display location information includes a first display reference location corresponding to a previous time and a second display reference location corresponding to a current time. In step S420 , the processing device 140 determines the actual display position of the virtual information Vf1 corresponding to the current time on the display 110 according to the distance between the first display reference position and the second display reference position. In step S422, the processing device 140 displays the virtual information Vf1 on the display 110 according to the actual display position.

詳細而言,圖6A是根據本揭露一範例實施例所繪示的不更新實際顯示位置的示意圖。圖6B是根據本揭露一範例實施例所繪示的更新實際顯示位置的示意圖。請參照圖6A與6B,於本實施範例中,顯示位置資訊包括對應至先前時間的第一顯示參考位置IP1_p與對應至當前時間的第二顯示參考位置IP1_c。第一顯示參考位置IP1_p與第二顯示參考位置IP1_c可以是根據圖4實施例之計算方式而決定的交點位置。更具體而言,基於對應至先前時間的先前使用者影像與先前目標物影像,處理裝置140可根據兩空間座標的連線而獲取顯示平面上的第一顯示參考位置IP1_p。接著,舉例而言,處理裝置140可將對應至先前時間的實際顯示位置設置為第一顯示參考位置IP1_p上方M個像素的位置,M值可以依據實際需求調整,例如參考使用者的視野範圍,使實際顯示位置落入使用者視野範圍內。之後,基於對應至當前時間的當前使用者影像與當前目標物影像,處理裝置140可根據兩空間座標的連線而獲取顯示平面上的第二顯示參考位置IP1_c。處理裝置140判斷第一顯示參考位置IP1_p與第二顯示參考位置IP1_c之間的距離Δd1是否大於預設臨界值。In detail, FIG. 6A is a schematic diagram of not updating the actual display position according to an exemplary embodiment of the present disclosure. FIG. 6B is a schematic diagram of updating the actual display position according to an exemplary embodiment of the present disclosure. Referring to FIGS. 6A and 6B , in this embodiment, the display position information includes a first display reference position IP1_p corresponding to a previous time and a second display reference position IP1_c corresponding to a current time. The first display reference position IP1_p and the second display reference position IP1_c may be the intersection positions determined according to the calculation method in the embodiment of FIG. 4 . More specifically, based on the previous user image and the previous object image corresponding to the previous time, the processing device 140 can obtain the first display reference position IP1_p on the display plane according to the connection line between the two spatial coordinates. Then, for example, the processing device 140 can set the actual display position corresponding to the previous time as the position M pixels above the first display reference position IP1_p, and the M value can be adjusted according to actual needs, for example, referring to the user's field of view, Make the actual display position fall within the user's field of vision. Afterwards, based on the current user image and the current object image corresponding to the current time, the processing device 140 can obtain the second display reference position IP1_c on the display plane according to the connection line between the two spatial coordinates. The processing device 140 determines whether the distance Δd1 between the first display reference position IP1_p and the second display reference position IP1_c is greater than a preset threshold.

若距離Δd1未大於預設臨界值,請參照圖6A,處理裝置140可不更新虛擬資訊Vf1於顯示器110上對應於當前時間的實際顯示位置,致使虛擬資訊Vf1將維持於顯示於第一顯示參考位置IP1_p正上方M個像素的位置。相反地,若距離Δd1大於預設臨界值,請參照圖6B,處理裝置140根據對應至當前時間的第二顯示參考位置IP1_c更新虛擬資訊Vf1於顯示器110上對應於當前時間的實際顯示位置,像是處理裝置140可將對應至當前時間的實際顯示位置設置為第二顯示參考位置IP1_c上方M個像素的位置。If the distance Δd1 is not greater than the preset threshold, please refer to FIG. 6A , the processing device 140 may not update the actual display position of the virtual information Vf1 corresponding to the current time on the display 110, so that the virtual information Vf1 will remain displayed at the first display reference position. The position of M pixels directly above IP1_p. Conversely, if the distance Δd1 is greater than the preset threshold, please refer to FIG. 6B , the processing device 140 updates the actual display position of the virtual information Vf1 corresponding to the current time on the display 110 according to the second display reference position IP1_c corresponding to the current time, as shown in FIG. Yes, the processing device 140 may set the actual display position corresponding to the current time as the position M pixels above the second display reference position IP1_c.

圖7是根據本揭露一範例實施例所繪示的資訊顯示方法的流程圖。請參照圖7,於步驟S702,第一資訊擷取裝置120擷取使用者影像,其中使用者個數可為一個以上。於步驟S704,第一資訊擷取裝置120對多張使用者影像進行人臉偵測而獲取各使用者影像中的人臉位置。步驟S702與步驟S704的細節可參照圖4中步驟S402與步驟S404的類似相關說明,於此不再贅述。FIG. 7 is a flowchart of an information display method according to an exemplary embodiment of the present disclosure. Please refer to FIG. 7 , in step S702 , the first information capture device 120 captures images of users, where the number of users may be more than one. In step S704, the first information capture device 120 performs face detection on multiple user images to obtain the face positions in each user image. The details of step S702 and step S704 can refer to similar related descriptions of step S402 and step S404 in FIG. 4 , which will not be repeated here.

於步驟S706,第一資訊擷取裝置120根據人臉位置進行座標轉換而獲取使用者U1的空間位置資訊。於此,第一資訊擷取裝置120可將人臉位置轉換為對應至預設影像解析度的畫素座標。於圖7的實施例中,使用者U1的空間位置資訊可包括使用者U1相對於第一影像感測器的水平偏移角以及垂直偏移角。詳細而言,根據人臉位置的二維畫素座標以及第一影像感測器的水平視角(Field of View,FOV),第一資訊擷取裝置120可計算出使用者U1相對於鏡頭光軸的水平偏移角。根據人臉位置的二維畫素座標以及第一影像感測器的垂直FOV,第一資訊擷取裝置120可計算出使用者U1相對於鏡頭光軸的垂直偏移角。In step S706, the first information capturing device 120 performs coordinate transformation according to the face position to obtain the spatial position information of the user U1. Here, the first information capturing device 120 can convert the face position into pixel coordinates corresponding to the default image resolution. In the embodiment of FIG. 7 , the spatial position information of the user U1 may include a horizontal offset angle and a vertical offset angle of the user U1 relative to the first image sensor. Specifically, according to the two-dimensional pixel coordinates of the face position and the horizontal field of view (Field of View, FOV) of the first image sensor, the first information capture device 120 can calculate the relative distance between the user U1 and the optical axis of the lens. the horizontal offset angle. According to the two-dimensional pixel coordinates of the face position and the vertical FOV of the first image sensor, the first information capturing device 120 can calculate the vertical offset angle of the user U1 relative to the optical axis of the lens.

於步驟S708,第二資訊擷取裝置130擷取目標物影像。於步驟S710,第二資訊擷取裝置130對多張目標物影像進行物件偵測而獲取目標物於各目標物影像中的物件邊界框。步驟S708與步驟S710的細節可參照圖4中步驟S410與步驟S412的類似相關說明,於此不再贅述。In step S708, the second information capturing device 130 captures the target image. In step S710 , the second information capturing device 130 performs object detection on multiple target images to obtain the object bounding box of the target in each target image. The details of step S708 and step S710 can refer to similar related descriptions of step S410 and step S412 in FIG. 4 , and will not be repeated here.

於步驟S712,第二資訊擷取裝置130根據物件邊界框進行座標轉換而獲取目標物Obj1的空間位置資訊。於此,第二資訊擷取裝置130可將物件邊界框轉換為對應至預設影像解析度的畫素座標。於圖7的實施例中,目標物Obj1的空間位置資訊可包括目標物Obj1相對於第一影像感測器的水平偏移角以及垂直偏移角。詳細而言,根據目標物影像中物件邊界框之垂直框邊界的畫素座標以及第一影像感測器的水平FOV,第一資訊擷取裝置120可計算出目標物Obj1相對於鏡頭光軸的水平偏移角,亦即物件邊界框之左右框邊界各自相對於鏡頭光軸的水平偏移角。根據目標物影像中物件邊界框之水平框邊界的畫素座標以及第一影像感測器的垂直FOV,第一資訊擷取裝置120可計算出目標物Obj1相對於鏡頭光軸的垂直偏移角,亦即物件邊界框之上下框邊界各自相對於鏡頭光軸的垂直偏移角。In step S712, the second information capturing device 130 performs coordinate transformation according to the object bounding box to obtain the spatial position information of the object Obj1. Here, the second information capturing device 130 can convert the object bounding box into pixel coordinates corresponding to the default image resolution. In the embodiment of FIG. 7 , the spatial position information of the object Obj1 may include a horizontal offset angle and a vertical offset angle of the object Obj1 relative to the first image sensor. Specifically, according to the pixel coordinates of the vertical frame boundary of the object's bounding box in the target image and the horizontal FOV of the first image sensor, the first information capturing device 120 can calculate the position of the target Obj1 relative to the optical axis of the lens. Horizontal offset angle, that is, the horizontal offset angle of the left and right borders of the object's bounding box relative to the optical axis of the lens. According to the pixel coordinates of the horizontal border of the object's bounding box in the target image and the vertical FOV of the first image sensor, the first information capture device 120 can calculate the vertical offset angle of the target Obj1 relative to the optical axis of the lens , that is, the vertical offset angles of the upper and lower borders of the object's bounding box relative to the optical axis of the lens.

於步驟S714,處理裝置140根據目標物Obj1相對於第二影像感測器的水平偏移角以及垂直偏移角、使用者U1相對於第一影像感測器的水平偏移角以及垂直偏移角、第一影像感測器的拍攝方向以及第二影像感測器的拍攝方向,決定虛擬資訊Vf1於顯示器110上的顯示位置資訊。In step S714, the processing device 140 according to the horizontal offset angle and the vertical offset angle of the target object Obj1 relative to the second image sensor, the horizontal offset angle and the vertical offset angle of the user U1 relative to the first image sensor The angle, the shooting direction of the first image sensor and the shooting direction of the second image sensor determine the display position information of the virtual information Vf1 on the display 110 .

詳細而言,圖8A與圖8B是根據本揭露一範例實施例所繪示的決定顯示位置資訊的示意圖。於一實施例中,第一影像感測器、第二影像感測器與顯示器110之間的相對位置關係可先定義。舉例而言,若以顯示器110上緣的中心點為原點且顯示平面為XY平面,則第一影像感測器之鏡頭的位置以及第二影像感測器之鏡頭的位置可根據實際裝設位置而定義出來。In detail, FIG. 8A and FIG. 8B are schematic diagrams of determining display location information according to an exemplary embodiment of the present disclosure. In one embodiment, the relative positional relationship among the first image sensor, the second image sensor and the display 110 can be defined in advance. For example, if the center point of the upper edge of the display 110 is taken as the origin and the display plane is the XY plane, the position of the lens of the first image sensor and the position of the lens of the second image sensor can be determined according to the actual installation. position is defined.

請先參照圖8A,以下將先以目標物Obj1之物件邊界框Roi1的左邊界BL為範例進行說明。第一資訊擷取裝置120可將使用者U1相對於第一影像感測器的水平偏移角θ3以及使用者U1距離第一影像感測器的深度資訊d1提供給處理裝置140。第二資訊擷取裝置130可將目標物Obj1相對於第二影像感測器的水平偏移角θ4以及目標物Obj1距離第二影像感測器的深度資訊d2提供給處理裝置140。Please refer to FIG. 8A first, and the following will take the left boundary BL of the object bounding box Roi1 of the object Obj1 as an example for illustration. The first information capturing device 120 can provide the processing device 140 with the horizontal offset angle θ3 of the user U1 relative to the first image sensor and the depth information d1 of the user U1 from the first image sensor. The second information capturing device 130 can provide the processing device 140 with the horizontal offset angle θ4 of the object Obj1 relative to the second image sensor and the depth information d2 of the object Obj1 from the second image sensor.

於是,處理裝置140可獲取角度θ5(θ5=180-θ3-θ4),並基於餘弦定理而根據深度資訊d1與深度資訊d2計算出使用者U1與左邊界BL之間的距離d3。接著,處理裝置140可基於正弦定理而依據角度θ5、深度資訊d1、深度資訊d2與距離d3計算出角度θ6,並根據角度θ6與水平偏移角θ3計算出角度θ7。於是,處理裝置140可計算使用者U1與顯示器110之中線L1之間的橫向偏移距離為d1*sinθ3,以及計算左邊界BL與顯示器110之中線L1之間的橫向偏移距離為d2*sinθ4。此外,處理裝置140還可計算出顯示平面上左邊界BL的交點Q的橫向偏移距離Qx1等於d1*sinθ3+d1*cosθ3*tanθ7。同理,物件邊界框Roi1右邊界的交點Q的橫向偏移距離亦可以類似的方式計算出。Therefore, the processing device 140 can obtain the angle θ5 (θ5=180−θ3−θ4), and calculate the distance d3 between the user U1 and the left boundary BL according to the depth information d1 and depth information d2 based on the law of cosines. Then, the processing device 140 can calculate the angle θ6 according to the angle θ5 , the depth information d1 , the depth information d2 and the distance d3 based on the law of sine, and calculate the angle θ7 according to the angle θ6 and the horizontal offset angle θ3 . Therefore, the processing device 140 can calculate the lateral offset distance between the user U1 and the central line L1 of the display 110 as d1*sinθ3, and calculate the lateral offset distance between the left border BL and the central line L1 of the display 110 as d2 *sin θ4. In addition, the processing device 140 can also calculate the lateral offset distance Qx1 of the intersection point Q of the left border BL on the display plane equal to d1*sinθ3+d1*cosθ3*tanθ7. Similarly, the lateral offset distance of the intersection point Q of the right boundary of the object bounding box Roi1 can also be calculated in a similar manner.

接著,請先參照圖8B,以下將先以目標物Obj1之物件邊界框Roi1的上邊界UL為範例進行說明。第一資訊擷取裝置120可將第一影像感測器的拍攝俯仰角θ1與使用者U1相對於第一影像感測器的垂直偏移角θ9提供給處理裝置140。第二資訊擷取裝置130可將第二影像感測器的拍攝俯仰角θ2與目標物Obj1相對於第二影像感測器的垂直偏移角θ8提供給處理裝置140。Next, please refer to FIG. 8B , and the following will take the upper boundary UL of the object bounding box Roi1 of the object Obj1 as an example for illustration. The first information capturing device 120 can provide the shooting elevation angle θ1 of the first image sensor and the vertical offset angle θ9 of the user U1 relative to the first image sensor to the processing device 140 . The second information capturing device 130 can provide the shooting elevation angle θ2 of the second image sensor and the vertical offset angle θ8 of the object Obj1 relative to the second image sensor to the processing device 140 .

於是,依據深度資訊d2、拍攝俯仰角θ2與垂直偏移角θ8,處理裝置140可計算上邊界UL與顯示器110的上緣之間的距離為d2sin(θ2+θ8)。依據深度資訊d1、拍攝俯仰角θ1與垂直偏移角θ9,處理裝置140可計算使用者U1與顯示器110的上緣之間的距離為d1sin(θ1+θ9)。於一實施例中,處理裝置140還可依據兩影像感測器之鏡頭與顯示器110的上緣之間的距離來進一步校正計算前述計算結果,而獲取使用者U1與目標物Obj與顯示器110的上緣之間的實際距離。之後,處理裝置140可基於下列公式(4)計算出顯示平面上上邊界UL的交點Q的直向偏移距離Eu1。

Figure 02_image007
公式(4) 其中,使用者U1與顯示器110的上緣之間的實際距離A1可表示為
Figure 02_image009
Figure 02_image011
1為第一影像感測器之鏡頭與顯示器110的上緣之間的距離。目標物Obj與顯示器110的上緣之間的實際距離A2可表示為
Figure 02_image013
Figure 02_image015
為第二影像感測器之鏡頭與顯示器110的上緣之間的距離。同理,物件邊界框Roi1下邊界的交點的直向偏移距離亦可以類似的方式計算出。 Therefore, according to the depth information d2, the shooting pitch angle θ2 and the vertical offset angle θ8, the processing device 140 can calculate the distance between the upper boundary UL and the upper edge of the display 110 as d2sin(θ2+θ8). According to the depth information d1, the shooting pitch angle θ1 and the vertical offset angle θ9, the processing device 140 can calculate the distance between the user U1 and the upper edge of the display 110 as d1sin(θ1+θ9). In one embodiment, the processing device 140 can further correct and calculate the aforementioned calculation result according to the distance between the lenses of the two image sensors and the upper edge of the display 110 , so as to obtain the distance between the user U1, the object Obj and the display 110 The actual distance between the upper edges. Afterwards, the processing device 140 can calculate the vertical offset distance Eu1 of the intersection point Q of the upper boundary UL on the display plane based on the following formula (4).
Figure 02_image007
Formula (4) where, the actual distance A1 between the user U1 and the upper edge of the display 110 can be expressed as
Figure 02_image009
,
Figure 02_image011
1 is the distance between the lens of the first image sensor and the upper edge of the display 110 . The actual distance A2 between the object Obj and the upper edge of the display 110 can be expressed as
Figure 02_image013
,
Figure 02_image015
is the distance between the lens of the second image sensor and the upper edge of the display 110 . Similarly, the vertical offset distance of the intersection point of the lower boundary of the object bounding box Roi1 can also be calculated in a similar manner.

基於上述說明,處理裝置140可獲取顯示平面上對應於物件邊界框Roi1的參考顯示物件框。詳細而言,基於圖8A所示之計算方法,處理裝置140可獲取顯示平面上參考顯示物件框的左右邊界分別與顯示器110之中線L1之間的橫向偏移距離。基於圖8B所示之計算方法,處理裝置140可獲取顯示平面上參考顯示物件框的上下邊界分別與顯示器110上緣之間的直向偏移距離。換言之,顯示平面上對應於物件邊界框Roi1的參考顯示物件框可由平行於顯示X軸的兩直線方程式與平行於顯示Y軸的兩直線方程式來表示。其中,參考顯示物件框的四個框頂點可由這些直線方程式的交點獲取。Based on the above description, the processing device 140 may acquire a reference display object box corresponding to the object bounding box Roi1 on the display plane. In detail, based on the calculation method shown in FIG. 8A , the processing device 140 can obtain the lateral offset distances between the left and right borders of the reference display object frame on the display plane and the center line L1 of the display 110 . Based on the calculation method shown in FIG. 8B , the processing device 140 can obtain the vertical offset distances between the upper and lower boundaries of the reference display object frame on the display plane and the upper edge of the display 110 respectively. In other words, the reference display object box corresponding to the object bounding box Roi1 on the display plane can be represented by two straight line equations parallel to the display X axis and two straight line equations parallel to the display Y axis. Among them, the four frame vertices of the reference display object frame can be obtained from the intersection points of these straight line equations.

之後,處理裝置140可根據顯示器110的尺寸與解析度獲取顯示位置資訊,而顯示位置資訊可為顯示器110的畫素座標。更詳細來說,處理裝置140可根據顯示器110的尺寸與解析度將參考顯示物件框轉換為顯示器110上的畫素座標,以獲取虛擬資訊Vf1於顯示器110上的顯示位置資訊。亦即,由畫素座標系來表示顯示平面上的參考顯示物件框。Afterwards, the processing device 140 can obtain display position information according to the size and resolution of the display 110 , and the display position information can be the pixel coordinates of the display 110 . In more detail, the processing device 140 can convert the reference display object frame into pixel coordinates on the display 110 according to the size and resolution of the display 110 to obtain display position information of the virtual information Vf1 on the display 110 . That is, the reference display object frame on the display plane is represented by the pixel coordinate system.

回到圖7,於步驟S716,處理裝置140比較對應於先前時間的第一參考顯示物件框以及對應於當前時間的第二參考顯示物件框之間的距離與預設臨界值,以決定虛擬資訊Vf1於顯示器110上對應於當前時間的實際顯示位置。於步驟S718,處理裝置140根據實際顯示位置顯示虛擬資訊Vf1於顯示器110上。Returning to FIG. 7, in step S716, the processing device 140 compares the distance between the first reference display object frame corresponding to the previous time and the second reference display object frame corresponding to the current time with a preset threshold value to determine the virtual information Vf1 corresponds to the actual display position of the current time on the display 110 . In step S718 , the processing device 140 displays the virtual information Vf1 on the display 110 according to the actual display position.

詳細而言,圖9A是根據本揭露一範例實施例所繪示的不更新實際顯示位置的示意圖。圖9B是根據本揭露一範例實施例所繪示的更新實際顯示位置的示意圖。請參照圖9A與9B,於本實施範例中,顯示位置資訊包括於對應於目標物影像中物件邊界框的參考顯示物件框。參考顯示物件框可包括對應於先前時間的第一參考顯示物件框Roi_p以及對應於當前時間的第二參考顯示物件框Roi_c。第一顯示參考位置包括第一參考顯示物件框Roi_p,而第二顯示參考位置包括第二參考顯示物件框Roi_c。In detail, FIG. 9A is a schematic diagram of not updating the actual display position according to an exemplary embodiment of the present disclosure. FIG. 9B is a schematic diagram of updating the actual display position according to an exemplary embodiment of the present disclosure. Referring to FIGS. 9A and 9B , in this embodiment, the display location information is included in the reference display object box corresponding to the object bounding box in the target image. The reference display object frame may include a first reference display object frame Roi_p corresponding to a previous time and a second reference display object frame Roi_c corresponding to a current time. The first display reference position includes a first reference display object frame Roi_p, and the second display reference position includes a second reference display object frame Roi_c.

對應於先前時間的第一參考顯示物件框Roi_p以及對應於當前時間的第二參考顯示物件框Roi_c可以是根據圖7、圖8A與圖8B所示實施例之計算方式而決定。更具體而言,基於對應至先前時間的先前使用者影像與先前目標物影像,處理裝置140可獲取顯示平面上的第一參考顯示物件框Roi_p。接著,舉例而言,假設第一參考顯示物件框Roi_p的左上頂點的畫素座標為(x1,y1),處理裝置140可將對應至先前時間的實際顯示位置設置為畫素座標為(x1+αw1, y1+βh1)。其中,w1為第一參考顯示物件框Roi_p的寬度,h1為第一參考顯示物件框Roi_p的高度。α與β可視實際而設置,本揭露對此不限制。The first reference display object frame Roi_p corresponding to the previous time and the second reference display object frame Roi_c corresponding to the current time may be determined according to the calculation methods shown in the embodiments shown in FIG. 7 , FIG. 8A and FIG. 8B . More specifically, based on the previous user image and the previous object image corresponding to the previous time, the processing device 140 may acquire the first reference display object frame Roi_p on the display plane. Next, for example, assuming that the pixel coordinates of the upper left vertex of the first reference display object frame Roi_p are (x1, y1), the processing device 140 can set the actual display position corresponding to the previous time as the pixel coordinates of (x1+ αw1, y1+βh1). Wherein, w1 is the width of the first reference display object frame Roi_p, and h1 is the height of the first reference display object frame Roi_p. α and β can be set according to the actual situation, which is not limited in this disclosure.

之後,基於對應至當前時間的當前使用者影像與當前目標物影像,處理裝置140可獲取顯示平面上的第二參考顯示物件框Roi_c。處理裝置140比較對應於先前時間的第一參考顯示物件框Roi_p以及對應於當前時間的第二參考顯示物件框Roi_c之間的距離Δd2與預設臨界值。Afterwards, based on the current user image and the current object image corresponding to the current time, the processing device 140 may acquire the second reference display object frame Roi_c on the display plane. The processing device 140 compares the distance Δd2 between the first reference display object frame Roi_p corresponding to the previous time and the second reference display object frame Roi_c corresponding to the current time with a preset threshold.

處理裝置140計算第一參考顯示物件框Roi_p的第一參考點與第二參考顯示物件框Roi_c的第二參考點之間的距離。第一參考點與第二參考點可例如是框中心點或框頂點。於圖9A與9B的範例中,第一參考點與第二參考點分別為第一參考顯示物件框Roi_p的左上頂點與第二參考顯示物件框Roi_c的左上頂點。The processing device 140 calculates the distance between the first reference point of the first reference display object frame Roi_p and the second reference point of the second reference display object frame Roi_c. The first reference point and the second reference point may be, for example, a frame center point or a frame vertex. In the examples of FIGS. 9A and 9B , the first reference point and the second reference point are respectively the upper left vertex of the first reference display object frame Roi_p and the upper left vertex of the second reference display object frame Roi_c.

於一實施例中,此預設臨界值可以是根據第二參考顯示物件框Roi_c的尺寸而決定。舉例而言,預設臨界值可等於min(γ*w2, γ*h2)。其中,w2為第二參考顯示物件框Roi_c的寬度,h2為第二參考顯示物件框Roi_c的高度。γ可為介於0到1之間的數值,其可視實際而設置,本揭露對此不限制。min(․)代表取最小值的函數。由此可見,預設臨界值可反應於參考物件框的尺寸而彈性設置。預設臨界值與參考物件框的尺寸可成正比關係。In an embodiment, the preset threshold may be determined according to the size of the second reference display object frame Roi_c. For example, the preset threshold may be equal to min(γ*w2, γ*h2). Wherein, w2 is the width of the second reference display object frame Roi_c, and h2 is the height of the second reference display object frame Roi_c. γ can be a value between 0 and 1, which can be set according to the actual situation, which is not limited in the present disclosure. min(․) represents the function of taking the minimum value. It can be seen that the preset threshold can be flexibly set in response to the size of the reference object frame. The preset critical value may be directly proportional to the size of the reference object frame.

若距離Δd2未大於預設臨界值,請參照圖9A,處理裝置140不更新虛擬資訊Vf1於顯示器110上對應於當前時間的實際顯示位置,虛擬資訊Vf1可維持於顯示於第一參考顯示物件框Roi_p的上方。相反地,若距離Δd2大於預設臨界值,請參照圖9B,處理裝置140根據對應至當前時間的第二參考顯示物件框Roi_c更新虛擬資訊Vf1於顯示器110上對應於當前時間的實際顯示位置。舉例而言,假設第二參考顯示物件框Roi_c的左上頂點的畫素座標為(x2,y2),處理裝置140可將對應至當前時間的實際顯示位置設置為畫素座標為(x2+αw2, y2+βh2)。其中,w2為第二參考顯示物件框Roi_c的寬度,h2為第二參考顯示物件框Roi_c的高度。α與β可視實際而設置,本揭露對此不限制。If the distance Δd2 is not greater than the preset threshold, please refer to FIG. 9A , the processing device 140 does not update the actual display position of the virtual information Vf1 on the display 110 corresponding to the current time, and the virtual information Vf1 can remain displayed in the first reference display object frame. Above Roi_p. Conversely, if the distance Δd2 is greater than the preset threshold, please refer to FIG. 9B , the processing device 140 updates the actual display position of the virtual information Vf1 corresponding to the current time on the display 110 according to the second reference display object frame Roi_c corresponding to the current time. For example, assuming that the pixel coordinates of the upper left vertex of the second reference display object frame Roi_c are (x2, y2), the processing device 140 may set the actual display position corresponding to the current time as the pixel coordinates of (x2+αw2, y2+βh2). Wherein, w2 is the width of the second reference display object frame Roi_c, and h2 is the height of the second reference display object frame Roi_c. α and β can be set according to the actual situation, which is not limited in this disclosure.

於一實施例中,反應於顯示位置資訊位於顯示器110的顯示邊緣區域,處理裝置140往遠離顯示器110之邊緣的方向調整顯示位置資訊。詳細而言,在處理裝置140獲取某一時間點的顯示位置資訊之後,若顯示位置資訊位於顯示器110的顯示邊緣區域,代表虛擬資訊可能會顯示於顯示器110的顯示邊緣區域或甚至是顯示不完全,導致使用者U1觀看上的困難度。於是,處理裝置140將進一步調整顯示位置資訊,例如處理裝置140可往遠離顯示器110之邊緣的方向調整顯示位置資訊。In one embodiment, the processing device 140 adjusts the display position information in a direction away from the edge of the display 110 in response to the fact that the display position information is located in the display edge area of the display 110 . In detail, after the processing device 140 obtains the display position information at a certain time point, if the display position information is located in the display edge area of the display 110, it means that the virtual information may be displayed in the display edge area of the display 110 or even displayed incompletely. , resulting in difficulty for the user U1 to watch. Therefore, the processing device 140 will further adjust the display position information, for example, the processing device 140 can adjust the display position information in a direction away from the edge of the display 110 .

舉例而言,當基於圖5B的計算方式而計算出來的交點位置位於顯示器110的顯示邊緣區域時,處理裝置140可將交點位置往遠離顯示器110之邊緣的方向調整,再根據調整後的交點位置決定實際顯示位置。或者,當基於圖8A與8B的計算方式而計算出來的參考顯示物件框位於顯示器110的顯示邊緣區域時,處理裝置140可將參考顯示物件框往遠離顯示器110之邊緣的方向調整,再根據調整後的參考顯示物件框決定實際顯示位置。For example, when the intersection position calculated based on the calculation method in FIG. 5B is located in the display edge area of the display 110, the processing device 140 can adjust the intersection position away from the edge of the display 110, and then according to the adjusted intersection position Determines the actual display position. Alternatively, when the reference display object frame calculated based on the calculation methods in FIGS. 8A and 8B is located in the display edge area of the display 110, the processing device 140 may adjust the reference display object frame to a direction away from the edge of the display 110, and then adjust The following reference display object box determines the actual display position.

於一實施例中,處理裝置140可根據下列公式(5)來調整顯示位置資訊於垂直軸向上的座標分量。

Figure 02_image017
公式(5) 其中,
Figure 02_image019
為顯示位置資訊的原始Y座標、F為調整權重、c為常數。F例如是介於1至100之間的數值,其可視實際而設置,本揭露對此不限制。常數c可根據原始Y座標所在顯示區域而決定。舉例而言,假設顯示位置資訊的原始Y座標位於顯示範圍的中心區域,則常數c可設置為等於原始Y座標,即不調整原始Y座標。假設顯示位置資訊的原始Y座標位於顯示範圍的邊緣區域,則常數c可設置為不等於原始Y座標,即往遠離顯示器110之邊緣的方向調整原始Y座標。同理,顯示位置資訊於水平軸向上的座標分量亦可以類似的方式進行調整。 In one embodiment, the processing device 140 can adjust the coordinate component of the display position information on the vertical axis according to the following formula (5).
Figure 02_image017
Formula (5) where,
Figure 02_image019
is the original Y coordinate of the displayed location information, F is the adjustment weight, and c is a constant. F is, for example, a value between 1 and 100, which can be set according to actual conditions, and the present disclosure is not limited thereto. The constant c can be determined according to the display area where the original Y coordinate is located. For example, assuming that the original Y coordinate of the displayed position information is located in the central area of the display range, the constant c can be set equal to the original Y coordinate, that is, the original Y coordinate is not adjusted. Assuming that the original Y coordinate of the display location information is located at the edge of the display range, the constant c can be set to be different from the original Y coordinate, that is, the original Y coordinate is adjusted away from the edge of the display 110 . Similarly, the coordinate components of the displayed location information on the horizontal axis can also be adjusted in a similar manner.

於一些實施例中,處理裝置140還可根據使用者的姿態資訊來決定虛擬資訊於顯示器110上的顯示位置資訊。第一資訊擷取裝置120可用以擷取使用者的姿態資訊。具體而言,第一資訊擷取裝置120可透過各類人體姿態辨識技術來辨識使用者的四肢、軀幹及/或頭部所呈現的姿態。像是,第一資訊擷取裝置120可根據影像資料辨識人體骨架或人體特徵點等等,從而識別出使用者的姿態。於一些實施例中,第一資訊擷取裝置120可對使用者影像進行人臉偵測而獲取各使用者影像中的人臉位置與人臉特徵點。本發明對於人臉特徵點的數量並不加以限制,其例如可包括嘴角特徵點、眉心特徵點或眼角特徵點等等。於一些實施例中,第一資訊擷取裝置120或處理裝置140可根據多幀使用者影像中的人臉特徵點來辨識出使用者的臉部中心位移量與頭部轉動角度。使用者的姿態資訊包括臉部中心位移量與頭部轉動角度。In some embodiments, the processing device 140 can also determine the display position information of the virtual information on the display 110 according to the gesture information of the user. The first information capturing device 120 can be used to capture the gesture information of the user. Specifically, the first information capturing device 120 can recognize the gestures of the user's limbs, torso and/or head through various human gesture recognition technologies. For example, the first information capturing device 120 can identify the human body skeleton or human body feature points according to the image data, so as to identify the posture of the user. In some embodiments, the first information capturing device 120 can perform face detection on user images to obtain the face positions and facial feature points in each user image. The present invention does not limit the number of facial feature points, which may include, for example, mouth corner feature points, eyebrow center feature points, eye corner feature points, and the like. In some embodiments, the first information capturing device 120 or the processing device 140 can identify the displacement of the user's face center and the head rotation angle according to the facial feature points in multiple frames of user images. The posture information of the user includes the displacement of the center of the face and the rotation angle of the head.

於一些實施例中,處理裝置140可根據使用者的姿態資訊來判斷使用者的頭部移動或轉動。反應於根據使用者的姿態資訊判定使用者的頭部移動,處理裝置140可根據使用者的空間位置資訊以及目標物的空間位置資訊決定目標物的虛擬資訊於顯示器110上的顯示位置資訊。亦即,當處理裝置140判定使用者的頭部移動時,處理裝置140可根據如圖2、圖4或圖7所示的流程來決定虛擬資訊的實際顯示位置。另一方面,反應於根據使用者的姿態資訊判定使用者的頭部轉動,處理裝置140可根據使用者的空間位置資訊與姿態資訊以及目標物的空間位置資訊決定目標物的虛擬資訊於顯示器110上的顯示位置資訊。當處理裝置140判定使用者的頭部轉動時,處理裝置140可根據使用者的頭部轉動角度來更新目標物的虛擬資訊於顯示器110上的顯示位置資訊。舉例而言,當使用者的頭部並沒有移動時,處理裝置140可利用使用者的頭部轉動角度來更新先前時間點所決定的顯示位置資訊。In some embodiments, the processing device 140 can determine the head movement or rotation of the user according to the posture information of the user. In response to determining the user's head movement based on the user's posture information, the processing device 140 can determine the display position information of the virtual information of the target on the display 110 according to the user's spatial position information and the target object's spatial position information. That is, when the processing device 140 determines that the user's head moves, the processing device 140 may determine the actual display position of the virtual information according to the flow shown in FIG. 2 , FIG. 4 or FIG. 7 . On the other hand, in response to determining the user's head rotation based on the user's posture information, the processing device 140 can determine the virtual information of the target object on the display 110 according to the user's spatial position information and posture information and the target object's spatial position information. Show location information on . When the processing device 140 determines that the user's head is turning, the processing device 140 may update the display position information of the virtual information of the target on the display 110 according to the rotation angle of the user's head. For example, when the user's head does not move, the processing device 140 can use the user's head rotation angle to update the display position information determined at the previous time point.

於一些實施例中,處理裝置140可根據姿態資訊中的臉部中心位移量來估計出一頭部行為指標。上述頭部行為指標可為臉部中心位移量除以頭部半徑基準值。舉例而言,頭部半徑基準值可為80mm。透過比對頭部行為指標以及頭部轉動角度的變化量,處理裝置140可判斷使用者的頭部移動或轉動,以反應於使用者的頭部出現移動或轉動來決定虛擬資訊於顯示器110上的顯示位置資訊。In some embodiments, the processing device 140 can estimate a head behavior index according to the face center displacement in the pose information. The above-mentioned head behavior index may be the face center displacement divided by the head radius reference value. For example, the base radius of the head may be 80mm. By comparing the head behavior index and the variation of the head rotation angle, the processing device 140 can determine the movement or rotation of the user's head, and determine the virtual information on the display 110 in response to the movement or rotation of the user's head. Show location information for .

於一些實施例中,若頭部行為指標小於等於頭部轉動角度的變化量,處理裝置140判定使用者的頭部移動,亦即使用者頭部自第一空間位置移動至第二空間位置。於是,可執行如圖2、圖4或圖7所示的流程來決定虛擬資訊的實際顯示位置。若頭部行為指標大於頭部轉動角度的變化量,處理裝置140判定使用者的頭部轉動,亦即使用者頭部轉向發生改變。於是,處理裝置140可根據使用者的空間位置資訊與頭部轉動角度以及目標物的空間位置資訊決定目標物的虛擬資訊於顯示器110上的顯示位置資訊。像是,處理裝置140可利用使用者的頭部轉動角度來更新先前時間點所決定的顯示位置資訊,並據以決定虛擬資訊的實際顯示位置。或者,處理裝置140可利用使用者的頭部轉動角度與使用者的空間位置資訊來識別另一目標物,以根據另一目標物的空間位置資訊以及使用者的空間位置資訊決定另一目標物的另一虛擬資訊於顯示器110上的顯示位置資訊。In some embodiments, if the head behavior index is less than or equal to the variation of the head rotation angle, the processing device 140 determines that the user's head moves, that is, the user's head moves from the first spatial position to the second spatial position. Therefore, the process shown in FIG. 2 , FIG. 4 or FIG. 7 can be executed to determine the actual display position of the virtual information. If the head behavior index is greater than the variation of the head rotation angle, the processing device 140 determines that the user's head is rotating, that is, the direction of the user's head changes. Therefore, the processing device 140 can determine the display position information of the virtual information of the target on the display 110 according to the spatial position information of the user, the head rotation angle and the spatial position information of the target object. For example, the processing device 140 can use the rotation angle of the user's head to update the display position information determined at the previous time point, and determine the actual display position of the virtual information accordingly. Alternatively, the processing device 140 may use the user's head rotation angle and the user's spatial position information to identify another target, so as to determine another target according to the spatial position information of the other target and the user's spatial position information. The location information of another virtual information is displayed on the display 110 .

本揭露的範例實施例所提出的資訊顯示方法及其處理裝置與資訊顯示系統,可即時地依據使用者位置與目標物位置,計算使視線投射於顯示平面上的顯示位置資訊,並據以產生虛擬資訊的實際顯示位置。藉此,即便目標物的移動多變或不可預測,根據實際顯示位置所顯示的虛擬資訊可即時地跟隨目標物的移動。此外,本揭露的範例實施例可依據先前時間的第一顯示參考位置與當前時間的第二顯示參考位置之間的距離決定是否更新虛擬資訊於顯示器上對應於當前時間的實際顯示位置。藉此,基於即時物件辨識而顯示的虛擬資訊可穩定且順暢地進行顯示,大幅提升使用者的觀看體驗。The information display method and its processing device and information display system proposed in the exemplary embodiments of the present disclosure can calculate the display position information for projecting the line of sight on the display plane according to the user position and the target object position in real time, and generate The actual display position of the virtual information. In this way, even if the movement of the target is variable or unpredictable, the virtual information displayed according to the actual display position can follow the movement of the target in real time. In addition, the exemplary embodiment of the present disclosure may determine whether to update the virtual information on the display at the actual display position corresponding to the current time according to the distance between the first display reference position at the previous time and the second display reference position at the current time. Thereby, the virtual information displayed based on real-time object recognition can be displayed stably and smoothly, greatly improving the user's viewing experience.

雖然本揭露已以範例實施例揭露如上,然其並非用以限定本揭露,任何所屬技術領域中具有通常知識者,在不脫離本揭露的精神和範圍內,當可作些許的更動與潤飾,故本揭露的保護範圍當視後附的申請專利範圍及其均等範圍所界定者為準。Although the present disclosure has been disclosed above with exemplary embodiments, it is not intended to limit the present disclosure. Anyone with ordinary knowledge in the technical field may make some changes and modifications without departing from the spirit and scope of the present disclosure. Therefore, the scope of protection of this disclosure should be defined by the appended scope of patent application and its equivalent scope.

100:資訊顯示系統 110:顯示器 120:第一資訊擷取裝置 130:第二資訊擷取裝置 140:處理裝置 141:記憶體 142:處理器 Vf1:虛擬資訊 RF1:參考顯示物件框 U1:使用者 Obj1:目標物 CX1、CX2:鏡頭光軸 HX1:顯示平面法線 IP1:交點位置 IP1_p:第一顯示參考位置 IP1_c:第二顯示參考位置 BL:左邊界 UL:上邊界 Q:交點 L1:中線 Roi1:物件邊界框 Roi_p:第一參考顯示物件框 Roi_c:第二參考顯示物件框 S210~S250、S241~S243、S402~S422、S702~S718:步驟 100:Information display system 110: Display 120: The first information retrieval device 130: the second information retrieval device 140: processing device 141: memory 142: Processor Vf1: Virtual Information RF1: Reference display object frame U1: User Obj1: target object CX1, CX2: Optical axis of lens HX1: display plane normal IP1: Intersection position IP1_p: first display reference position IP1_c: Second display reference position BL: left border UL: upper boundary Q: Intersection L1: midline Roi1: object bounding box Roi_p: first reference display object box Roi_c: Second reference display object frame S210~S250, S241~S243, S402~S422, S702~S718: steps

圖1A是根據本揭露一範例實施例所繪示的資訊顯示系統的方塊圖。 圖1B是根據本揭露一範例實施例所繪示的資訊顯示系統的示意圖。 圖2是根據本揭露一範例實施例所繪示的資訊顯示方法的流程圖。 圖3是根據本揭露一範例實施例所繪示的決定對應於當前時間的實際顯示位置的流程圖。 圖4是根據本揭露一範例實施例所繪示的資訊顯示方法的流程圖。 圖5A是根據本揭露一範例實施例所繪示的第一影像感測器與第二影像感測器的拍攝方向的示意圖。 圖5B是根據本揭露一範例實施例所繪示的決定顯示位置資訊的示意圖。 圖6A是根據本揭露一範例實施例所繪示的不更新實際顯示位置的示意圖。 圖6B是根據本揭露一範例實施例所繪示的更新實際顯示位置的示意圖。 圖7是根據本揭露一範例實施例所繪示的資訊顯示方法的流程圖。 圖8A與圖8B是根據本揭露一範例實施例所繪示的決定顯示位置資訊的示意圖。 圖9A是根據本揭露一範例實施例所繪示的不更新實際顯示位置的示意圖。 圖9B是根據本揭露一範例實施例所繪示的更新實際顯示位置的示意圖。 FIG. 1A is a block diagram of an information display system according to an exemplary embodiment of the present disclosure. FIG. 1B is a schematic diagram of an information display system according to an exemplary embodiment of the present disclosure. FIG. 2 is a flowchart of an information display method according to an exemplary embodiment of the present disclosure. FIG. 3 is a flow chart of determining an actual display position corresponding to the current time according to an exemplary embodiment of the present disclosure. FIG. 4 is a flowchart of an information display method according to an exemplary embodiment of the present disclosure. FIG. 5A is a schematic diagram illustrating shooting directions of the first image sensor and the second image sensor according to an exemplary embodiment of the present disclosure. FIG. 5B is a schematic diagram of determining to display location information according to an exemplary embodiment of the present disclosure. FIG. 6A is a schematic diagram of not updating the actual display position according to an exemplary embodiment of the present disclosure. FIG. 6B is a schematic diagram of updating the actual display position according to an exemplary embodiment of the present disclosure. FIG. 7 is a flowchart of an information display method according to an exemplary embodiment of the present disclosure. FIG. 8A and FIG. 8B are schematic diagrams of determining display location information according to an exemplary embodiment of the present disclosure. FIG. 9A is a schematic diagram of not updating the actual display position according to an exemplary embodiment of the present disclosure. FIG. 9B is a schematic diagram of updating the actual display position according to an exemplary embodiment of the present disclosure.

S210~S250:步驟 S210~S250: steps

Claims (20)

一種資訊顯示系統,包括: 可透光的顯示器; 第一資訊擷取裝置,用以擷取一使用者的空間位置資訊; 第二資訊擷取裝置,用以擷取一目標物的空間位置資訊; 處理裝置,連接於所述顯示器、所述第一資訊擷取裝置以及所述第二資訊擷取裝置,經配置以: 根據所述使用者的空間位置資訊以及所述目標物的空間位置資訊決定所述目標物的一虛擬資訊於所述顯示器上的顯示位置資訊,其中所述顯示位置資訊包括對應於先前時間的第一顯示參考位置與對應至當前時間的第二顯示參考位置; 根據所述第一顯示參考位置與所述第二顯示參考位置之間的距離決定所述虛擬資訊於所述顯示器上對應於所述當前時間的實際顯示位置;以及 根據所述實際顯示位置顯示所述虛擬資訊於所述顯示器上。 An information display system, comprising: Light-transmitting displays; a first information retrieval device, configured to retrieve spatial location information of a user; The second information capturing device is used for capturing spatial position information of a target; a processing device connected to the display, the first information capture device and the second information capture device, configured to: Determining display position information of virtual information of the target on the display according to the spatial position information of the user and the spatial position information of the target, wherein the display position information includes the first time corresponding to the previous time a display reference position and a second display reference position corresponding to the current time; determining an actual display position of the virtual information corresponding to the current time on the display according to a distance between the first display reference position and the second display reference position; and and displaying the virtual information on the display according to the actual display position. 如請求項1所述的資訊顯示系統,其中所述處理裝置經配置以: 反應於所述第一顯示參考位置與所述第二顯示參考位置之間的距離大於一預設臨界值,根據對應至所述當前時間的所述第二顯示參考位置更新所述虛擬資訊於所述顯示器上對應於所述當前時間的所述實際顯示位置;以及 反應於所述第一顯示參考位置與所述第二顯示參考位置之間的距離未大於一預設臨界值,不更新所述虛擬資訊於所述顯示器上對應於所述當前時間的所述實際顯示位置。 The information display system as claimed in claim 1, wherein the processing device is configured to: In response to a distance between the first display reference position and the second display reference position being greater than a predetermined threshold, updating the virtual information at the second display reference position corresponding to the current time said actual display position on said display corresponding to said current time; and In response to the fact that the distance between the first display reference position and the second display reference position is not greater than a predetermined threshold, not updating the virtual information on the display corresponding to the actual time at the current time Show location. 如請求項1所述的資訊顯示系統,其中所述第一資訊擷取裝置用以擷取該使用者的姿態資訊,而所述處理裝置更經配置以: 根據所述使用者的姿態資訊判定所述使用者的頭部移動或轉動;以及 反應於根據所述使用者的姿態資訊判定所述使用者的頭部移動,根據所述使用者的空間位置資訊以及所述目標物的空間位置資訊決定所述目標物的所述虛擬資訊於所述顯示器上的所述顯示位置資訊,或反應於根據所述使用者的姿態資訊判定所述使用者的頭部轉動,根據所述使用者的空間位置資訊與姿態資訊以及所述目標物的空間位置資訊決定所述目標物的所述虛擬資訊於所述顯示器上的所述顯示位置資訊。 The information display system as described in claim 1, wherein the first information capturing device is used to capture the user's posture information, and the processing device is further configured to: determining head movement or rotation of the user based on posture information of the user; and In response to determining the user's head movement according to the user's posture information, determining the virtual information of the target object based on the user's spatial position information and the target object's spatial position information The display position information on the display, or in response to determining the user’s head rotation based on the user’s posture information, according to the user’s spatial position information and posture information and the space of the target The position information determines the display position information of the virtual information of the target on the display. 如請求項1所述的資訊顯示系統,其中所述第一資訊擷取裝置包括用以拍攝所述使用者的第一影像感測器,且所述第二資訊擷取裝置包括用以拍攝所述目標物的第二影像感測器。The information display system according to claim 1, wherein the first information capture device includes a first image sensor for photographing the user, and the second information capture device includes a sensor for photographing the user The second image sensor of the target object. 如請求項4所述的資訊顯示系統,其中所述第二資訊擷取裝置對多張目標物影像進行物件偵測而獲取所述目標物於各所述目標物影像中的物件邊界框,且所述第二資訊擷取裝置根據所述物件邊界框進行座標轉換而獲取所述目標物的空間位置資訊, 其中所述第一資訊擷取裝置對多張使用者影像進行人臉偵測而獲取各所述使用者影像中的人臉位置,且所述第一資訊擷取裝置根據所述人臉位置進行座標轉換而獲取所述使用者的空間位置資訊。 The information display system as described in claim 4, wherein the second information capturing device performs object detection on a plurality of target images to obtain the object bounding box of the target in each of the target images, and The second information acquisition device performs coordinate conversion according to the object bounding box to obtain the spatial position information of the target object, Wherein the first information capturing device performs face detection on a plurality of user images to obtain the face position in each of the user images, and the first information capturing device performs face detection according to the face position coordinate conversion to obtain the spatial location information of the user. 如請求項4所述的資訊顯示系統,其中所述目標物的空間位置資訊包括相機座標系下的目標物相機座標,所述使用者的空間位置資訊包括相機座標系下的使用者相機座標, 其中所述處理裝置根據所述第一影像感測器的拍攝方向將所述使用者相機座標轉換為相對於所述顯示器的使用者位置座標,並根據所述第二影像感測器的拍攝方向將所述目標物相機座標轉換為相對於所述顯示器的目標物位置座標, 其中所述處理裝置根據所述目標物位置座標與所述使用者位置座標之間的連線決定所述虛擬資訊於所述顯示器上的所述顯示位置資訊。 The information display system as described in claim 4, wherein the spatial position information of the target includes the target camera coordinates in the camera coordinate system, and the user spatial position information includes the user camera coordinates in the camera coordinate system, Wherein the processing device converts the user camera coordinates into user position coordinates relative to the display according to the shooting direction of the first image sensor, and converting the object camera coordinates to object position coordinates relative to the display, Wherein the processing device determines the display position information of the virtual information on the display according to the connection line between the target position coordinates and the user position coordinates. 如請求項6所述的資訊顯示系統,其中所述顯示位置資訊包括所述連線相交於所述顯示器的顯示平面的一交點位置。The information display system according to claim 6, wherein the display position information includes a position of an intersection point where the connection line intersects the display plane of the display. 如請求項4所述的資訊顯示系統,其中所述目標物的空間位置資訊包括所述目標物相對於所述第二影像感測器的水平偏移角以及垂直偏移角,所述使用者的空間位置資訊包括所述使用者相對於所述第一影像感測器的水平偏移角以及垂直偏移角, 其中所述處理裝置根據所述目標物相對於所述第二影像感測器的水平偏移角以及垂直偏移角、所述使用者相對於所述第一影像感測器的水平偏移角以及垂直偏移角、所述第一影像感測器的拍攝方向以及所述第二影像感測器的拍攝方向,決定所述虛擬資訊於所述顯示器上的所述顯示位置資訊。 The information display system according to claim 4, wherein the spatial position information of the target includes a horizontal offset angle and a vertical offset angle of the target relative to the second image sensor, and the user The spatial position information includes the horizontal offset angle and the vertical offset angle of the user relative to the first image sensor, Wherein the processing device is based on the horizontal deviation angle and the vertical deviation angle of the target object relative to the second image sensor, the horizontal deviation angle of the user relative to the first image sensor And the vertical offset angle, the shooting direction of the first image sensor and the shooting direction of the second image sensor determine the display position information of the virtual information on the display. 如請求項8所述的資訊顯示系統,其中所述第二資訊擷取裝置對多張目標物影像進行物件偵測而獲取所述目標物於各所述目標物影像中的一物件邊界框,所述顯示位置資訊包括於所述顯示器的顯示平面上對應於各所述物件邊界框的參考顯示物件框, 其中所述參考顯示物件框包括對應於所述先前時間的第一參考顯示物件框以及對應於所述當前時間的第二參考顯示物件框,所述第一顯示參考位置包括所述第一參考顯示物件框,而所述第二顯示參考位置包括所述第二參考顯示物件框, 其中所述處理裝置比較對應於所述先前時間的所述第一參考顯示物件框以及對應於所述當前時間的所述第二參考顯示物件框之間的距離與一預設臨界值,以決定所述虛擬資訊於所述顯示器上對應於所述當前時間的所述實際顯示位置。 The information display system as described in claim 8, wherein the second information capturing device performs object detection on a plurality of target images to obtain an object bounding box of the target in each of the target images, the display position information includes a reference display object frame corresponding to each of the object bounding boxes on the display plane of the display, Wherein the reference display object frame includes a first reference display object frame corresponding to the previous time and a second reference display object frame corresponding to the current time, and the first display reference position includes the first reference display object frame an object frame, and the second display reference location includes the second reference display object frame, Wherein the processing device compares the distance between the first reference display object frame corresponding to the previous time and the second reference display object frame corresponding to the current time with a preset threshold value to determine The virtual information corresponds to the actual display position of the current time on the display. 如請求項9所述的資訊顯示系統,其中所述預設臨界值是根據所述第二參考顯示物件框的尺寸而決定。The information display system according to claim 9, wherein the preset threshold is determined according to the size of the second reference display object frame. 如請求項9所述的資訊顯示系統,其中所述處理裝置計算所述第一參考顯示物件框的第一參考點與所述第二參考顯示物件框的第二參考點之間的所述距離,所述第一參考點與所述第二參考點為框中心點或框頂點。The information display system according to claim 9, wherein the processing means calculates the distance between the first reference point of the first reference display object frame and the second reference point of the second reference display object frame , the first reference point and the second reference point are frame center points or frame vertices. 如請求項1所述的資訊顯示系統,其中反應於所述顯示位置資訊位於所述顯示器的顯示邊緣區域,所述處理裝置往遠離所述顯示器之邊緣的方向調整所述顯示位置資訊。The information display system according to claim 1, wherein the processing device adjusts the display position information in a direction away from the edge of the display in response to the display position information being located in a display edge area of the display. 如請求項1所述的資訊顯示系統,其中所述實際顯示位置與所述顯示位置資訊為所述顯示器的畫素座標,所述處理裝置根據所述顯示器的尺寸與解析度獲取所述顯示位置資訊。The information display system according to claim 1, wherein the actual display position and the display position information are pixel coordinates of the display, and the processing device obtains the display position according to the size and resolution of the display Information. 一種資訊顯示方法,適用於具有可透光的顯示器、第一資訊擷取裝置、第二資訊擷取裝置以及處理裝置的資訊顯示系統,而所述方法包括: 利用所述第一資訊擷取裝置擷取一使用者的空間位置資訊; 利用所述第二資訊擷取裝置擷取一目標物的空間位置資訊; 根據所述使用者的空間位置資訊以及所述目標物的空間位置資訊決定所述目標物的一虛擬資訊於所述顯示器上的顯示位置資訊,其中所述顯示位置資訊包括對應於先前時間的第一顯示參考位置與對應至當前時間的第二顯示參考位置; 根據所述第一顯示參考位置與所述第二顯示參考位置之間的距離決定所述虛擬資訊於所述顯示器上對應於所述當前時間的實際顯示位置;以及 根據所述實際顯示位置顯示所述虛擬資訊於所述顯示器上。 An information display method is applicable to an information display system having a light-transmissive display, a first information capture device, a second information capture device, and a processing device, and the method includes: using the first information capturing device to capture spatial location information of a user; using the second information capturing device to capture spatial position information of a target; Determining display position information of virtual information of the target on the display according to the spatial position information of the user and the spatial position information of the target, wherein the display position information includes the first time corresponding to the previous time a display reference position and a second display reference position corresponding to the current time; determining an actual display position of the virtual information corresponding to the current time on the display according to a distance between the first display reference position and the second display reference position; and and displaying the virtual information on the display according to the actual display position. 如請求項14所述的資訊顯示方法,其中根據所述第一顯示參考位置與所述第二顯示參考位置之間的所述距離決定所述虛擬資訊在所述當前時間於所述顯示器上的所述實際顯示位置的步驟包括: 反應於所述第一顯示參考位置與所述第二顯示參考位置之間的距離大於一預設臨界值,根據對應至所述當前時間的所述第二顯示參考位置決定所述虛擬資訊於所述顯示器上對應於所述當前時間的所述實際顯示位置;以及 反應於所述第一顯示參考位置與所述第二顯示參考位置之間的距離未大於所述預設臨界值,根據對應於所述先前時間的所述第一顯示參考位置決定所述虛擬資訊於所述顯示器上對應於所述當前時間的所述實際顯示位置。 The information display method according to claim 14, wherein the position of the virtual information on the display at the current time is determined according to the distance between the first display reference position and the second display reference position The step of the actual display position comprises: Responding to the fact that the distance between the first display reference position and the second display reference position is greater than a preset threshold, determining the virtual information at the second display reference position corresponding to the current time said actual display position on said display corresponding to said current time; and In response to the fact that the distance between the first display reference position and the second display reference position is not greater than the predetermined threshold, the virtual information is determined according to the first display reference position corresponding to the previous time The actual display position corresponding to the current time on the display. 如請求項14所述的資訊顯示方法,其中所述方法更包括: 利用所述第一資訊擷取裝置擷取一使用者的姿態資訊, 其中根據所述使用者的空間位置資訊以及所述目標物的空間位置資訊決定所述目標物的所述虛擬資訊於所述顯示器上的所述顯示位置資訊的步驟包括: 根據所述使用者的姿態資訊判定所述使用者的頭部移動或轉動;以及 反應於根據所述使用者的姿態資訊判定所述使用者的頭部移動,根據所述使用者的空間位置資訊以及所述目標物的空間位置資訊決定所述目標物的所述虛擬資訊於所述顯示器上的所述顯示位置資訊,或反應於根據所述使用者的姿態資訊判定所述使用者的頭部轉動,根據所述使用者的空間位置資訊與姿態資訊以及所述目標物的空間位置資訊決定所述目標物的所述虛擬資訊於所述顯示器上的所述顯示位置資訊。 The information display method as described in claim 14, wherein the method further includes: using the first information capturing device to capture a user's posture information, The step of determining the display position information of the virtual information of the target on the display according to the spatial position information of the user and the spatial position information of the target includes: determining head movement or rotation of the user based on posture information of the user; and In response to determining the user's head movement according to the user's posture information, determining the virtual information of the target object based on the user's spatial position information and the target object's spatial position information The display position information on the display, or in response to determining the user’s head rotation based on the user’s posture information, according to the user’s spatial position information and posture information and the space of the target The position information determines the display position information of the virtual information of the target on the display. 如請求項14所述的資訊顯示方法,其中所述第一資訊擷取裝置包括用以拍攝所述使用者的第一影像感測器,且所述第二資訊擷取裝置包括用以拍攝所述目標物的第二影像感測器, 其中利用所述第二資訊擷取裝置擷取所述目標物的空間位置資訊的步驟包括: 利用所述第二資訊擷取裝置對多張目標物影像進行物件偵測而獲取所述目標物於各所述目標物影像中的物件邊界框,並根據所述物件邊界框進行座標轉換而獲取所述目標物的空間位置資訊, 其中利用所述第一資訊擷取裝置擷取所述使用者的空間位置資訊的步驟包括: 利用所述第一資訊擷取裝置對多張使用者影像進行人臉偵測而獲取各所述使用者影像中的人臉位置,並根據所述人臉位置進行座標轉換而獲取所述使用者的空間位置資訊。 The information display method as described in claim 14, wherein the first information capturing device includes a first image sensor for photographing the user, and the second information capturing device includes a sensor for photographing the user the second image sensor of the target object, Wherein, the step of using the second information capturing device to capture the spatial position information of the target includes: Using the second information acquisition device to perform object detection on multiple target images to obtain the object bounding box of the target in each of the target images, and perform coordinate transformation according to the object bounding box to obtain spatial location information of the target, Wherein, the step of using the first information retrieval device to retrieve the user's spatial location information includes: Use the first information capture device to perform face detection on multiple user images to obtain the face position in each user image, and perform coordinate transformation according to the face position to obtain the user The spatial location information of . 如請求項17所述的資訊顯示方法,其中所述第二資訊擷取裝置對多張目標物影像進行物件偵測而獲取各所述目標物於所述目標物影像中的一物件邊界框,所述顯示位置資訊包括於所述顯示器的顯示平面上對應於各所述物件邊界框的參考顯示物件框, 其中所述參考顯示物件框包括對應於所述先前時間的第一參考顯示物件框以及對應於所述當前時間的第二參考顯示物件框,所述第一顯示參考位置包括所述第一參考顯示物件框,而所述第二顯示參考位置包括所述第二參考顯示物件框, 其中根據所述第一顯示參考位置與所述第二顯示參考位置之間的所述距離決定所述虛擬資訊在所述當前時間於所述顯示器上的所述實際顯示位置的步驟包括: 比較對應於所述先前時間的所述第一參考顯示物件框以及對應於所述當前時間的所述第二參考顯示物件框之間的距離與一預設臨界值,以決定所述虛擬資訊於所述顯示器上對應於所述當前時間的實際顯示位置。 The information display method according to claim 17, wherein the second information capturing device performs object detection on multiple target images to obtain an object bounding box of each target in the target image, the display position information includes a reference display object frame corresponding to each of the object bounding boxes on the display plane of the display, Wherein the reference display object frame includes a first reference display object frame corresponding to the previous time and a second reference display object frame corresponding to the current time, and the first display reference position includes the first reference display object frame an object frame, and the second display reference location includes the second reference display object frame, The step of determining the actual display position of the virtual information on the display at the current time according to the distance between the first display reference position and the second display reference position comprises: comparing the distance between the first reference display object frame corresponding to the previous time and the second reference display object frame corresponding to the current time with a preset threshold value, so as to determine when the virtual information is The actual display position corresponding to the current time on the display. 如請求項14所述的資訊顯示方法,其中反應於所述顯示位置資訊位於所述顯示器的顯示邊緣區域,所述處理裝置往遠離所述顯示器之邊緣的方向調整所述顯示位置資訊。The information display method according to claim 14, wherein the processing device adjusts the display position information in a direction away from the edge of the display in response to the fact that the display position information is located in a display edge area of the display. 一種處理裝置,連接於可透光的顯示器、第一資訊擷取裝置以及第二資訊擷取裝置,其中所述第一資訊擷取裝置用以擷取一使用者的空間位置資訊,所述第二資訊擷取裝置用以擷取一目標物的空間位置資訊,該處理裝置包括: 記憶體,用以儲存資料;以及 處理器,連接所述記憶體並經配置: 自所述第一資訊擷取裝置取得所述使用者的空間位置資訊; 自所述第二資訊擷取裝置取得所述目標物的空間位置資訊; 根據所述使用者的空間位置資訊以及所述目標物的空間位置資訊決定所述目標物的一虛擬資訊於所述顯示器上的顯示位置資訊,其中所述顯示位置資訊包括對應於先前時間的第一顯示參考位置與對應至當前時間的第二顯示參考位置; 根據所述第一顯示參考位置與所述第二顯示參考位置之間的距離決定所述虛擬資訊於所述顯示器上對應於所述當前時間的實際顯示位置;以及 根據所述實際顯示位置顯示所述虛擬資訊於所述顯示器上。 A processing device connected to a light-transmitting display, a first information capturing device, and a second information capturing device, wherein the first information capturing device is used to capture spatial position information of a user, and the second information capturing device Two information retrieval devices are used to retrieve the spatial position information of a target, and the processing device includes: memory for storing data; and a processor coupled to said memory and configured to: obtaining spatial location information of the user from the first information retrieval device; Obtaining spatial position information of the target from the second information capturing device; Determining display position information of virtual information of the target on the display according to the spatial position information of the user and the spatial position information of the target, wherein the display position information includes the first time corresponding to the previous time a display reference position and a second display reference position corresponding to the current time; determining an actual display position of the virtual information corresponding to the current time on the display according to a distance between the first display reference position and the second display reference position; and and displaying the virtual information on the display according to the actual display position.
TW111128679A 2021-09-09 2022-07-29 Method, processing device, and display system for information display TWI844902B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202210997341.5A CN115794004A (en) 2021-09-09 2022-08-19 Information display method, processing device thereof and information display system
US17/939,900 US20230070721A1 (en) 2021-09-09 2022-09-07 Method, processing device, and display system for information display

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US202163242044P 2021-09-09 2021-09-09
US63/242,044 2021-09-09

Publications (2)

Publication Number Publication Date
TW202311817A true TW202311817A (en) 2023-03-16
TWI844902B TWI844902B (en) 2024-06-11

Family

ID=

Similar Documents

Publication Publication Date Title
TWI722280B (en) Controller tracking for multiple degrees of freedom
US11127380B2 (en) Content stabilization for head-mounted displays
US11436742B2 (en) Systems and methods for reducing a search area for identifying correspondences between images
US11568555B2 (en) Dense depth computations aided by sparse feature matching
US11451760B2 (en) Systems and methods for correcting rolling shutter artifacts
US11587255B1 (en) Collaborative augmented reality eyewear with ego motion alignment
US12020448B2 (en) Systems and methods for updating continuous image alignment of separate cameras
US11430086B2 (en) Upsampling low temporal resolution depth maps
WO2023072030A1 (en) Automatic focusing method and apparatus for lens, and electronic device and computer-readable storage medium
TWI844902B (en) Method, processing device, and display system for information display
TW202311817A (en) Method, processing device, and display system for information display
TWI792106B (en) Method, processing device, and display system for information display
US11450014B2 (en) Systems and methods for continuous image alignment of separate cameras
US11516452B2 (en) Systems and methods for temporal corrections for parallax reprojection
US20230070721A1 (en) Method, processing device, and display system for information display
TWI823740B (en) Active interactive navigation system and active interactive navigation method
US20120038750A1 (en) Apparatus and method for displaying three-dimensional (3d) object
TWI779332B (en) Augmented reality system and display method for anchor virtual object thereof
US20240221220A1 (en) Collaborative augmented reality eyewear with ego motion alignment