TWI842650B - System, method and computer program product for assisting multiple users to choice activity in virtual world - Google Patents

System, method and computer program product for assisting multiple users to choice activity in virtual world Download PDF

Info

Publication number
TWI842650B
TWI842650B TW112143065A TW112143065A TWI842650B TW I842650 B TWI842650 B TW I842650B TW 112143065 A TW112143065 A TW 112143065A TW 112143065 A TW112143065 A TW 112143065A TW I842650 B TWI842650 B TW I842650B
Authority
TW
Taiwan
Prior art keywords
activity
action
semantics
virtual world
virtual
Prior art date
Application number
TW112143065A
Other languages
Chinese (zh)
Inventor
張勝仁
楊仕丞
Original Assignee
中華電信股份有限公司
Filing date
Publication date
Application filed by 中華電信股份有限公司 filed Critical 中華電信股份有限公司
Application granted granted Critical
Publication of TWI842650B publication Critical patent/TWI842650B/en

Links

Images

Abstract

A system, a method and a computer program product for assisting multiple users to choice an activity in a virtual world are provided. Speeches and motions of a plurality of VR devices accessing a virtual world are detected. a semantic of a speech is recognized by using an AI semantic model, and a motion is recognized by using an AI motion model. A recognition result of the semantics, a recognition result of the motion, and an association between the recognition result of the semantics and the recognition result of the motion are analyzed to produce an analyzation result. A recommended activity is determined from an activity database according to the analyzation result. A virtual image assistant in the virtual world provides the recommended activity, thereby transferring the VR devices to a scene of the recommended activity in the virtual world.

Description

在虛擬世界中協助多人選擇活動之系統、方法及電腦程式產品 Systems, methods and computer program products for assisting multiple people in selecting activities in a virtual world

本案係關於虛擬實境之領域,詳而言之,係關於一種在虛擬世界中協助多人選擇活動之系統、方法及電腦程式產品。 This case is related to the field of virtual reality, and more specifically, to a system, method and computer program product for assisting multiple people in selecting activities in a virtual world.

在元宇宙虛擬世界的應用中,當同一團體多人以虛擬實境(virtual reality,VR)裝置進入虛擬世界(例如元宇宙平台)時,因為元宇宙提供的虛擬活動太多,而在多人想要一起到同一活動會場同樂的情況下,人多意見也很多,故難以決定出適合此團體的活動。 In the application of the Metaverse virtual world, when multiple people in the same group enter the virtual world (such as the Metaverse platform) with a virtual reality (VR) device, because the Metaverse provides too many virtual activities, and when multiple people want to have fun together at the same event venue, there are many opinions, so it is difficult to decide on an activity suitable for this group.

目前技術中,雖有協助推薦適合活動,然,通常都是罐頭式回答,或預設好題目讓使用者選擇。另外,因為元宇宙虛擬世界是以數位的方式存在於雲端,所以無論是活動或是場景都可以無限延伸,會有很多種類的活動,而目前雖有提供參與者以語音方式詢問的技術,然,通常都只是擷取已定義好的關鍵字進行搜尋,此方法無法精準、快速或確實推薦最適合的活動。 Although current technology can help recommend suitable activities, it usually provides canned answers or preset questions for users to choose from. In addition, because the virtual world of the Metaverse exists digitally in the cloud, both activities and scenes can be infinitely extended, and there will be many types of activities. Although there is technology that allows participants to ask questions by voice, it usually only captures defined keywords for search. This method cannot accurately, quickly or accurately recommend the most suitable activities.

因此,如何在虛擬世界中協助多人討論以精準、快速或確實地推薦適合活動,為本領域待解決之問題。 Therefore, how to assist multiple people in discussing in the virtual world to accurately, quickly or accurately recommend suitable activities is a problem to be solved in this field.

為解決上述問題及其他問題,本案揭示一種在虛擬世界中協助多人選擇活動之系統、方法、電腦程式產品、電腦可讀取記錄媒體。 To solve the above problems and other problems, this case discloses a system, method, computer program product, and computer-readable recording medium for assisting multiple people in selecting activities in a virtual world.

本案所揭之在虛擬世界中協助多人選擇活動之系統,係包括:虛擬形象助理模組,用於顯示虛擬形象助理於虛擬世界中;語音及語意偵測模組,用於偵測進入該虛擬世界的多個虛擬實境裝置之語音及該語音的語意;手勢及動作偵測模組,用於偵測進入該虛擬世界的該多個虛擬實境裝置之手勢及動作;AI辨識模組,用於辨識該語音、該語意、該手勢、該動作,以判斷該語音的辨識結果是否為開啟協助語音,利用AI語意模型判斷該語意的辨識結果是否為問題語意,利用AI動作模型判斷該手勢的辨識結果是否為開啟協助手勢,及利用該AI動作模型判斷該動作的辨識結果是否為贊同動作;活動分析模組,用於分析該AI語意模型之該語意的辨識結果、該AI動作模型之該動作的辨識結果、及該語意的辨識結果與該動作的辨識結果之間的關聯性;以及活動決定模組,用於根據該活動分析模組之分析結果自活動資料庫中決定出推薦活動,以由顯示於該虛擬世界中之該虛擬形象助理提供該推薦活動。 The system disclosed in this case for assisting multiple people in selecting activities in a virtual world includes: a virtual image assistant module for displaying a virtual image assistant in the virtual world; a voice and semantic detection module for detecting the voices of multiple virtual reality devices entering the virtual world and the semantics of the voices; a gesture and action detection module for detecting the gestures and actions of the multiple virtual reality devices entering the virtual world; an AI recognition module for recognizing the voice, the semantics, the gestures, and the actions to determine whether the recognition result of the voice is an assistive voice, and using an AI semantic model to determine whether the recognition result of the semantics is an assistive voice. For the question semantics, the AI action model is used to determine whether the recognition result of the gesture is to open the collaborative gesture, and the AI action model is used to determine whether the recognition result of the action is an approval action; the activity analysis module is used to analyze the recognition result of the semantics of the AI semantic model, the recognition result of the action of the AI action model, and the correlation between the recognition result of the semantics and the recognition result of the action; and the activity determination module is used to determine the recommended activity from the activity database according to the analysis result of the activity analysis module, so that the virtual image assistant displayed in the virtual world provides the recommended activity.

本案所揭之在虛擬世界中協助多人選擇活動方法,係包括:偵測進入虛擬世界的多個虛擬實境裝置之語音及該語音的語意;偵測進入該虛擬世界之該多個虛擬實境裝置之手勢及動作;辨識該語音及該語意,以判斷該語音的辨識結果是否為開啟協助語音,及利用AI語意模型判斷該語意的辨識結果是否 為問題語意;利用AI動作模型辨識該手勢及動作,以利用AI動作模型判斷該手勢的辨識結果是否為開啟協助手勢,及利用該AI動作模型判斷該動作的辨識結果是否為贊同動作;分析該AI語意模型之該語意的辨識結果、該AI動作模型之該動作的辨識結果、及該語意的辨識結果與該動作的辨識結果之間的關聯性,以產生分析結果;以及根據該分析結果自活動資料庫中決定出推薦活動,以由虛擬形象助理在該虛擬世界中提供該推薦活動。 The method disclosed in this case for assisting multiple people in selecting activities in a virtual world includes: detecting the voices of multiple virtual reality devices entering the virtual world and the meaning of the voices; detecting the gestures and actions of the multiple virtual reality devices entering the virtual world; recognizing the voices and the meanings to determine whether the recognition result of the voices is an assist voice, and using an AI semantic model to determine whether the recognition result of the meanings is a question meaning; using an AI action model to recognize the gestures and actions, and using the AI action model to determine whether the recognition result of the meanings is a question meaning; Determine whether the recognition result of the gesture is a collaborative gesture, and use the AI action model to determine whether the recognition result of the action is an approval action; analyze the recognition result of the semantics of the AI semantic model, the recognition result of the action of the AI action model, and the correlation between the recognition result of the semantics and the recognition result of the action to generate an analysis result; and determine a recommended activity from the activity database based on the analysis result, so that the virtual image assistant provides the recommended activity in the virtual world.

本案所揭之在虛擬世界中協助多人選擇活動之電腦程式產品,經由電腦載入程式以執行本案所揭之在虛擬世界中協助多人選擇活動之方法。 The computer program product disclosed in this case assists multiple people in selecting activities in a virtual world, and the method disclosed in this case assists multiple people in selecting activities in a virtual world by loading the program into a computer.

本案所揭之在虛擬世界中協助多人選擇活動之電腦可讀取記錄媒體,儲存有指令,並可利用計算設備或電腦透過處理器及/或記憶體執行電腦可讀取記錄媒體,以於執行電腦可讀取記錄媒體時執行本案所揭之在虛擬世界中協助多人選擇活動之方法。 The computer-readable recording medium disclosed in this case for assisting multiple people in selecting activities in a virtual world stores instructions, and can use a computing device or a computer to execute the computer-readable recording medium through a processor and/or a memory, so as to execute the method for assisting multiple people in selecting activities in a virtual world disclosed in this case when executing the computer-readable recording medium.

於一實施例中,該AI語意模型係利用與活動有關的問答集作為訓練資料進行訓練,以於該語意的辨識結果為問題語意時,該AI語意模型係產生對應的回答語意,而由該虛擬形象助理在該虛擬世界中根據該回答語意透過語音、圖片、影片、或活動現場觀賞窗來提供回答。 In one embodiment, the AI semantic model is trained using a question-and-answer set related to an activity as training data, so that when the semantic recognition result is the question semantics, the AI semantic model generates the corresponding answer semantics, and the virtual image assistant provides answers in the virtual world based on the answer semantics through voice, pictures, videos, or event viewing windows.

於一實施例中,該AI動作模型係利用虛擬實境裝置動作資料作為訓練資料進行訓練,以於該虛擬形象助理提供該回答時,該AI動作模型辨識各該虛擬實境裝置對於該回答之反應動作且該AI語意模型辨識各該虛擬實境裝置對於該回答之反應語意,而由該AI辨識模組記錄各該反應動作的辨識結果、提供各該反應動作之虛擬實境裝置的數量、各該反應語意的辨識結果、提供各該反 應語意之虛擬實境裝置的數量、各該反應動作的辨識結果與各該反應語意的辨識結果之間的關聯性,以供該活動分析模組進行分析。 In one embodiment, the AI action model is trained using the action data of the virtual reality device as training data, so that when the virtual image assistant provides the answer, the AI action model recognizes the reaction action of each virtual reality device to the answer and the AI semantic model recognizes the reaction semantics of each virtual reality device to the answer, and the AI action model recognizes the reaction action of each virtual reality device to the answer. The recognition module records the recognition results of each reaction action, the number of virtual reality devices that provide each reaction action, the recognition results of each reaction semantics, the number of virtual reality devices that provide each reaction semantics, and the correlation between the recognition results of each reaction action and the recognition results of each reaction semantics, so as to provide the activity analysis module for analysis.

於一實施例中,該AI動作模型係利用各種手勢圖像作為訓練資料進行訓練,以於該手勢及動作偵測模組偵測到該多個虛擬實境裝置的手勢且該AI動作模型對該手勢的辨識結果為開啟協助手勢或者於該語音及語意偵測模組偵測到該多個虛擬實境裝置的語意且該AI語意模型對該語意的辨識結果為開啟協助語意時,令該虛擬形象助理出現在該虛擬世界中。 In one embodiment, the AI action model is trained using various gesture images as training data, so that when the gesture and action detection module detects the gestures of the multiple virtual reality devices and the AI action model recognizes the gesture as turning on the assist gesture, or when the voice and semantic detection module detects the semantics of the multiple virtual reality devices and the AI semantic model recognizes the semantics as turning on the assist semantic, the virtual image assistant appears in the virtual world.

於一實施例中,本案所揭之在虛擬世界中協助多人選擇活動方法更包括活動控制模組,其中,若該活動決定模組所決定之推薦活動在該虛擬世界正在進行中,則該活動控制模組將該多個虛擬實境裝置轉移進入該虛擬世界的該推薦活動中,而若該活動決定模組所決定之推薦活動在該虛擬世界不在進行中,則該活動控制模組將該推薦活動加入行事曆,待該推薦活動開始前,再通知該多個虛擬實境裝置。 In one embodiment, the method for assisting multiple people in selecting activities in a virtual world disclosed in this case further includes an activity control module, wherein if the recommended activity determined by the activity determination module is ongoing in the virtual world, the activity control module transfers the multiple virtual reality devices into the recommended activity in the virtual world, and if the recommended activity determined by the activity determination module is not ongoing in the virtual world, the activity control module adds the recommended activity to the calendar, and notifies the multiple virtual reality devices before the recommended activity starts.

於一實施例中,本案所揭之在虛擬世界中協助多人選擇活動方法更包括活動預覽模組,係提供活動現場觀賞窗以播放該推薦活動或其他活動,以於偵測到該虛擬實境裝置在該虛擬世界中的虛擬化身撞擊或接觸該活動現場觀賞窗時,允許對應該虛擬化身的虛擬實境裝置轉移進入該虛擬世界中的該推薦活動或該其他活動。 In one embodiment, the method for assisting multiple people in selecting activities in a virtual world disclosed in this case further includes an activity preview module, which provides an activity live viewing window to play the recommended activity or other activities, so that when it is detected that the virtual reality device in the virtual world hits or touches the virtual reality device in the virtual world, the virtual reality device corresponding to the virtual reality device is allowed to transfer to the recommended activity or the other activity in the virtual world.

藉由本案所揭之在虛擬世界中協助多人選擇活動之系統、方法、電腦程式產品、電腦可讀取記錄媒體,提供虛擬化的智慧助理以協助多參與者進行活動討論,並記錄下活動討論中多個參與者透過VR裝置的語音或動作來計算 分析再至活動資料庫中搜索,以決定出推薦活動,並將VR裝置的參與者移動到該推薦活動的會場,以讓沉浸在虛擬世界活動時能更為方便自在。 Through the system, method, computer program product, and computer-readable recording medium disclosed in this case for assisting multiple people in selecting activities in a virtual world, a virtualized intelligent assistant is provided to assist multiple participants in discussing activities, and the voice or actions of multiple participants in the activity discussion are recorded for calculation and analysis through VR devices, and then searched in the activity database to determine recommended activities, and the participants of the VR device are moved to the venue of the recommended activity, so that it is more convenient and comfortable to immerse in virtual world activities.

11、12:VR裝置 11, 12: VR device

20:系統 20: System

21:虛擬形象助理模組 21: Avatar Assistant Module

22:語音及語意偵測模組 22: Voice and semantic detection module

23:手勢及動作偵測模組 23: Gesture and motion detection module

24:AI辨識模組 24: AI recognition module

241:AI語意模型 241:AI semantic model

242:AI動作模型 242: AI action model

25:活動分析模組 25: Activity analysis module

26:活動決定模組 26: Activity decision module

27:活動控制模組 27: Activity control module

28:活動預覽模組 28: Event preview module

30:活動資料庫 30:Activity database

S201~S207:步驟 S201~S207: Steps

S301~S303:步驟 S301~S303: Steps

S401~S405:步驟 S401~S405: Steps

S501~S504:步驟 S501~S504: Steps

S601~S605:步驟 S601~S605: Steps

圖1係為本案之在虛擬世界中協助多人選擇活動之系統之架構示意圖。 Figure 1 is a schematic diagram of the architecture of the system that assists multiple people in selecting activities in the virtual world.

圖2係為本案之在虛擬世界中協助多人選擇活動之方法之步驟示意圖。 Figure 2 is a schematic diagram of the steps of the method for assisting multiple people in selecting activities in a virtual world.

圖3係為本案之在虛擬世界中協助多人選擇活動之方法中開啟協助之步驟示意圖。 FIG3 is a schematic diagram of the steps of enabling assistance in the method of assisting multiple people to select activities in a virtual world in this case.

圖4係為本案之在虛擬世界中協助多人選擇活動之方法中討論活動之步驟示意圖。 Figure 4 is a schematic diagram of the steps of discussing activities in the method of assisting multiple people in selecting activities in a virtual world in this case.

圖5係為本案之在虛擬世界中協助多人選擇活動之方法中終止問答後之步驟示意圖。 FIG5 is a schematic diagram of the steps after the question and answer session is terminated in the method of assisting multiple people in selecting activities in a virtual world in this case.

圖6係為本案之在虛擬世界中協助多人選擇活動之方法中轉移至推薦活動之步驟示意圖。 FIG6 is a schematic diagram of the steps of transitioning to recommending activities in the method of assisting multiple people in selecting activities in a virtual world in this case.

以下藉由特定的實施例說明本案之實施方式,熟習此項技藝之人士可由本文所揭示之內容輕易地瞭解本案之其他優點及功效。本說明書所附圖式所繪示之結構、比值、大小等均僅用於配合說明書所揭示之內容,以供熟悉此技藝之人士之瞭解與閱讀,非用於限定本案可實施之限定條件,故任何修飾、改 變或調整,在不影響本案所能產生之功效及所能達成之目的下,均應仍落在本案所揭示之技術內容得能涵蓋之範圍內。 The following specific examples are used to illustrate the implementation of this case. People familiar with this technology can easily understand the other advantages and effects of this case from the content disclosed in this article. The structures, ratios, sizes, etc. shown in the attached figures of this manual are only used to match the content disclosed in the manual for people familiar with this technology to understand and read, and are not used to limit the conditions under which this case can be implemented. Therefore, any modification, change or adjustment should still fall within the scope of the technical content disclosed in this case without affecting the effects and purposes that can be achieved by this case.

於本文中所用之術語「包括」、「包含」、「具有」、「含有」或其任何其他變體都旨在涵蓋非排他性的包含。除非另有說明,單數形式的措辭,如「一」、「一個」、「該」也適用於複數形式,而「或」、「及/或」等措辭可互換使用。 As used herein, the terms "include", "comprising", "having", "containing" or any other variations thereof are intended to cover a non-exclusive inclusion. Unless otherwise indicated, singular forms such as "a", "an", "the" may also be used in the plural, and "or", "and/or" and the like may be used interchangeably.

請參閱圖1,係為本案之在虛擬世界中協助多人選擇活動之系統之架構示意圖,系統20包括虛擬形象助理模組21、語音及語意偵測模組22、手勢及動作偵測模組23、人工智慧(artificial intelligence,AI)辨識模組24、活動分析模組25、活動決定模組26、活動控制模組27、活動預覽模組28。 Please refer to Figure 1, which is a schematic diagram of the architecture of the system for assisting multiple people in selecting activities in the virtual world. The system 20 includes a virtual image assistant module 21, a voice and semantic detection module 22, a gesture and action detection module 23, an artificial intelligence (AI) recognition module 24, an activity analysis module 25, an activity decision module 26, an activity control module 27, and an activity preview module 28.

系統20應用於進入虛擬世界的虛擬實境裝置(即VR裝置)11、12,並連接活動資料庫30以搜尋一推薦活動。VR裝置11、12在本文中可指具有VR APP的手機或可載入VR APP以呈現VR畫面之裝置、VR穿戴裝置、VR穿戴裝置配合手持控制器、VR穿戴裝置配合內部定位、或VR穿戴裝置配合外部定位等。另外,系統20可例如為伺服器,其中之各模組或單元均可為軟體、硬體或韌體;若為硬體,則可為具有資料處理與運算能力之處理單元、處理器、或電腦主機;若為軟體或韌體,則可包括處理單元、處理器、電腦或電腦主機可執行之指令,且可安裝於同一硬體裝置或分布於不同的複數硬體裝置。 The system 20 is applied to a virtual reality device (i.e., VR device) 11, 12 that enters a virtual world and connects to an activity database 30 to search for a recommended activity. The VR devices 11, 12 herein may refer to a mobile phone with a VR APP or a device that can load a VR APP to present a VR image, a VR wearable device, a VR wearable device with a handheld controller, a VR wearable device with an internal positioning device, or a VR wearable device with an external positioning device, etc. In addition, the system 20 may be, for example, a server, in which each module or unit may be software, hardware or firmware; if it is hardware, it may be a processing unit, processor, or computer host with data processing and computing capabilities; if it is software or firmware, it may include instructions executable by a processing unit, processor, computer or computer host, and may be installed on the same hardware device or distributed on different multiple hardware devices.

虛擬形象助理模組21用於將虛擬形象助理出現在虛擬世界中,或顯示虛擬形象助理於虛擬世界中。例如,當VR裝置11、12之在虛擬世界中的虛擬化身說話時,該虛擬形象助理會將頭朝向該虛擬化身以呈現傾聽的姿態,而當該虛擬形象助理回答時,也會面向提出問題的虛擬化身來回答問題。 The virtual image assistant module 21 is used to make the virtual image assistant appear in the virtual world, or to display the virtual image assistant in the virtual world. For example, when the virtual avatar of the VR device 11, 12 in the virtual world speaks, the virtual image assistant will turn its head toward the virtual avatar to present a listening posture, and when the virtual image assistant answers, it will also answer the question facing the virtual avatar that asked the question.

語音及語意偵測模組22用於偵測進入該虛擬世界的VR裝置11、12之語音及該語音的語意。換言之,當偵測到虛擬世界中有虛擬化身發出語音時,語音及語意偵測模組22可將所偵測到的語音轉換成語意文字,以供AI辨識模組24進行處理。於一實施例中,將語音轉換成語意的技術例如為自然語言分析技術,透過深度學習語意分析技術可理解用戶多樣化、口語化長句之表達用語,例如可自VR裝置11、12之語音理解到與協助參加活動相關的語意,接著再將語意文字提供至AI辨識模組24將其輸入至AI語意模型241。 The voice and semantic detection module 22 is used to detect the voice and the semantics of the VR devices 11 and 12 entering the virtual world. In other words, when a virtual avatar in the virtual world is detected to speak, the voice and semantic detection module 22 can convert the detected voice into semantic text for processing by the AI recognition module 24. In one embodiment, the technology for converting speech into semantics is, for example, natural language analysis technology. Through deep learning semantic analysis technology, the user's diverse, colloquial long sentences can be understood. For example, the semantics related to assisting in participating in activities can be understood from the speech of VR devices 11 and 12, and then the semantic text is provided to the AI recognition module 24 to input it into the AI semantic model 241.

手勢及動作偵測模組23用於偵測進入該虛擬世界的VR裝置11、12之手勢及動作。換言之,VR裝置11、12可將其手勢及動作等影像或資料傳至系統20,而系統20再傳至AI辨識模組24進行處理。 The gesture and action detection module 23 is used to detect the gestures and actions of the VR devices 11 and 12 entering the virtual world. In other words, the VR devices 11 and 12 can transmit images or data such as gestures and actions to the system 20, and the system 20 transmits them to the AI recognition module 24 for processing.

AI辨識模組24用於辨識來自語音及語意偵測模組22的語音及語意以及辨識來自手勢及動作偵測模組23的手勢及動作,以利用AI語意模型241判斷該語意的辨識結果是否為問題語意或終止問答語意,以及利用AI動作模型242判斷該手勢的辨識結果是否為開啟協助手勢或結束協助手勢,及利用AI動作模型242判斷該動作的辨識結果是否為贊同動作。於一實施例中,AI辨識模組24根據語音及語意偵測模組22所提供之語意,分析出與活動有關的關鍵字以成為活動關鍵字,供後續活動分析模組25進行分析。 The AI recognition module 24 is used to recognize the voice and semantics from the voice and semantic detection module 22 and recognize the gestures and actions from the gesture and action detection module 23, so as to use the AI semantic model 241 to determine whether the recognition result of the semantics is a question semantics or a termination question and answer semantics, and use the AI action model 242 to determine whether the recognition result of the gesture is to open the collaborative gesture or to end the collaborative gesture, and use the AI action model 242 to determine whether the recognition result of the action is an approval action. In one embodiment, the AI recognition module 24 analyzes the activity-related keywords based on the semantics provided by the voice and semantic detection module 22 to become activity keywords for subsequent analysis by the activity analysis module 25.

於一實施例中,AI語意模型241可利用與活動有關的問答集作為訓練資料進行訓練。於一實施例中,該AI動作模型242可利用虛擬實境裝置動作資料作為訓練資料進行訓練。於一實施例中,AI動作模型242可利用各種手勢圖像作為訓練資料進行訓練。 In one embodiment, the AI semantic model 241 can be trained using a question-and-answer set related to an activity as training data. In one embodiment, the AI action model 242 can be trained using virtual reality device action data as training data. In one embodiment, the AI action model 242 can be trained using various gesture images as training data.

詳言之,當手勢及動作偵測模組23偵測到VR裝置11、12的手勢且AI動作模型242對該手勢的辨識結果為開啟協助手勢時,令虛擬形象助理模組21顯示虛擬形象助理於該虛擬世界中,而辨識到結束協助手勢時,虛擬形象助理不再顯示於虛擬世界中。另外,當語音及語意偵測模組22偵測到VR裝置11、12的語音,然後AI辨識模組24的辨識結果為開啟協助語音時,令虛擬形象助理模組21顯示虛擬形象助理於該虛擬世界中,而辨識到結束協助語音時,虛擬形象助理不再顯示於虛擬世界中。換言之,在該虛擬世界中可透過語音或手勢的方式呼叫該虛擬形象助理。 In detail, when the gesture and action detection module 23 detects the gesture of the VR devices 11 and 12 and the AI action model 242 recognizes the gesture as turning on the collaboration gesture, the virtual image assistant module 21 displays the virtual image assistant in the virtual world. When the collaboration gesture is recognized to end, the virtual image assistant is no longer displayed in the virtual world. In addition, when the voice and semantic detection module 22 detects the voice of the VR device 11, 12, and the AI recognition module 24 recognizes that the assisting voice is turned on, the virtual image assistant module 21 displays the virtual image assistant in the virtual world, and when the assisting voice is recognized to be ended, the virtual image assistant is no longer displayed in the virtual world. In other words, the virtual image assistant can be called by voice or gesture in the virtual world.

接著,當AI語意模型241之語意的辨識結果為問題語意時,AI語意模型241產生對應的回答語意,而由該虛擬形象助理在該虛擬世界中根據該回答語意透過語音、圖片、影片、或活動現場觀賞窗來提供回答。另外,於該虛擬形象助理提供該回答時,AI動作模型242可辨識VR裝置11、12對於該回答之反應動作且AI語意模型241可辨識VR裝置11、12對於該回答之反應語意,或者當任一VR裝置11、12提出問題時,AI動作模型242可辨識其他的VR裝置11、12對於該問題之反應動作且AI語意模型241可辨識其他的VR裝置11、12對於該問題之反應語意。接著,AI辨識模組24記錄各該反應動作的辨識結果、提供各該反應動作之VR裝置11、12的數量、各該反應語意的辨識結果、各提供該反應語意之VR裝置11、12的數量、以及各該反應動作的辨識結果與各該反應語意的辨識結果之間的關聯性等。例如:將辨識結果為贊同或認可的動作或手勢,與當下虛擬世界出現的語音(無論是問題、答案、或是討論等)關聯起來。 Then, when the semantics of the AI semantic model 241 is recognized as the question semantics, the AI semantic model 241 generates the corresponding answer semantics, and the virtual image assistant provides an answer in the virtual world based on the answer semantics through voice, pictures, videos, or an event live viewing window. In addition, when the virtual image assistant provides the answer, the AI action model 242 can recognize the reaction action of the VR devices 11 and 12 to the answer and the AI semantic model 241 can recognize the reaction semantics of the VR devices 11 and 12 to the answer, or when any VR device 11 and 12 asks a question, the AI action model 242 can recognize the reaction action of the other VR devices 11 and 12 to the question and the AI semantic model 241 can recognize the reaction semantics of the other VR devices 11 and 12 to the question. Next, the AI recognition module 24 records the recognition results of each reaction action, the number of VR devices 11, 12 providing each reaction action, the recognition results of each reaction semantics, the number of VR devices 11, 12 providing each reaction semantics, and the correlation between the recognition results of each reaction action and the recognition results of each reaction semantics. For example, the action or gesture with the recognition result of approval or recognition is associated with the voice (whether it is a question, answer, or discussion, etc.) appearing in the virtual world at the moment.

再而,當AI語意模型241之語意的辨識結果為終止問答語意時,AI模型24將上述所記錄下的所有資料,也就是在終止問答語意之前,所有關於活動討論之語意和動作(或手勢)的辨識結果、彼此之間的關聯性、提供語意和動作(或手勢)的時間、語意和動作(或手勢)的VR裝置等資料,傳至活動分析模組25。在此,語意和動作(或手勢)包括各種關於問題、答案或反應之語意和動作(或手勢)。 Furthermore, when the semantics recognition result of the AI semantics model 241 is the termination of the question-and-answer semantics, the AI model 24 transmits all the data recorded above, that is, the recognition results of all the semantics and actions (or gestures) discussed about the activity before the termination of the question-and-answer semantics, the correlation between each other, the time of providing the semantics and actions (or gestures), the VR device of the semantics and actions (or gestures), etc., to the activity analysis module 25. Here, the semantics and actions (or gestures) include various semantics and actions (or gestures) about questions, answers or reactions.

活動分析模組25用於分析AI語意模型241之語意的辨識結果、AI動作模型242之動作的辨識結果、及語意的辨識結果與動作的辨識結果之間的關聯性。詳言之,活動分析模組25將本次活動討論所記錄下的所有資料,包括語意(例如問題語意、回答語意)和動作或手勢(例如贊同動作、手勢動作)的辨識結果、對於問題語意或回答語意所反應出的語意或動作或手勢的辨識結果、彼此之間的關聯性、提供語意和動作或手勢的時間、語意和動作或手勢的VR裝置,進行分析計算,以產生與活動有關的活動關鍵字與加權重資料,再組合成加權重資料篩選計算式。於一實施例中,活動分析模組25紀錄所有參與活動的VR裝置使用者參與討論的語意,並透過AI語意分析技術篩選出與活動相關的語詞為活動關鍵字,活動關鍵字為後續到活動資料庫30作搜尋的主要輸入值。活動分析模組25依每個活動關鍵字出現在討論中的次數,與當下記錄到該活動關鍵字時紀錄其他人透過VR裝置的動作判斷是否有贊同或認可的動作,辨識到每個動作定義的加權分數進行加權計算。 The activity analysis module 25 is used to analyze the semantic recognition results of the AI semantic model 241, the action recognition results of the AI action model 242, and the correlation between the semantic recognition results and the action recognition results. In detail, the activity analysis module 25 analyzes and calculates all the data recorded in the discussion of this activity, including the recognition results of semantics (such as question semantics, answer semantics) and actions or gestures (such as approval actions, hand gestures), the recognition results of semantics or actions or gestures reflected in the question semantics or answer semantics, the correlation between them, the time of providing semantics and actions or gestures, and the VR device of semantics and actions or gestures, to generate activity keywords and weighted data related to the activity, and then combine them into a weighted data screening formula. In one embodiment, the activity analysis module 25 records the semantics of all VR device users participating in the activity and selects activity-related words as activity keywords through AI semantic analysis technology. The activity keywords are the main input values for subsequent searches in the activity database 30. The activity analysis module 25 determines whether there are actions of approval or recognition of other people through VR devices when the activity keyword is recorded, and identifies the weighted score defined for each action for weighted calculation.

活動決定模組26用於根據活動分析模組25之分析結果自活動資料庫30中決定出推薦活動,以由該虛擬形象助理在該虛擬世界中提供該推薦活動。詳言之,活動決定模組26將活動分析模組25所產生之活動關鍵字與加權重 資料,轉換成搜尋語法以至活動資料庫30進行搜尋,最後搜尋出的活動依加權平均數演算法得出平均分數最高的活動,即為本次活動討論的推薦活動。於一實施例中,活動分析模組25將與討論活動時紀錄有關的活動關鍵字與根據活動關鍵字出現次數與當下記錄到的有贊同動作加權重資料,再組合成加權重資料篩選計算式(例如有記錄到活關鍵字「演唱會」,並記錄到五次(一次一分),其中五次有三次記錄到其他人有比讚的動作(比讚的動作+2分,所以2x3=6,加6分),則利用「演唱會」搜尋出來的活動加權11分),活動決定模組26再轉換成搜尋語法,至活動資料庫30進行搜尋,藉以最後搜尋出依加權平均數演算法得出平均分數最高的活動,即為本次活動討論的推薦活動。 The activity decision module 26 is used to determine the recommended activity from the activity database 30 according to the analysis result of the activity analysis module 25, so that the virtual image assistant can provide the recommended activity in the virtual world. In detail, the activity decision module 26 converts the activity keywords and weighted data generated by the activity analysis module 25 into a search syntax to search the activity database 30. Finally, the activity with the highest average score obtained by the weighted average algorithm is the recommended activity for this activity discussion. In one embodiment, the activity analysis module 25 combines the activity keywords related to the activity discussion with the weighted data based on the number of occurrences of the activity keywords and the currently recorded like actions into a weighted data screening formula (for example, if the activity keyword "concert" is recorded five times (one point per time), and three of the five times other people have like actions recorded (like actions +2 points, so 2x3=6, plus 6 points), then the activities searched using "concert" are weighted 11 points), and the activity decision module 26 then converts it into a search syntax and searches the activity database 30, so as to finally search out the activity with the highest average score according to the weighted average algorithm, which is the recommended activity for this activity discussion.

此外,若活動決定模組26所決定之推薦活動在該虛擬世界正在進行中,則活動控制模組27將VR裝置11、12轉移進入該虛擬世界的該推薦活動中;若活動決定模組26所決定之推薦活動在該虛擬世界不在進行中,則活動控制模組27將該推薦活動加入行事曆,待該推薦活動開始前,再通知VR裝置11、12。於一實施例中,活動控制模組27可於推薦活動的前一天或前30分鐘發簡訊通知參與者,而簡訊內容可包括活動會場連結,VR裝置11、12的參與者點選該活動會場連結即可直接加入該活動,而參與者點擊連結加入該活動後,活動控制模組27會通知當時一同參與討論並已在活動會場的其他參與者。 In addition, if the recommended activity determined by the activity determination module 26 is ongoing in the virtual world, the activity control module 27 transfers the VR devices 11 and 12 to the recommended activity in the virtual world; if the recommended activity determined by the activity determination module 26 is not ongoing in the virtual world, the activity control module 27 adds the recommended activity to the calendar and notifies the VR devices 11 and 12 before the recommended activity starts. In one embodiment, the activity control module 27 may send a text message to the participants one day or 30 minutes before the recommended activity, and the text message content may include a link to the activity venue. The participants of the VR devices 11 and 12 can directly join the activity by clicking on the link to join the activity. After the participants click on the link to join the activity, the activity control module 27 will notify other participants who participated in the discussion and were already at the activity venue at that time.

活動預覽模組28可提供活動現場觀賞窗以播放一活動,更於偵測到VR裝置11、12在虛擬世界中的虛擬化身撞擊或接觸該活動現場觀賞窗時,允許對應該虛擬化身的VR裝置11、12轉移進入該虛擬世界中的活動。另外,活動預覽模組28除了利用活動現場觀賞窗撥放正在進行的活動之外,當同一時間有其他要參加或推薦的活動也在進行中時,活動預覽模組28可利用另一活動 現場觀賞窗來播放此另一活動,而參與者可透過呼叫而轉換到該另一活動現場觀賞窗,當虛擬化身撞擊該一活動現場觀賞窗即可瞬間轉移到該另一活動的現場。 The activity preview module 28 can provide an activity live viewing window to play an activity, and when it is detected that the virtual avatar of the VR device 11, 12 in the virtual world hits or touches the activity live viewing window, the VR device 11, 12 corresponding to the virtual avatar is allowed to transfer to the activity in the virtual world. In addition, in addition to using the event live viewing window to play the ongoing event, the event preview module 28 can also use another event live viewing window to play the other event when there are other events to be participated in or recommended at the same time. Participants can switch to the other event live viewing window through calling, and when the virtual avatar hits the event live viewing window, it can be instantly transferred to the scene of the other event.

藉由圖1所示之本案之在虛擬世界中協助多人選擇活動之系統的一或多個實施例可知,一群人戴上VR裝置11、12透過進入元宇宙的虛擬世界後,想在同一場景同樂,發現虛擬世界的活動種類很多,而群體人多意見也多,此時可呼叫虛擬形象助理出來協助,經由與虛擬形象助理之間的問答互動,同時系統20也會根據問題與回答的內容以及進行問答時多人的反應語音和反應動作,最後再依所記錄或儲存的資料計算出篩選活動條件式,至活動資料庫30進行搜尋,藉以產生一推薦活動,並協助多人一同參與該推薦活動。 According to one or more embodiments of the system for assisting multiple people in selecting activities in a virtual world shown in FIG. 1 , a group of people wearing VR devices 11 and 12 enter the virtual world of the metaverse and want to have fun together in the same scene. They find that there are many types of activities in the virtual world and there are many opinions because there are many people in the group. At this time, they can call a virtual image assistant to assist. Through the question-and-answer interaction with the virtual image assistant, the system 20 will also calculate the screening activity conditional formula based on the content of the questions and answers and the response voices and response actions of multiple people during the question-and-answer session, and finally calculate the screening activity conditional formula based on the recorded or stored data, and search the activity database 30 to generate a recommended activity and assist multiple people to participate in the recommended activity together.

請參閱圖2,係為本案之在虛擬世界中協助多人選擇活動之方法之步驟示意圖。須說明的是,圖2所示之方法可由伺服器或電腦設備所執行。 Please refer to Figure 2, which is a schematic diagram of the steps of the method for assisting multiple people in selecting activities in a virtual world. It should be noted that the method shown in Figure 2 can be executed by a server or a computer device.

步驟S201和步驟S202,偵測進入虛擬世界中多個VR裝置之語音及語音的語意和偵測進入虛擬世界中多個VR裝置之手勢及動作。具體言之,當多人穿戴VR裝置進入虛擬世界後,伺服器偵測VR裝置參與者的語音以及動作及手勢,其中,語音可由VR裝置接收以傳輸至伺服器,而手勢及/或動作可由VR裝置偵測而以訊號方式傳輸至伺服器。 Step S201 and step S202 detect the voice and the meaning of the voice of multiple VR devices in the virtual world and detect the gestures and actions of multiple VR devices in the virtual world. Specifically, when multiple people wear VR devices to enter the virtual world, the server detects the voice, action and gestures of the VR device participants, wherein the voice can be received by the VR device and transmitted to the server, and the gestures and/or actions can be detected by the VR device and transmitted to the server in the form of signals.

步驟S203和步驟S204,利用AI語意模型辨識語意和利用AI動作模型辨識手勢及/動作。於一實施例中,伺服器透過語意分析器、語意雲、或具有相似功能的人工智慧模型轉換為語意,以擷取其中的關鍵字,接著利用AI語意模型對語意及其中的關鍵字進行辨識。於另一實施例中,語音的辨識結果可為開啟協助語音或結束協助語音,語意的辨識結果例如問題語意或終止問答語 意,手勢的辨識結果可為開啟協助手勢或結束協助手勢,動作的辨識結果為贊同動作。辨識的步驟例如後述之步驟S301、S302、S401、S402、S404、S501、S502所示。 Step S203 and step S204, use AI semantic model to identify semantics and use AI action model to identify gestures and/or actions. In one embodiment, the server converts into semantics through semantic analyzer, semantic cloud, or artificial intelligence model with similar functions to extract keywords therein, and then uses AI semantic model to identify semantics and keywords therein. In another embodiment, the recognition result of voice can be to open the assisting voice or to end the assisting voice, the recognition result of semantics can be, for example, question semantics or to end the question and answer semantics, the recognition result of gestures can be to open the assisting gesture or to end the assisting gesture, and the recognition result of actions can be an approving action. The identification steps are shown in steps S301, S302, S401, S402, S404, S501, and S502 described below.

步驟S205,分析AI語意模型之語意的辨識結果、AI動作模型之動作的辨識結果、語意的辨識結果與動作的辨識結果之間的關聯性,以產生分析結果。於一實施例中,所述關聯性可例如:提供語音/語意的時間和VR裝置、提供動作的時間和VR裝置、當一VR裝置提供語音/語意或動作時其他VR裝置的反應語意或反應動作、當虛擬形象助理提供回答時各個VR裝置的反應動作或語意及其反應的時間、各VR裝置有反應的關鍵詞等。於另一實施例中,分析結果為與活動有關的活動關鍵字與加權重資料。詳言之,系統紀錄所有參與活動的VR裝置使用者參與討論的語意,並透過AI語意分析篩選出與活動相關的語詞為活動關鍵字,活動關鍵字為後續到活動資料庫作搜尋的主要輸入值。而每個活動關鍵字依其出現在討論中的次數,與當下記錄到該活動關鍵字時紀錄其他人透過VR裝置的動作判斷是否有贊同或認可的動作,辨識到每個動作定義的加權分數進行加權計算。 Step S205, analyzing the semantic recognition results of the AI semantic model, the action recognition results of the AI action model, and the correlation between the semantic recognition results and the action recognition results to generate analysis results. In one embodiment, the correlation may be, for example: the time and VR device of providing voice/semantics, the time and VR device of providing actions, the reaction semantics or reaction actions of other VR devices when a VR device provides voice/semantics or actions, the reaction actions or semantics of each VR device and the time of reaction when the virtual image assistant provides answers, and the keywords that each VR device has responded to. In another embodiment, the analysis results are activity keywords and weighted data related to the activity. In detail, the system records the semantics of all VR device users participating in the event, and selects activity-related words as activity keywords through AI semantic analysis. Activity keywords are the main input values for subsequent searches in the activity database. Each activity keyword is weighted by the number of times it appears in the discussion and the actions of other people through VR devices when the activity keyword is recorded. It determines whether there is an action of approval or recognition, and identifies the weighted score defined for each action for weighted calculation.

分析之步驟例如後述之步驟S405、S503。 The analysis steps are, for example, steps S405 and S503 described later.

步驟S206,根據分析結果自活動資料庫中決定出推薦活動。於一實施例中,將與討論活動時紀錄有關的活動關鍵字與根據活動關鍵字出現次數與當下記錄到的有贊同動作加權重資料,再組合成加權重資料篩選計算式(例如有記錄到活關鍵字「演唱會」,並記錄到五次(一次一分),其中五次有三次記錄到其他人有比讚的動作(比讚的動作+2分,所以2x3=6,加6分),則利用「演唱會」搜尋出來的活動加權11分再轉換成搜尋語法,至活動資料庫進行搜尋,藉 以最後搜尋出依加權平均數演算法得出平均分數最高的活動,即為本次活動討論的推薦活動。決定出推薦活動之步驟例如後述之步驟S504、S601。 Step S206, based on the analysis results, a recommended activity is determined from the activity database. In one embodiment, the activity keywords recorded during the discussion of the activity are weighted according to the number of occurrences of the activity keywords and the currently recorded like actions, and then combined into a weighted data screening formula (for example, the activity keyword "concert" is recorded and recorded five times (one point per time), and three of the five times other people's like actions are recorded (like actions +2 points) points, so 2x3=6, add 6 points), then use the "concert" search to weight the activities by 11 points and convert them into search syntax, and search the activity database to finally find the activity with the highest average score according to the weighted average algorithm, which is the recommended activity discussed in this event. The steps for determining the recommended activities are, for example, steps S504 and S601 described below.

步驟S207,令虛擬形象助理在虛擬空間中提供推薦活動。於一實施例中,推薦活動能透過觀賞窗的方式撥放,以供虛擬世界中的虛擬化身撞擊或接觸該活動現場觀賞窗時,能轉移進入虛擬世界中的推薦活動的現場。提供推薦活動之步驟例如後述之步驟S602~605。 Step S207, allowing the virtual image assistant to provide recommended activities in the virtual space. In one embodiment, the recommended activities can be displayed in the form of a viewing window, so that when the virtual avatar in the virtual world hits or touches the viewing window of the activity site, it can be transferred to the recommended activity site in the virtual world. The step of providing recommended activities is, for example, steps S602~605 described later.

請參閱圖3~6,其分別為本案之在虛擬世界中協助多人選擇活動之方法中開啟協助、討論活動期間、終止問答後、轉移至推薦活動之步驟示意圖。 Please refer to Figures 3 to 6, which are schematic diagrams of the steps of starting assistance, discussing activities, ending the Q&A session, and moving to recommending activities in the method of assisting multiple people in selecting activities in the virtual world.

首先,步驟S301~S303係說明:藉由辨識到開啟協助的語音和手勢後,虛擬形象助理開始協助同一團體參與者討論活動。 First, steps S301 to S303 illustrate that after recognizing the voice and gestures for starting assistance, the virtual image assistant begins to assist the participants in the same group to discuss the activity.

步驟S301,辨識VR裝置之語音或手勢;步驟S302,語音的辨識結果為開啟協助語意、或手勢的辨識結果為開啟協助手勢;步驟S303,虛擬形象助理出現在虛擬空間中。 Step S301, recognize the voice or gesture of the VR device; Step S302, the recognition result of the voice is to turn on the assistant meaning, or the recognition result of the gesture is to turn on the assistant gesture; Step S303, the virtual image assistant appears in the virtual space.

其次,步驟S401~S405係說明:在活動討論中所有語音或語意(無論是虛擬形象助理的回答、VR裝置參與者的問題或反應)、動作或手勢(無論是VR裝置參與者的問題或反應)、時間點、關鍵字等等,皆會被記錄下來以供後續分析。 Secondly, steps S401 to S405 explain that all voice or meaning (whether it is the answer of the virtual image assistant, the question or reaction of the VR device participant), action or gesture (whether it is the question or reaction of the VR device participant), time point, keywords, etc. in the activity discussion will be recorded for subsequent analysis.

步驟S401,辨識VR裝置之語意;步驟S402,語意的辨識結果為問題語意;步驟S403,產生對應的回答語意,可由虛擬形象助理根據該回答語意透過語音、圖片、影片、或活動現場觀賞窗來提供回答;步驟S404,辨識所有VR裝置對問題語意或回答語意之反應語意和反應動作(例如贊同的語意或動作);步驟S405,記錄各動作的辨識結果、提供各動作之VR裝置的數量、各語 意的辨識結果、提供各語意之VR裝置的數量、各動作的辨識結果與各語意的辨識結果之間的關聯性,以供後續分析。 Step S401, identify the semantics of VR devices; Step S402, the semantics identification result is the question semantics; Step S403, generate the corresponding answer semantics, which can be answered by the virtual image assistant through voice, pictures, videos, or event viewing windows according to the answer semantics; Step S404, identify the reaction semantics and reaction actions (such as the semantics or actions of approval) of all VR devices to the question semantics or answer semantics; Step S405, record the identification results of each action, the number of VR devices providing each action, the identification results of each semantics, the number of VR devices providing each semantics, and the correlation between the identification results of each action and the identification results of each semantics for subsequent analysis.

再者,步驟S501~S504係說明:根據此次活動討論的所有問題和答案及反應進行分析,將分析完的加權重資料計算後,再至活動資料庫中搜尋出推薦活動。 Furthermore, steps S501 to S504 explain: analyzing all the questions, answers and responses discussed in this activity, calculating the weighted data after the analysis, and then searching the activity database for recommended activities.

步驟S501,辨識VR裝置之語意;步驟S502,語意的辨識結果為終止問答語意;步驟S503,對在終止問答語意之前所記錄的所有資料進行分析;步驟S504,根據分析結果自活動資料庫中決定出推薦活動。 Step S501, identify the semantics of the VR device; Step S502, the semantics identification result is the termination of the question and answer semantics; Step S503, analyze all the data recorded before the termination of the question and answer semantics; Step S504, determine the recommended activities from the activity database based on the analysis results.

此外,步驟S601~S605係說明:當產生一推薦活動後,虛擬形象助理可將同一團體所有人瞬移至該推薦活動的會場,或者協助參與者加入行事曆,等活動開始時透過簡訊通知VR裝置參與者,更可連結直接加入該活動。 In addition, steps S601 to S605 explain that when a recommended event is generated, the virtual image assistant can teleport all members of the same group to the venue of the recommended event, or assist participants in adding the event to the calendar, and notify VR device participants via SMS when the event starts, and can also directly join the event through a link.

步驟S601,根據分析結果自活動資料庫中決定出推薦活動;步驟S602,判斷推薦活動是否正在進行中,其中,若推薦活動正在進行中,進至步驟S603,反之,若推薦活動不在進行中,進至步驟S604;步驟S603,將多個VR裝置轉移進入虛擬空間的推薦活動中;步驟S604,將推薦活動加入行事曆;步驟S605,推薦活動開始前,通知多個VR裝置。 Step S601, determine the recommended activity from the activity database according to the analysis result; Step S602, determine whether the recommended activity is in progress, if the recommended activity is in progress, proceed to step S603, otherwise, if the recommended activity is not in progress, proceed to step S604; Step S603, transfer multiple VR devices to the recommended activity in the virtual space; Step S604, add the recommended activity to the calendar; Step S605, notify multiple VR devices before the recommended activity starts.

另外,可提供活動現場觀賞窗以播放一推薦活動或其他正在進行中活動,而於偵測到VR裝置在虛擬世界中的虛擬化身撞擊或接觸該活動現場觀賞窗時,允許VR裝置11、12轉移進入該虛擬世界中的推薦活動或其他正在進行中活動。 In addition, an event viewing window can be provided to play a recommended event or other ongoing event, and when it is detected that the virtual avatar of the VR device in the virtual world hits or touches the event viewing window, the VR devices 11 and 12 are allowed to transfer to the recommended event or other ongoing event in the virtual world.

除了上述一或多個實施例之外,本案提供一種電腦程式產品,經由電腦載入程式後執行上述方法;另外,電腦程式(產品)除可儲存於記錄媒體 外,亦可在網路上直接傳輸提供,電腦程式(產品)係為載有電腦可讀取之程式且不限外在形式之物,所述電腦包括但不限於具有處理器之電子裝置,例如手機或平板等;此外,本案還提供一種電腦可讀取記錄媒體,係應用於具有處理器及/或記憶體之計算設備或電腦中,且電腦可讀取記錄媒體儲存有指令,並可利用計算設備或電腦透過處理器及/或記憶體執行電腦可讀取記錄媒體,以於執行電腦可讀取記錄媒體時執行上述方法及/或內容,所述電腦可讀取紀錄媒體(例如硬碟、軟碟、光碟、USB隨身碟)係儲存有該電腦程式(產品)。 In addition to one or more of the above embodiments, this case provides a computer program product, which executes the above method after the program is loaded into a computer; in addition, the computer program (product) can be stored in a recording medium and can also be directly transmitted and provided on the Internet. The computer program (product) is a thing that carries a program that can be read by a computer and is not limited to an external form. The computer includes but is not limited to an electronic device with a processor, such as a mobile phone or a tablet; In addition, this case also provides a computer that can The readable recording medium is applied to a computing device or computer having a processor and/or memory, and the computer-readable recording medium stores instructions, and the computing device or computer can execute the computer-readable recording medium through the processor and/or memory to execute the above method and/or content when executing the computer-readable recording medium. The computer-readable recording medium (such as a hard disk, a floppy disk, an optical disk, a USB flash drive) stores the computer program (product).

根據以上本案之在虛擬世界中協助多人選擇活動之系統、方法、電腦程式產品、電腦可讀取記錄媒體之一或多個實施例,以下提供一具體實施例之舉例說明。 Based on one or more embodiments of the system, method, computer program product, and computer-readable recording medium for assisting multiple people in selecting activities in a virtual world, an example of a specific embodiment is provided below.

同一團體多人戴上VR裝置進入元宇宙空間,多位朋友想一起到同一會場參加活動,但進入虛擬世界後發現活動好多,無從選擇,且多位都各有喜好的項目,這時小華說不如我們請『智慧助理』出來協助我們選擇吧,這時『智慧助理』以非玩家腳色(Non-Player Character,NPC)的方式出現,並開始協助大家選擇在元宇宙空間適合大家的活動。 Many people in the same group put on VR devices and entered the Metaverse. Many friends wanted to go to the same venue to participate in activities together, but after entering the virtual world, they found that there were too many activities to choose from, and many of them had their own favorites. At this time, Xiaohua said that we should ask the "Smart Assistant" to help us choose. At this time, the "Smart Assistant" appeared in the form of a non-player character (NPC) and began to help everyone choose activities suitable for everyone in the Metaverse.

小華首先說我想參加運動類型活動,可推薦類似的活動嗎?這時『智慧助理』列出幾個運動類型活動,並用語音、圖片、影片一一介紹,其中有一個活動還是正在進行中的活動,所以讓大家透過實景窗讓大家看到活動中實際的影像和聲音。這時當智慧助理在介紹活動時會偵測多人的語音與動作,並儲存語音與動作之間的關聯性,接者小明說躲避球活動好像不錯喔,這時小華和小張用VR裝置比出讚的動作。這些討論過程中與活動選擇相關的語意與VR動作都會被偵測,並記錄下關聯性資料。 Xiaohua first said that he wanted to participate in sports activities. Can you recommend similar activities? At this time, the "Smart Assistant" listed several sports activities and introduced them one by one with voice, pictures, and videos. One of the activities was still in progress, so everyone could see the actual images and sounds of the activity through the real-time window. At this time, when the smart assistant introduced the activities, it would detect the voices and actions of multiple people and store the correlation between voices and actions. Then Xiaoming said that the dodgeball activity seemed good. At this time, Xiaohua and Xiaozhang used VR devices to give thumbs-up actions. During these discussions, the semantics and VR actions related to the activity selection will be detected and the correlation data will be recorded.

接著,『活動智慧助理』將剛剛大家討論中與活動相關的問答語意與相關反應動作儲存下來,並依據儲存當下其他人贊同的語意或VR回饋動作進行加權計算並計算出加權平均計算式,依據此計算式到活動資料庫篩選,最後決定推薦正在進行的「躲避球活動」。 Next, the "Activity Smart Assistant" stores the question and answer semantics and related reaction actions related to the activity in the discussion just now, and performs weighted calculations based on the semantics or VR feedback actions that others agree with at the time, and calculates a weighted average calculation formula. Based on this calculation formula, it screens the activity database and finally decides to recommend the ongoing "Dodgeball Activity".

在推薦正在進行的「躲避球活動」後,智慧助理即刻將此團體多人一起傳送到「躲避球活動」會場中,並判斷最適合這個團體人數最佳的降落點。 After recommending the ongoing "dodgeball activity", the smart assistant immediately transported the group to the "dodgeball activity" venue and determined the best landing point for the group.

綜上所述,不同於以往的罐頭式回答或預設好題目讓使用者選擇,本案能讓虛擬世界元宇宙中的參與者可共同討論,更可與虛擬形象助理互動,以藉由助理的協助討論並提供最適合的活動(即決定出一推薦活動),來參加活動。另外,除了與虛擬形象助理互動外,本案藉由虛擬形象助理不僅能偵測VR裝置的語音問題,更會觀察與紀錄團體參與者們之間的互動相關的語音與動作,將語音理解為語意再辨識出是何種語意後,提供相應的回答,進而配合手勢或動作進行加權,藉以增加推薦活動的精準性。又,本案能除了提供最適合的活動外,也可以將所有同一團體的參與者瞬移到活動場所中,更可協助參與者安排活動行程。 In summary, unlike the canned answers or preset questions for users to choose from in the past, this solution allows participants in the virtual world metaverse to discuss together and interact with the virtual image assistant, so that the assistant can help them discuss and provide the most suitable activities (i.e., decide on a recommended activity) to participate in the activity. In addition, in addition to interacting with the virtual image assistant, this solution uses the virtual image assistant to not only detect voice problems of the VR device, but also observe and record the voice and actions related to the interaction between the group participants, understand the voice as meaning, and then identify what kind of meaning it is, provide corresponding answers, and then weight them with gestures or actions to increase the accuracy of the recommended activities. In addition to providing the most suitable activities, this case can also teleport all participants of the same group to the event venue, and can also assist participants in arranging the event schedule.

上述實施例僅例示性說明本案之功效,而非用於限制本案,任何熟習此項技藝之人士均可在不違背本案之精神及範疇下對上述該些實施態樣進行修飾與改變。因此本案之權利保護範圍,應如後述之申請專利範圍所列。 The above embodiments are only illustrative of the effects of this case, and are not intended to limit this case. Anyone familiar with this technology can modify and change the above embodiments without violating the spirit and scope of this case. Therefore, the scope of protection of this case should be as listed in the scope of the patent application described below.

S201~S207:步驟 S201~S207: Steps

Claims (11)

一種在虛擬世界中協同多人討論之系統,係包括: A system for collaborative multi-person discussion in a virtual world, comprising: 虛擬形象助理模組,用於顯示虛擬形象助理於虛擬世界中; Virtual image assistant module, used to display the virtual image assistant in the virtual world; 語音及語意偵測模組,用於偵測進入該虛擬世界的多個虛擬實境裝置之語音及該語音的語意; The voice and semantic detection module is used to detect the voice and semantics of multiple virtual reality devices entering the virtual world; 手勢及動作偵測模組,用於偵測進入該虛擬世界的該多個虛擬實境裝置之手勢及動作; A gesture and action detection module, used to detect the gestures and actions of the multiple virtual reality devices entering the virtual world; AI辨識模組,用於辨識該語音、該語意、該手勢、該動作,以判斷該語音的辨識結果是否為開啟協助語音,利用AI語意模型判斷該語意的辨識結果是否為問題語意,利用AI動作模型判斷該手勢的辨識結果是否為開啟協助手勢,及利用該AI動作模型判斷該動作的辨識結果是否為贊同動作; The AI recognition module is used to recognize the voice, the meaning, the gesture, and the action to determine whether the recognition result of the voice is the voice to turn on the assistance, use the AI semantic model to determine whether the recognition result of the meaning is the question meaning, use the AI action model to determine whether the recognition result of the gesture is the gesture to turn on the assistance, and use the AI action model to determine whether the recognition result of the action is an approval action; 活動分析模組,用於分析該AI語意模型之該語意的辨識結果、該AI動作模型之該動作的辨識結果、及該語意的辨識結果與該動作的辨識結果之間的關聯性,以產生分析結果;以及 An activity analysis module is used to analyze the semantic recognition result of the AI semantic model, the action recognition result of the AI action model, and the correlation between the semantic recognition result and the action recognition result to generate an analysis result; and 活動決定模組,用於根據該活動分析模組之分析結果自活動資料庫中決定出推薦活動,以由顯示於該虛擬世界中之該虛擬形象助理提供該推薦活動。 The activity determination module is used to determine the recommended activities from the activity database according to the analysis results of the activity analysis module, so that the virtual image assistant displayed in the virtual world can provide the recommended activities. 如請求項1所述之系統,其中,該AI語意模型係利用與活動有關的問答集作為訓練資料進行訓練,以於該語意的辨識結果為問題語意時,該AI語意模型係產生對應的回答語意,而由該虛擬形象助理在該虛擬世界中根據該回答語意透過語音、圖片、影片、或活動現場觀賞窗來提供回答。 The system as claimed in claim 1, wherein the AI semantic model is trained using a question-and-answer set related to an activity as training data, so that when the semantic recognition result is the question semantics, the AI semantic model generates a corresponding answer semantics, and the virtual image assistant provides an answer in the virtual world based on the answer semantics through voice, pictures, videos, or an event viewing window. 如請求項2所述之系統,其中,該AI動作模型係利用虛擬實境裝置動作資料作為訓練資料進行訓練,以於該虛擬形象助理提供該回答時,該AI 動作模型辨識各該虛擬實境裝置對於該回答之反應動作且該AI語意模型辨識各該虛擬實境裝置對於該回答之反應語意,而由該AI辨識模組記錄各該反應動作的辨識結果、提供各該反應動作之虛擬實境裝置的數量、各該反應語意的辨識結果、提供各該反應語意之虛擬實境裝置的數量、各該反應動作的辨識結果與各該反應語意的辨識結果之間的關聯性,以供該活動分析模組進行分析。 The system as claimed in claim 2, wherein the AI action model is trained using virtual reality device action data as training data, so that when the virtual image assistant provides the answer, the AI action model recognizes the response action of each virtual reality device to the answer and the AI semantic model recognizes the response semantics of each virtual reality device to the answer, The AI recognition module records the recognition results of each reaction action, the number of virtual reality devices that provide each reaction action, the recognition results of each reaction semantics, the number of virtual reality devices that provide each reaction semantics, and the correlation between the recognition results of each reaction action and the recognition results of each reaction semantics, so as to provide the activity analysis module for analysis. 如請求項1所述之系統,其中,該AI動作模型係利用各種手勢圖像作為訓練資料進行訓練,以於該手勢及動作偵測模組偵測到該多個虛擬實境裝置的手勢且該AI動作模型對該手勢的辨識結果為開啟協助手勢或者於該語音及語意偵測模組偵測到該多個虛擬實境裝置的語意且該AI語意模型對該語意的辨識結果為開啟協助語意時,令該虛擬形象助理出現在該虛擬世界中。 The system as claimed in claim 1, wherein the AI action model is trained using various gesture images as training data, so that when the gesture and action detection module detects the gestures of the multiple virtual reality devices and the AI action model recognizes the gestures as turning on the assist gesture, or when the voice and semantic detection module detects the semantics of the multiple virtual reality devices and the AI semantic model recognizes the semantics as turning on the assist semantics, the virtual image assistant appears in the virtual world. 如請求項1所述之系統,更包括活動控制模組,其中,若該活動決定模組所決定之推薦活動在該虛擬世界正在進行中,則該活動控制模組將該多個虛擬實境裝置轉移進入該虛擬世界的該推薦活動中,而若該活動決定模組所決定之推薦活動在該虛擬世界不在進行中,則該活動控制模組將該推薦活動加入行事曆,待該推薦活動開始前,再通知該多個虛擬實境裝置。 The system as described in claim 1 further includes an activity control module, wherein if the recommended activity determined by the activity determination module is ongoing in the virtual world, the activity control module transfers the multiple virtual reality devices into the recommended activity in the virtual world, and if the recommended activity determined by the activity determination module is not ongoing in the virtual world, the activity control module adds the recommended activity to the calendar, and notifies the multiple virtual reality devices before the recommended activity starts. 一種在虛擬世界中協同多人討論之方法,係包括: A method for collaborative discussion among multiple people in a virtual world, comprising: 偵測進入虛擬世界的多個虛擬實境裝置之語音及該語音的語意; Detecting the voices of multiple virtual reality devices entering the virtual world and the meaning of the voices; 偵測進入該虛擬世界的該多個虛擬實境裝置之手勢及動作; Detecting the gestures and actions of the multiple virtual reality devices entering the virtual world; 辨識該語音及該語意,以判斷該語音的辨識結果是否為開啟協助語音,及利用AI語意模型判斷該語意的辨識結果是否為問題語意; Identify the voice and the meaning to determine whether the recognition result of the voice is to turn on the assistive voice, and use the AI semantic model to determine whether the recognition result of the meaning is the question meaning; 辨識該手勢及動作,以利用AI動作模型判斷該手勢的辨識結果是否為開啟協助手勢,及利用該AI動作模型判斷該動作的辨識結果是否為贊同動作; Recognize the gesture and action, and use the AI action model to determine whether the recognition result of the gesture is a gesture to turn on the collaboration, and use the AI action model to determine whether the recognition result of the action is an approval action; 分析該AI語意模型之該語意的辨識結果、該AI動作模型之該動作的辨識結果、及該語意的辨識結果與該動作的辨識結果之間的關聯性,以產生分析結果;以及 Analyze the semantic recognition result of the AI semantic model, the action recognition result of the AI action model, and the correlation between the semantic recognition result and the action recognition result to generate an analysis result; and 根據該分析結果自活動資料庫中決定出推薦活動,以由虛擬形象助理在該虛擬世界中提供該推薦活動。 According to the analysis result, a recommended activity is determined from the activity database, and the virtual image assistant provides the recommended activity in the virtual world. 如請求項6所述之方法,其中,該AI語意模型係利用與活動有關的問答集作為訓練資料進行訓練,以於該語意的辨識結果為問題語意時,該AI語意模型係產生對應的回答語意,而由該虛擬形象助理在該虛擬世界中根據該回答語意透過語音、圖片、影片、或活動現場觀賞窗來提供回答。 The method as described in claim 6, wherein the AI semantic model is trained using a question-and-answer set related to the activity as training data, so that when the semantic recognition result is the question semantics, the AI semantic model generates the corresponding answer semantics, and the virtual image assistant provides an answer in the virtual world based on the answer semantics through voice, pictures, videos, or event live viewing windows. 如請求項7所述之方法,其中,該AI動作模型係利用虛擬實境裝置動作資料作為訓練資料進行訓練,以於該虛擬形象助理提供該回答時,該AI動作模型辨識各該虛擬實境裝置對於該回答之反應動作且該AI語意模型辨識各該虛擬實境裝置對於該回答之反應語意,進而記錄各該反應動作的辨識結果、提供各該反應動作之虛擬實境裝置的數量、各該反應語意的辨識結果、提供各該反應語意之虛擬實境裝置的數量、各該反應動作的辨識結果與各該反應語意的辨識結果之間的關聯性,藉以供進行分析。 The method as described in claim 7, wherein the AI action model is trained using virtual reality device action data as training data, so that when the virtual image assistant provides the answer, the AI action model identifies the reaction action of each virtual reality device to the answer and the AI semantic model identifies the reaction semantics of each virtual reality device to the answer, and then records the recognition results of each reaction action, the number of virtual reality devices providing each reaction action, the recognition results of each reaction semantics, the number of virtual reality devices providing each reaction semantics, and the correlation between the recognition results of each reaction action and the recognition results of each reaction semantics, so as to provide analysis. 如請求項6所述之方法,其中,該AI動作模型係利用各種手勢圖像作為訓練資料進行訓練,以於偵測到該多個虛擬實境裝置的手勢且該AI動作模型對該手勢的辨識結果為開啟協助手勢或者於偵測到該多個虛擬實境裝置的語意且該AI語意模型對該語意的辨識結果為開啟協助語意時,令該虛擬形象助理出現在該虛擬世界中。 The method as described in claim 6, wherein the AI action model is trained using various gesture images as training data, so that when the gestures of the multiple virtual reality devices are detected and the AI action model recognizes the gesture as turning on the assist gesture, or when the semantics of the multiple virtual reality devices are detected and the AI semantic model recognizes the semantics as turning on the assist semantics, the virtual image assistant appears in the virtual world. 如請求項6所述之方法,更包括:若該推薦活動在該虛擬世界正在進行中,則將該多個虛擬實境裝置轉移進入該虛擬世界的該推薦活動中,而若該推薦活動在該虛擬世界不在進行中,則將該推薦活動加入行事曆,待該推薦活動開始前,再通知該多個虛擬實境裝置。 The method as described in claim 6 further includes: if the recommended activity is ongoing in the virtual world, the multiple virtual reality devices are transferred to the recommended activity in the virtual world; if the recommended activity is not ongoing in the virtual world, the recommended activity is added to the calendar, and before the recommended activity starts, the multiple virtual reality devices are notified. 一種電腦程式產品,經由電腦載入程式以執行請求項6-10中任一項所述之方法。 A computer program product, which is loaded into a computer to execute the method described in any one of claim 6-10.
TW112143065A 2023-11-08 System, method and computer program product for assisting multiple users to choice activity in virtual world TWI842650B (en)

Publications (1)

Publication Number Publication Date
TWI842650B true TWI842650B (en) 2024-05-11

Family

ID=

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TW202336562A (en) 2021-10-29 2023-09-16 美商元平台技術有限公司 Controlling interactions with virtual objects

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TW202336562A (en) 2021-10-29 2023-09-16 美商元平台技術有限公司 Controlling interactions with virtual objects

Similar Documents

Publication Publication Date Title
CN107203953B (en) Teaching system based on internet, expression recognition and voice recognition and implementation method thereof
US9244533B2 (en) Camera navigation for presentations
US20190147367A1 (en) Detecting interaction during meetings
Bigham et al. Human-computer interaction and collective intelligence
US20200082928A1 (en) Assisting psychological cure in automated chatting
US20090258703A1 (en) Motion Assessment Using a Game Controller
US20160004299A1 (en) Systems and methods for assessing, verifying and adjusting the affective state of a user
CN109154948B (en) Method and apparatus for providing content
JP2011039860A (en) Conversation system, conversation method, and computer program using virtual space
US20220172442A1 (en) A smartphone, a host computer, a system and a method for a virtual object on augmented reality
JP4889690B2 (en) Conversation display method and server apparatus
Loth et al. Ghost-in-the-Machine reveals human social signals for human–robot interaction
Vu et al. Emotion recognition based on human gesture and speech information using RT middleware
CN114125529A (en) Method, equipment and storage medium for generating and demonstrating video
CN112287767A (en) Interaction control method, device, storage medium and electronic equipment
JP6302381B2 (en) Operation support apparatus, operation support method, and operation support program
CN114513694A (en) Scoring determination method and device, electronic equipment and storage medium
TWI842650B (en) System, method and computer program product for assisting multiple users to choice activity in virtual world
Sarkar et al. AVCAffe: a large scale audio-visual dataset of cognitive load and affect for remote work
CN111767386A (en) Conversation processing method and device, electronic equipment and computer readable storage medium
WO2023221233A1 (en) Interactive mirroring apparatus, system and method
US20240171418A1 (en) Information processing device and information processing method
KR102590988B1 (en) Apparatus, method and program for providing metaverse service to exercise with avatar
US11886773B1 (en) Systems and methods for a voice-controlled interactive application
US20240235872A1 (en) Systems and methods for decentralized generation of a summary of a virtual meeting