CN106534527A - Navigation voice processing device and method - Google Patents

Navigation voice processing device and method Download PDF

Info

Publication number
CN106534527A
CN106534527A CN201610976685.2A CN201610976685A CN106534527A CN 106534527 A CN106534527 A CN 106534527A CN 201610976685 A CN201610976685 A CN 201610976685A CN 106534527 A CN106534527 A CN 106534527A
Authority
CN
China
Prior art keywords
voice
navigation
prompt
played
speaker
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610976685.2A
Other languages
Chinese (zh)
Inventor
周龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nubia Technology Co Ltd
Original Assignee
Nubia Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nubia Technology Co Ltd filed Critical Nubia Technology Co Ltd
Priority to CN201610976685.2A priority Critical patent/CN106534527A/en
Publication of CN106534527A publication Critical patent/CN106534527A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72433User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/36Input/output arrangements for on-board computers
    • G01C21/3626Details of the output of route guidance instructions
    • G01C21/3629Guidance using speech or audio output, e.g. text-to-speech

Landscapes

  • Engineering & Computer Science (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Automation & Control Theory (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Human Computer Interaction (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Navigation (AREA)

Abstract

The embodiment of the invention discloses a navigation voice processing device and method. The device provided by the embodiment of the invention comprises an identification module which is used for identifying a prompt word label of a to-be-played voice in a navigation program, wherein the prompt word label comprises a left prompt word and a right prompt word; and a voice playing module which is used for playing the to-be-played voice through a left side loudspeaker or a right side loudspeaker of a terminal device according to the prompt word label of the to-be-played voice in the navigation program. According to the embodiment of the method and the device, through a reasonable design and a mode of carrying out navigation voice playing on the basis of the double loudspeakers, the double loudspeakers are effectively applied to navigation playing, the navigation voice playing effect is more stereoscopic, the indication is relatively high, and the application value of the double loudspeakers is improved.

Description

Navigation voice processing device and method
Technical Field
The present application relates to, but not limited to, the field of computer technologies, and in particular, to a navigation speech processing apparatus and method.
Background
With the development of the intelligent terminal and the application and popularization of the intelligent terminal, the needs of the user for the intelligent terminal gradually develop towards the direction of differentiation, and the better sound quality experience effect becomes an important direction for the pursuit of the user.
Currently, more and more manufacturers of smart terminals (e.g., smartphones) typically employ a dual speaker configuration when designing and generating smart terminals. In the intelligent terminal with the single loudspeaker, a user can only feel the sound, the tone color and the volume of music, but cannot feel the horizontal movement of the sound from left to right and the like; in the intelligent terminal provided with the double loudspeakers, a user can feel the position information of the left loudspeaker and the right loudspeaker and can generate the effect of stereo surround sound by matching with the sound channel processing of each loudspeaker, so that the sound quality effect of playing is obviously improved. The stereo sound effect that two speakers produced mainly uses in the application scene of broadcast audio frequency at present for promote the audio frequency audio effect, do not have the broadcast effect of abundant utilization two speakers in other application scenes at intelligent terminal.
To sum up, the application range of the dual speakers configured in the intelligent terminal in the prior art is single, and the advantages of the dual speaker playing effect are not fully utilized.
Disclosure of Invention
In order to solve the above technical problems, embodiments of the present invention provide a navigation voice processing apparatus and method, which effectively apply dual speakers to navigation play through a reasonable design of a navigation voice play mode through the dual speakers, so that the navigation voice play effect is more stereoscopic, the instruction is stronger, and the application value of the dual speakers is improved.
The embodiment of the invention provides a navigation voice processing device, which comprises:
the recognition module is used for recognizing a prompt language tag of voice to be played in the navigation program, wherein the prompt language tag comprises: a left prompt and a right prompt;
and the voice playing module is used for playing the voice to be played by adopting a left side loudspeaker or a right side loudspeaker of the terminal equipment according to the prompt language label of the voice to be played, which is identified by the identification module.
Optionally, in the navigation voice processing apparatus as described above, the apparatus further includes:
and the adding module is used for adding a prompt language label to the navigation voice in the navigation program before the recognition module recognizes the prompt language label of the voice to be played in the navigation program.
Optionally, in the navigation voice processing apparatus as described above, the playing of the voice to be played by the voice playing module using the left speaker or the right speaker of the terminal device according to the prompt language tag of the voice to be played, which is identified by the identifying module, includes:
when the prompt language label is the left prompt language, playing the voice to be played through a left side loudspeaker of the terminal equipment;
when the prompt label is the right prompt, the voice to be played is played through a right loudspeaker of the terminal equipment;
the prompt language label also comprises a non-directional prompt language, and the voice playing module is further used for playing the voice to be played through a left side loudspeaker and a right side loudspeaker of the terminal device when the prompt language label is the non-directional prompt language.
Optionally, in the navigation voice processing apparatus as described above, the left speaker and the right speaker are speakers provided on the terminal device; or,
the left side loudspeaker and the right side loudspeaker are external loudspeakers of the terminal equipment.
Optionally, in the navigation voice processing apparatus described above, when the left speaker and the right speaker are speakers arranged on the terminal device, the recognition module is further configured to recognize, before the voice playing module plays the to-be-played voice, that the speaker arranged at the left end of the navigation interface of the terminal device is the left speaker and that the speaker arranged at the right end of the navigation interface is the right speaker when the terminal device is in the landscape screen mode.
The embodiment of the invention provides a navigation voice processing method, which comprises the following steps:
identifying a prompt language tag of voice to be played in a navigation program, wherein the prompt language tag comprises: a left prompt and a right prompt;
and according to the prompt language label of the voice to be played, playing the voice to be played by adopting a left side loudspeaker or a right side loudspeaker of the terminal equipment.
Optionally, in the navigation voice processing method described above, before the recognizing a prompt language tag of the voice to be played in the navigation program, the method further includes:
and adding a prompt language tag to the navigation voice in the navigation program.
Optionally, in the navigation voice processing method, the playing the voice to be played by using the left speaker or the right speaker of the terminal device according to the prompt language tag of the voice to be played includes:
when the prompt language label is the left prompt language, playing the voice to be played through a left side loudspeaker of the terminal equipment;
when the prompt label is the right prompt, the voice to be played is played through a right loudspeaker of the terminal equipment;
the prompt label further comprises a non-directional prompt, and the method further comprises:
and when the prompt language label is the non-directional prompt language, playing the voice to be played through a left side loudspeaker and a right side loudspeaker of the terminal equipment.
Optionally, in the navigation voice processing method described above, the left speaker and the right speaker are speakers disposed on the terminal device; or,
the left side loudspeaker and the right side loudspeaker are external loudspeakers of the terminal equipment.
Optionally, in the navigation voice processing method described above, when the left speaker and the right speaker are speakers arranged on the terminal device, before the voice to be played is played, the method further includes:
when the terminal equipment is in a horizontal screen mode, a loudspeaker arranged at the left end of a navigation interface of the terminal equipment is identified as the left side loudspeaker, and a loudspeaker arranged at the right end of the navigation interface is identified as the right side loudspeaker.
The navigation voice processing device and the navigation voice processing method provided by the embodiment of the invention have the advantages that the prompt language label of voice to be played in a navigation program is identified through the identification module, the prompt language label comprises a left prompt language and a right prompt language, then the voice playing module plays the voice to be played by adopting a left side loudspeaker or a right side loudspeaker of the terminal equipment according to the prompt language label of the voice to be played identified by the identification module, and when the navigation voice is played by adopting the navigation voice processing device provided by the embodiment of the invention, the navigation voice with different prompt language labels can be stereoscopically played by utilizing the position characteristics of the two loudspeakers, so that the direction sources of different navigation voices are reflected, and the navigation voice is presented with the characteristic of directionality; according to the embodiment of the invention, the navigation voice playing mode is carried out through the double loudspeakers through reasonable design, the double loudspeakers are effectively applied to the navigation playing, the navigation voice playing effect is more stereoscopic, the indication is stronger, and the application value of the double loudspeakers is improved.
Drawings
The accompanying drawings are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the example serve to explain the principles of the invention and not to limit the invention.
Fig. 1 is a schematic hardware structure of an alternative mobile terminal for implementing various embodiments of the present invention;
FIG. 2 depicts a communication system in which a mobile terminal according to the present invention is capable of operating;
fig. 3 is a schematic structural diagram of a navigation speech processing apparatus according to an embodiment of the present invention;
FIG. 4 is a schematic diagram of a navigation speech processing apparatus according to an embodiment of the present invention, in which two speakers transmit audio;
FIG. 5 is a schematic structural diagram of another navigation speech processing apparatus according to an embodiment of the present invention;
FIG. 6 is a flowchart illustrating a navigation voice processing method according to an embodiment of the present invention;
FIG. 7 is a flowchart illustrating another navigation voice processing method according to an embodiment of the present invention;
fig. 8 is a flowchart illustrating a navigation speech processing method according to another embodiment of the present invention.
Detailed Description
It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
A mobile terminal implementing various embodiments of the present invention will now be described with reference to the accompanying drawings. In the following description, suffixes used to represent elements such as "module", "means", or "unit" are used only for the convenience of description of the present invention, and have no specific meaning in themselves. Thus, "module" and "component" may be used in a mixture.
The mobile terminal may be implemented in various forms. For example, the terminal described in the present invention may include a mobile terminal such as a mobile phone, a smart phone, a notebook computer, a digital broadcast receiver, a PDA (personal digital assistant), a PAD (tablet computer), a PMP (portable multimedia player), a navigation device, and the like, and a stationary terminal such as a digital TV, a desktop computer, and the like. In the following, it is assumed that the terminal is a mobile terminal. However, it will be understood by those skilled in the art that the configuration according to the embodiment of the present invention can be applied to a fixed type terminal in addition to elements particularly used for moving purposes.
Fig. 1 is a schematic hardware structure of an optional mobile terminal for implementing various embodiments of the present invention.
The mobile terminal 100 may include a wireless communication unit 110, an a/V (audio/video) input unit 120, a user input unit 130, a sensing unit 140, an output unit 150, a memory 160, an interface unit 170, a controller 180, and a power supply unit 190, etc. Fig. 1 illustrates a mobile terminal having various components, but it is to be understood that not all illustrated components are required to be implemented. More or fewer components may alternatively be implemented. Elements of the mobile terminal will be described in detail below.
The wireless communication unit 110 typically includes one or more components that allow radio communication between the mobile terminal 100 and a wireless communication system or network. For example, the wireless communication unit may include at least one of a broadcast receiving module 111, a mobile communication module 112, a wireless internet module 113, a short-range communication module 114, and a location information module 115.
The broadcast receiving module 111 receives a broadcast signal and/or broadcast associated information from an external broadcast management server via a broadcast channel. The broadcast channel may include a satellite channel and/or a terrestrial channel. The broadcast management server may be a server that generates and transmits a broadcast signal and/or broadcast associated information or a server that receives a previously generated broadcast signal and/or broadcast associated information and transmits it to a terminal. The broadcast signal may include a TV broadcast signal, a radio broadcast signal, a data broadcast signal, and the like. Also, the broadcast signal may further include a broadcast signal combined with a TV or radio broadcast signal. The broadcast associated information may also be provided via a mobile communication network, and in this case, the broadcast associated information may be received by the mobile communication module 112. The broadcast signal may exist in various forms, for example, it may exist in the form of an Electronic Program Guide (EPG) of Digital Multimedia Broadcasting (DMB), an Electronic Service Guide (ESG) of digital video broadcasting-handheld (DVB-H), and the like. The broadcast receiving module 111 may receive a signal broadcast by using various types of broadcasting systems. In particular, the broadcast receiving module 111 may receive digital broadcasting by using a digital broadcasting system such as a data broadcasting system of multimedia broadcasting-terrestrial (DMB-T), digital multimedia broadcasting-satellite (DMB-S), digital video broadcasting-handheld (DVB-H), forward link media (MediaFLO @), terrestrial digital broadcasting integrated service (ISDB-T), and the like. The broadcast receiving module 111 may be constructed to be suitable for various broadcasting systems that provide broadcast signals as well as the above-mentioned digital broadcasting systems. The broadcast signal and/or broadcast associated information received via the broadcast receiving module 111 may be stored in the memory 160 (or other type of storage medium).
The mobile communication module 112 transmits and/or receives radio signals to and/or from at least one of a base station (e.g., access point, node B, etc.), an external terminal, and a server. Such radio signals may include voice call signals, video call signals, or various types of data transmitted and/or received according to text and/or multimedia messages.
The wireless internet module 113 supports wireless internet access of the mobile terminal. The module may be internally or externally coupled to the terminal. The wireless internet access technology to which the module relates may include WLAN (wireless LAN) (Wi-Fi), Wibro (wireless broadband), Wimax (worldwide interoperability for microwave access), HSDPA (high speed downlink packet access), and the like.
The short-range communication module 114 is a module for supporting short-range communication. Some examples of short-range communication technologies include bluetooth (TM), Radio Frequency Identification (RFID), infrared data association (IrDA), Ultra Wideband (UWB), zigbee (TM), and the like.
The location information module 115 is a module for checking or acquiring location information of the mobile terminal. A typical example of the location information module is a GPS (global positioning system). According to the current technology, the GPS module 115 calculates distance information and accurate time information from three or more satellites and applies triangulation to the calculated information, thereby accurately calculating three-dimensional current location information according to longitude, latitude, and altitude. Currently, a method for calculating position and time information uses three satellites and corrects an error of the calculated position and time information by using another satellite. In addition, the GPS module 115 can calculate speed information by continuously calculating current position information in real time.
The a/V input unit 120 is used to receive an audio or video signal. The a/V input unit 120 may include a camera 121 and a microphone 1220, and the camera 121 processes image data of still pictures or video obtained by an image capturing apparatus in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display module 151. The image frames processed by the cameras 121 may be stored in the memory 160 (or other storage medium) or transmitted via the wireless communication unit 110, and two or more cameras 121 may be provided according to the construction of the mobile terminal. The microphone 122 may receive sounds (audio data) via the microphone in a phone call mode, a recording mode, a voice recognition mode, or the like, and can process such sounds into audio data. The processed audio (voice) data may be converted into a format output transmittable to a mobile communication base station via the mobile communication module 112 in case of a phone call mode. The microphone 122 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting audio signals.
The user input unit 130 may generate key input data according to a command input by a user to control various operations of the mobile terminal. The user input unit 130 allows a user to input various types of information, and may include a keyboard, dome sheet, touch pad (e.g., a touch-sensitive member that detects changes in resistance, pressure, capacitance, and the like due to being touched), scroll wheel, joystick, and the like. In particular, when the touch pad is superimposed on the display module 151 in the form of a layer, a touch screen may be formed.
The sensing unit 140 detects a current state of the mobile terminal 100 (e.g., an open or closed state of the mobile terminal 100), a position of the mobile terminal 100, presence or absence of contact (i.e., touch input) by a user with the mobile terminal 100, an orientation of the mobile terminal 100, acceleration or deceleration movement and direction of the mobile terminal 100, and the like, and generates a command or signal for controlling an operation of the mobile terminal 100. For example, when the mobile terminal 100 is implemented as a slide-type mobile phone, the sensing unit 140 may sense whether the slide-type phone is opened or closed. In addition, the sensing unit 140 can detect whether the power supply unit 190 supplies power or whether the interface unit 170 is coupled with an external device.
The interface unit 170 serves as an interface through which at least one external device is connected to the mobile terminal 100. For example, the external device may include a wired or wireless headset port, an external power supply (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The identification module may store various information for authenticating a user using the mobile terminal 100 and may include a User Identity Module (UIM), a Subscriber Identity Module (SIM), a Universal Subscriber Identity Module (USIM), and the like. In addition, a device having an identification module (hereinafter, referred to as an "identification device") may take the form of a smart card, and thus, the identification device may be connected with the mobile terminal 100 via a port or other connection means. The interface unit 170 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the mobile terminal 100 or may be used to transmit data between the mobile terminal and the external device.
In addition, when the mobile terminal 100 is connected with an external cradle, the interface unit 170 may serve as a path through which power is supplied from the cradle to the mobile terminal 100 or may serve as a path through which various command signals input from the cradle are transmitted to the mobile terminal. Various command signals or power input from the cradle may be used as signals for recognizing whether the mobile terminal is accurately mounted on the cradle. The output unit 150 is configured to provide output signals (e.g., audio signals, video signals, alarm signals, vibration signals, etc.) in a visual, audio, and/or tactile manner. The output unit 150 may include a display module 151, an audio output module 152, an alarm module 153, and the like.
The display module 151 may display information processed in the mobile terminal 100. For example, when the mobile terminal 100 is in a phone call mode, the display module 151 may display a User Interface (UI) or a Graphical User Interface (GUI) related to a call or other communication (e.g., text messaging, multimedia file downloading, etc.). When the mobile terminal 100 is in a video call mode or an image capturing mode, the display module 151 may display a captured image and/or a received image, a UI or GUI showing a video or an image and related functions, and the like.
Meanwhile, when the display module 151 and the touch pad are stacked on each other in the form of layers to form a touch screen, the display module 151 may serve as an input device and an output device. The display module 151 may include at least one of a Liquid Crystal Display (LCD), a thin film transistor LCD (TFT-LCD), an Organic Light Emitting Diode (OLED) display, a flexible display, a three-dimensional (3D) display, and the like. Some of these displays may be configured to be transparent to allow a user to view from the outside, which may be referred to as transparent displays, and a typical transparent display may be, for example, a TOLED (transparent organic light emitting diode) display or the like. Depending on the particular desired implementation, the mobile terminal 100 may include two or more display units (or other display devices), for example, the mobile terminal may include an external display unit (not shown) and an internal display unit (not shown). The touch screen may be used to detect a touch input pressure as well as a touch input position and a touch input area.
The audio output module 152 may convert audio data received by the wireless communication unit 110 or stored in the memory 160 into an audio signal and output as sound when the mobile terminal is in a call signal reception mode, a call mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output module 152 may provide audio output related to a specific function performed by the mobile terminal 100 (e.g., a call signal reception sound, a message reception sound, etc.). The audio output module 152 may include a speaker, a buzzer, and the like.
The alarm module 153 may provide an output to notify the mobile terminal 100 of the occurrence of an event. Typical events may include call reception, message reception, key signal input, touch input, and the like. In addition to audio or video output, the alarm module 153 may provide output in different ways to notify the occurrence of an event. For example, the alarm module 153 may provide an output in the form of a vibration, and when a call, a message, or some other incoming communication (incomingmunication) is received, the alarm module 153 may provide a tactile output (i.e., a vibration) to inform the user thereof. By providing such a tactile output, the user can recognize the occurrence of various events even when the user's mobile phone is in the user's pocket. The alarm module 153 may also provide an output notifying the occurrence of an event via the display module 151 or the audio output module 152.
The memory 160 may store software programs and the like for processing and controlling operations performed by the controller 180, or may temporarily store data (e.g., a phonebook, messages, still images, videos, and the like) that has been or will be output. Also, the memory 160 may store data regarding various ways of vibration and audio signals output when a touch is applied to the touch screen.
The memory 160 may include at least one type of storage medium including a flash memory, a hard disk, a multimedia card, a card-type memory (e.g., SD or DX memory, etc.), a Random Access Memory (RAM), a Static Random Access Memory (SRAM), a read-only memory (ROM), an electrically erasable programmable read-only memory (EEPROM), a programmable read-only memory (PROM), a magnetic memory, a magnetic disk, an optical disk, and the like. Also, the mobile terminal 100 may cooperate with a network storage device that performs a storage function of the memory 160 through a network connection.
The controller 180 generally controls the overall operation of the mobile terminal. For example, the controller 180 performs control and processing related to voice calls, data communications, video calls, and the like. In addition, the controller 180 may include a multimedia module 181 for reproducing (or playing back) multimedia data, and the multimedia module 181 may be constructed within the controller 180 or may be constructed separately from the controller 180. The controller 180 may perform a pattern recognition process to recognize a handwriting input or a picture drawing input performed on the touch screen as a character or an image.
The power supply unit 190 receives external power or internal power and provides appropriate power required to operate various elements and components under the control of the controller 180.
The various embodiments described herein may be implemented in a computer-readable medium using, for example, computer software, hardware, or any combination thereof. For a hardware implementation, the embodiments described herein may be implemented using at least one of an Application Specific Integrated Circuit (ASIC), a Digital Signal Processor (DSP), a Digital Signal Processing Device (DSPD), a Programmable Logic Device (PLD), a Field Programmable Gate Array (FPGA), a processor, a controller, a microcontroller, a microprocessor, an electronic unit designed to perform the functions described herein, and in some cases, such embodiments may be implemented in the controller 180. For a software implementation, the implementation such as a process or a function may be implemented with a separate software module that allows performing at least one function or operation. The software codes may be implemented by software applications (or programs) written in any suitable programming language, which may be stored in the memory 160 and executed by the controller 180.
Up to this point, mobile terminals have been described in terms of their functionality. Hereinafter, a slide-type mobile terminal among various types of mobile terminals, such as a folder-type, bar-type, swing-type, slide-type mobile terminal, and the like, will be described as an example for the sake of brevity. Accordingly, the present invention can be applied to any type of mobile terminal, and is not limited to a slide type mobile terminal.
The mobile terminal 100 as shown in fig. 1 may be configured to operate with communication systems such as wired and wireless communication systems and satellite-based communication systems that transmit data via frames or packets.
A communication system in which a mobile terminal according to the present invention is operable will now be described with reference to fig. 2.
Such communication systems may use different air interfaces and/or physical layers. For example, the air interface used by the communication system includes, for example, Frequency Division Multiple Access (FDMA), Time Division Multiple Access (TDMA), Code Division Multiple Access (CDMA), and Universal Mobile Telecommunications System (UMTS) (in particular, Long Term Evolution (LTE)), global system for mobile communications (GSM), and the like. By way of non-limiting example, the following description relates to a CDMA communication system, but such teachings are equally applicable to other types of systems.
Referring to fig. 2, the CDMA wireless communication system may include a plurality of mobile terminals 100, a plurality of Base Stations (BSs) 270, Base Station Controllers (BSCs) 275, and a Mobile Switching Center (MSC) 280. The MSC280 is configured to interface with a Public Switched Telephone Network (PSTN) 290. The MSC280 is also configured to interface with a BSC275, which may be coupled to the base station 270 via a backhaul. The backhaul may be constructed according to any of several known interfaces including, for example, E1/T1, ATM, IP, PPP, frame Relay, HDSL, ADSL, or xDSL. It will be understood that a system as shown in fig. 2 may include multiple BSCs 2750.
Each BS270 may serve one or more sectors (or regions), each sector covered by a multi-directional antenna or an antenna pointing in a particular direction being radially distant from the BS 270. Alternatively, each partition may be covered by two or more antennas for diversity reception. Each BS270 may be configured to support multiple frequency allocations, with each frequency allocation having a particular frequency spectrum (e.g., 1.25MHz,5MHz, etc.).
The intersection of partitions with frequency allocations may be referred to as a CDMA channel. The BS270 may also be referred to as a Base Transceiver Subsystem (BTS) or other equivalent terminology. In such a case, the term "base station" may be used to generically refer to a single BSC275 and at least one BS 270. The base stations may also be referred to as "cells". Alternatively, each sector of a particular BS270 may be referred to as a plurality of cell sites.
As shown in fig. 2, a Broadcast Transmitter (BT)295 transmits a broadcast signal to the mobile terminal 100 operating within the system. A broadcast receiving module 111 as shown in fig. 1 is provided at the mobile terminal 100 to receive a broadcast signal transmitted by the BT 295. In fig. 2, several Global Positioning System (GPS) satellites 300 are shown. The satellite 300 assists in locating at least one of the plurality of mobile terminals 100.
In fig. 2, a plurality of satellites 300 are depicted, but it is understood that useful positioning information may be obtained with any number of satellites. The GPS module 115 as shown in fig. 1 is generally configured to cooperate with satellites 300 to obtain desired positioning information. Other techniques that can track the location of the mobile terminal may be used instead of or in addition to GPS tracking techniques. In addition, at least one GPS satellite 300 may selectively or additionally process satellite DMB transmission.
As a typical operation of the wireless communication system, the BS270 receives reverse link signals from various mobile terminals 100. The mobile terminal 100 is generally engaged in conversations, messaging, and other types of communications. Each reverse link signal received by a particular base station 270 is processed within the particular BS 270. The obtained data is forwarded to the associated BSC 275. The BSC provides call resource allocation and mobility management functions including coordination of soft handoff procedures between BSs 270. The BSCs 275 also route the received data to the MSC280, which provides additional routing services for interfacing with the PSTN 290. Similarly, the PSTN290 interfaces with the MSC280, the MSC interfaces with the BSCs 275, and the BSCs 275 accordingly control the BS270 to transmit forward link signals to the mobile terminal 100.
Based on the above mobile terminal hardware structure and communication system, various embodiments of the apparatus and method of the present invention are proposed. In order to make the objects, technical solutions and advantages of the present invention more apparent, embodiments of the present invention will be described in detail below with reference to the accompanying drawings. It should be noted that the embodiments and features of the embodiments in the present application may be arbitrarily combined with each other without conflict.
The steps illustrated in the flow charts of the figures may be performed in a computer system such as a set of computer-executable instructions. Also, while a logical order is shown in the flow diagrams, in some cases, the steps shown or described may be performed in an order different than here.
The technical solution of the present invention is described in detail below with specific embodiments, and the terminal device in the following embodiments of the present invention is an intelligent terminal capable of executing a navigation program, such as a smart phone, a vehicle-mounted intelligent terminal, a Personal Digital Assistant (PDA), a tablet computer, or the like. The following specific embodiments of the present invention may be combined, and the same or similar concepts or processes may not be described in detail in some embodiments.
Fig. 3 is a schematic structural diagram of a navigation speech processing apparatus according to an embodiment of the present invention. The navigation voice processing apparatus provided in this embodiment is suitable for a case where a dual speaker is used to play a voice in a navigation application, and the navigation voice processing apparatus is usually implemented by a method combining hardware and software, and the navigation voice processing apparatus may be integrated in a controller of a terminal device and may be invoked by the controller. As shown in fig. 3, the navigation voice processing apparatus provided in this embodiment may include: a recognition module 10 and a voice playing module 20.
The recognition module 10 is configured to recognize a prompt language tag of a voice to be played in a navigation program, where the prompt language tag includes: the left prompt and the right prompt.
The navigation voice processing device provided by the embodiment of the invention provides a mode for navigating and playing navigation voice through terminal equipment, the navigation program described in the embodiment of the invention can be a navigation Application program (APP) installed on the terminal equipment, such as a Baidu map, a Google map or Gauder navigation installed on a smart phone, and the navigation program can also be driving navigation configured on a vehicle.
In an embodiment of the present invention, the recognition module 10 may recognize some prompt tags of the voice to be played from the navigation program, where the prompt tags generally include a left prompt or a right prompt, and may be tags corresponding to each navigation voice in a navigation voice library of the navigation program, for example. For example, a voice having left-side directivity such as a navigation voice "driving left ahead", "turning left ahead", or the like corresponds to a cue tag having left-side directivity, i.e., a left cue, and a voice having right-side directivity such as a navigation voice "driving right ahead", "turning right ahead", or the like corresponds to a cue tag having right-side directivity, i.e., a right cue; therefore, the navigation voices in the navigation voice library of the navigation program can be classified through the prompt language tags, so that each navigation voice has a corresponding prompt language tag.
It should be noted that the types of the prompt tags may be, for example, a fixed number, and the prompt tags of the same type may be carried in the navigation voice with the same attribute; the types of the prompt tags are classified based on, for example, the voice directionality, such as the left-side directionality, the right-side directionality, and the non-directionality.
And the voice playing module 20 is configured to play the voice to be played by using a left speaker or a right speaker of the terminal device according to the prompt language tag of the voice to be played, which is identified by the identifying module 10.
In the embodiment of the present invention, the recognition module 10 may recognize a prompt tag of the voice to be played, for example, a tag with directivity, so that the voice playing module 20 may play the voice to be played through different speakers of the terminal device according to the prompt tag recognized by the recognition module 10. In the embodiment of the present invention, the speakers of the terminal device are configured as dual speakers, the dual speakers may be a left speaker and a right speaker that are configured on both sides of the terminal device, or may be speakers that are configured on both upper and lower ends of the terminal device, and when the terminal device enters a navigation program and is in a specific application scenario, the speakers at both ends of the terminal device may be regarded as the left speaker and the right speaker. In practical application, according to different prompt language tags identified by the identification module 10, a left speaker or a right speaker is adopted to play navigation voice; for example, if the cue word label recognized by the recognition module 10 is a left cue word, that is, the playback content corresponding to the left speaker, the left speaker plays the cue word, and if the cue word label recognized by the recognition module 10 is a right cue word, that is, the playback content corresponding to the right speaker, the right speaker plays the cue word.
Taking the terminal device in this embodiment as an example of a smart phone, the dual speakers of the smart phone are usually configured at the top and the bottom, and if the smart phone is in the landscape mode, the two speakers of the smart phone are located at the left and right sides, and at this time, the sound emitted by the speakers is captured by the ear closer to the ear first and captured by the ear farther from the ear; moreover, the user can determine the direction of the sound source by the sequence of the sound arriving at the ear, as shown in fig. 4, which is a schematic diagram of a principle that a dual speaker propagates audio in the navigation voice processing apparatus provided in the embodiment of the present invention, as can be seen from fig. 4, if the left speaker of the terminal device plays audio and the left ear of the user captures the sound first, the user can determine that the audio is played by the left speaker, and if the right speaker of the terminal device plays audio and the right ear of the user captures the sound first, the user can determine that the audio is played by the right speaker.
For example, based on the above-mentioned principle of playing audio by two speakers, when the terminal device is used for navigation, especially for driving navigation, the two speakers of the terminal device satisfy the position requirements of the left speaker and the right speaker, i.e. the conditions for generating stereo effect by the two speakers. When the navigation program of the terminal device plays the voice, the voice to be played in the navigation program is analyzed by the recognition module 10, and the voice to be played with the left attribute (for example, xx is left turn, xx is left front, etc.) can be emitted from the speaker (namely, left side speaker) of the terminal device close to the left ear of the user according to the prompt language label, and then the voice can be received by the left ear of the user firstly, so that the user feels that the voice is transmitted from the left; similarly, a voice to be played with "right attribute" (e.g., right turn, xx on the front right, etc.) is emitted from a speaker of the terminal device near the right ear of the user (i.e., right speaker), and is received by the right ear of the user first, so that the user feels that the voice is coming from the right side.
The navigation voice processing apparatus provided in this embodiment recognizes a prompt label of a voice to be played in a navigation program through the recognition module 10, where the prompt label includes a left prompt and a right prompt, and then the voice playing module 20 plays the voice to be played through a left speaker or a right speaker of the terminal device according to the prompt label of the voice to be played recognized by the recognition module 10; the embodiment effectively applies the double loudspeakers to navigation playing through reasonable design in a mode of performing navigation voice playing through the double loudspeakers, achieves the effect of navigation voice playing rich in stereoscopic impression, and has stronger indicativity, thereby improving the application value of the double loudspeakers.
Compared with the prior art, the navigation voice processing device provided by the embodiment of the invention has the advantages that the double loudspeakers are applied to the entertainment scene for playing music to improve the audio effect, the position characteristics of the double loudspeakers are utilized, the double loudspeakers are combined with the navigation voice playing based on the principle of sound propagation, and the navigation voice processing device with higher application value for user navigation, particularly driving navigation, is provided.
Optionally, fig. 5 is a schematic structural diagram of another navigation speech processing apparatus according to an embodiment of the present invention. On the basis of the structure of the navigation voice processing device shown in fig. 3, the device provided in this embodiment further includes:
and the adding module 30 is used for adding a cue language tag to the navigation voice in the navigation program before the recognition module 10 recognizes the cue language tag of the voice to be played in the navigation program.
Because the recognition module 10 in the embodiment of the present invention can recognize the prompt language tag of the voice to be played of the navigation program, so as to play the voice through different speakers, however, when the navigation program in the prior art plays the voice, the speakers do not need to be distinguished, and all the navigation voice can be played simultaneously from the two speakers configured in the terminal device; therefore, the apparatus provided in the embodiment of the present invention, before the recognition module 10 performs recognition, first needs to add a prompt tag to the navigation voice in the navigation program, for example, the navigation program has a navigation voice library, and the adding module 30 adds a corresponding prompt tag to all the navigation voices in the navigation voice library, where the prompt tag may be generally divided according to the directionality of the navigation voice, and for example, may include the following embodiments: the left prompt and the right prompt. Based on the contents of the above embodiments, the left direction cue may be added to the "forward left turn", "xx before left, etc." navigation voice, and the right direction cue may be added to the "forward right turn", "xx before left, etc." navigation voice.
Based on the category of the prompt language tag in this embodiment, the voice playing module 20 uses the left speaker or the right speaker of the terminal device to play the voice to be played according to the prompt language tag to be played, which is identified by the identification module 10, and may include:
when the prompt language label is a left prompt language, playing the voice to be played through a left loudspeaker of the terminal equipment;
and when the prompt language label is the right prompt language, playing the voice to be played through a right loudspeaker of the terminal equipment.
Furthermore, the prompt label in the embodiment of the present invention may further include a non-directional prompt, and the non-directional prompt may be added to navigation voices such as "go straight ahead", "ahead roadblock, and slow passing"; correspondingly, in the apparatus provided in the embodiment of the present invention, the voice playing module 20 is further configured to play the voice to be played through a left speaker and a right speaker of the terminal device when the prompt label is the non-directional prompt. In this embodiment, the navigation voice may also be played in a manner of combining the left speaker and the right speaker, for example, if the prompt tag recognized by the recognition module 10 is a non-directional prompt, that is, if the prompt tag does not have a corresponding relationship with a specific speaker, the navigation voice may be played through the left speaker and the right speaker at the same time.
In the embodiment of the invention, the prompt language label is associated with the direction characteristic, and a specific mode of playing voice through different speakers according to the prompt language label is provided, so that the voice direction played by the navigation program and the sounding direction of the speakers felt by the user have the same straightness, the indication effect of playing voice by the navigation program is favorably improved, and the intuitive feeling of the user can be improved. In addition, the embodiment of the invention fully utilizes the stereophonic technology configured by the double loudspeakers, applies the technology to a new scene, and improves the application value of the double loudspeakers.
Optionally, in an application scenario of the embodiment of the present invention, the left speaker and the right speaker of the terminal device may be external speakers of the terminal device, for example, when the terminal device is a vehicle-mounted navigation device, the left speaker is a speaker installed in a left door (a door beside a cab) of a vehicle, the right speaker is a speaker installed in a right door (a door beside a co-cab) of the vehicle, and for another example, the terminal device is a smart phone or an independent navigation device, and the left speaker and the right speaker may be external speakers electrically connected to the terminal device or connected to bluetooth.
Optionally, in another application scenario of the embodiment of the present invention, the left speaker and the right speaker of the terminal device may be speakers disposed on the terminal device, for example, the terminal device is a smartphone or a tablet computer, and the dual speakers are configured at the top and the bottom of the terminal device, and when the terminal device is in the landscape mode, the dual speakers may be regarded as the left speaker and the right speaker. In the application scenario of the embodiment of the present invention, the recognition module 10 is further configured to recognize that, before the voice playing module 20 plays the voice to be played, when the terminal device is in the landscape mode, the speaker arranged at the left end of the navigation interface of the terminal device is the left speaker, and the speaker arranged at the right end of the navigation interface is the right speaker.
It can be understood that when the terminal device starts the navigation program and executes the navigation task, the terminal device may be placed in a landscape mode to improve the display effect, and generally, the terminal device has a judgment mode of automatic screen selection, so that it can be judged that the terminal device has entered a landscape mode; however, according to different habits of users, the top of the terminal device is placed on the left side by using habits of some users, the bottom of the terminal device is placed on the left side by using habits of some users, and the bottom of the terminal device is placed on the left side by using arbitrary horizontal screen.
Compared with the traditional voice navigation, the navigation voice processing device provided by the embodiment of the invention has the advantages that the stereo effect is generated when the navigation voice is played, the direction information of the voice prompt is recognized from the content of the navigation voice, and the direction information of the navigation voice prompt can be sensed from the auditory reception, so that the navigation is realized for a person in the corresponding direction in front like the person speaks. The prompt effect of navigation voice playing is improved, the impression of the user receiving the prompt voice is more profound, the accuracy is high, and mistakes are avoided.
Fig. 6 is a flowchart illustrating a navigation voice processing method according to an embodiment of the present invention. The navigation voice processing method provided by the embodiment is suitable for the situation that the voice in the navigation application is played by adopting the double loudspeakers, the navigation voice processing method can be executed by the navigation voice processing device provided by the embodiment of the invention, the device is usually realized by a method of combining hardware and software, and the device can be integrated in a controller of terminal equipment and can be called and used by the controller. As shown in fig. 6, the navigation voice processing method of the present embodiment may include the following steps, i.e., S110 to S120:
s110, identifying a prompt language tag of the voice to be played in the navigation program, wherein the prompt language tag comprises: the left prompt and the right prompt.
The navigation voice processing method provided by the embodiment of the invention provides a mode of navigating through the terminal equipment and playing navigation voice, the navigation program described in the embodiment of the invention can be a navigation APP installed on the terminal equipment, such as a Baidu map, a Google map or a Goods navigation installed on a smart phone, and the navigation program can also be driving navigation configured on a vehicle.
In the embodiment of the present invention, the terminal device may recognize some prompt language tag of the voice to be played from the navigation program, where the prompt language tag usually includes a left prompt language or a right prompt language, and may be a tag corresponding to each navigation voice in a navigation voice library of the navigation program, for example. For example, a voice having left-side directivity such as a navigation voice "driving left ahead", "turning left ahead", or the like corresponds to a cue tag having left-side directivity, i.e., a left cue, and a voice having right-side directivity such as a navigation voice "driving right ahead", "turning right ahead", or the like corresponds to a cue tag having right-side directivity, i.e., a right cue; therefore, the navigation voices in the navigation voice library of the navigation program can be classified through the prompt language tags, so that each navigation voice has a corresponding prompt language tag.
It should be noted that the types of the prompt tags may be, for example, a fixed number, and the prompt tags of the same type may be carried in the navigation voice with the same attribute; the types of the prompt tags are classified based on, for example, the voice directionality, such as the left-side directionality, the right-side directionality, and the non-directionality.
And S120, playing the voice to be played by adopting a left side loudspeaker or a right side loudspeaker of the terminal equipment according to the prompt language label of the voice to be played.
In the embodiment of the present invention, in S110, the cue word tag of the voice to be played is already recognized, for example, the cue word tag has directionality, so that the voice to be played can be played through different speakers of the terminal device according to the recognized cue word tag. In the embodiment of the present invention, the speakers of the terminal device are configured as dual speakers, which may be a left speaker and a right speaker configured at both sides of the terminal device, or speakers configured at both upper and lower ends of the terminal device, and when the terminal device enters the navigation program and is in a specific application scenario, the speakers at both upper and lower ends of the terminal device may be regarded as the left speaker and the right speaker. In practical application, the navigation voice can be played by adopting a left loudspeaker or a right loudspeaker according to different identified prompt language labels; for example, if the recognized cue word label is a left cue word, that is, the playback content corresponding to the left speaker, the left speaker plays the text, and if the recognized cue word label is a right cue word, that is, the playback content corresponding to the right speaker, the right speaker plays the text.
Taking the terminal device in this embodiment as an example of a smart phone, the dual speakers of the smart phone are usually configured at the top and the bottom, and if the smart phone is in the landscape mode, the two speakers of the smart phone are located at the left and right sides, and at this time, the sound emitted by the speakers is captured by the ear closer to the ear first and captured by the ear farther from the ear; moreover, the user can determine the direction of the sound source by the sequence of the sounds arriving at the ears, and referring to the schematic diagram of the principle of dual-speaker audio transmission shown in fig. 4, it can be seen from fig. 4 that if the left speaker of the terminal device plays the audio and the sound is captured by the left ear of the user first, the user can determine that the audio is played by the left speaker, and if the right speaker of the terminal device plays the audio and the sound is captured by the right ear of the user first, the user can determine that the audio is played by the right speaker.
For example, based on the above-mentioned principle of playing audio by two speakers, when the terminal device is used for navigation, especially for driving navigation, the two speakers of the terminal device satisfy the position requirements of the left speaker and the right speaker, i.e. the conditions for generating stereo effect by the two speakers. When the navigation program of the terminal equipment plays voice, the voice to be played in the navigation program is analyzed, and the voice to be played with left attribute (such as left turn, xx in the front left and the like) can be emitted from a loudspeaker (namely a left loudspeaker) of the terminal equipment close to the left ear of the user according to the prompt language label, and then the voice can be received by the left ear of the user firstly, so that the user feels that the voice is transmitted from the left of the user; similarly, a voice to be played with "right attribute" (e.g., right turn, xx on the front right, etc.) is emitted from a speaker of the terminal device near the right ear of the user (i.e., right speaker), and is received by the right ear of the user first, so that the user feels that the voice is coming from the right side.
According to the navigation voice processing method provided by the embodiment, the prompt label of the voice to be played in the navigation program is identified, the prompt label comprises a left prompt and a right prompt, then the voice to be played is played by adopting a left loudspeaker or a right loudspeaker of the terminal equipment according to the identified prompt label of the voice to be played, and when the navigation voice is played by adopting the method provided by the embodiment, the navigation voice with different prompt labels can be played stereoscopically by utilizing the position characteristics of the two loudspeakers, so that the direction sources of different navigation voices are embodied, and the navigation voice is presented with the characteristic of directionality; the embodiment effectively applies the double loudspeakers to navigation playing through reasonable design in a mode of performing navigation voice playing through the double loudspeakers, achieves the effect of navigation voice playing rich in stereoscopic impression, and has stronger indicativity, thereby improving the application value of the double loudspeakers.
Compared with the prior art, the navigation voice processing method provided by the embodiment of the invention has the advantages that the application scene related to the life of a user is considered, the position characteristics of the double loudspeakers are utilized, the double loudspeakers are combined with the navigation voice playing based on the principle of sound propagation, and the navigation voice processing method with higher application value to the user navigation, particularly the driving navigation, is provided.
Optionally, fig. 7 is a schematic flowchart of another navigation speech processing method according to an embodiment of the present invention. On the basis of the embodiment shown in fig. 6, before S110, the method provided in this embodiment may further include:
and S100, adding a prompt language tag to navigation voice in the navigation program.
Because the prompt language label of the voice to be played of the navigation program can be identified in the S110 of the embodiment of the present invention, the voice is played through different speakers, however, when the navigation program in the prior art plays the voice, the speakers do not need to be distinguished, and all the navigation voice can be played simultaneously from the dual speakers configured in the terminal device; therefore, before the identification in S110, the method provided in the embodiment of the present invention first needs to add a prompt tag to the navigation voice in the navigation program, for example, the navigation program has a navigation voice library, that is, corresponding prompt tags are added to all the navigation voices in the navigation voice library, where the prompt tags may be generally divided according to the directionality of the navigation voice, and for example, the method may include: the left prompt and the right prompt. Based on the contents of the above embodiments, the left direction cue may be added to the "forward left turn", "xx before left, etc." navigation voice, and the right direction cue may be added to the "forward right turn", "xx before left, etc." navigation voice.
Based on the category of the hint label in this embodiment, the implementation manner of S120 in this embodiment of the present invention may include:
s121, when the prompt label is a left prompt, playing the voice to be played through a left loudspeaker of the terminal equipment;
and S122, when the prompt label is the right prompt, playing the voice to be played through a right loudspeaker of the terminal equipment.
Furthermore, the prompt label in the embodiment of the present invention may further include a non-directional prompt, and the non-directional prompt may be added to navigation voices such as "go straight ahead", "ahead roadblock, and slow passing"; correspondingly, the method provided by the embodiment of the invention can further comprise the following steps:
and S130, when the prompt language label is the non-directional prompt language, playing the voice to be played through a left side loudspeaker and a right side loudspeaker of the terminal equipment. In this embodiment, the navigation voice may also be played in a manner of combining the left speaker and the right speaker, for example, if the prompt tag identified in S110 is a non-directional prompt, that is, does not have a corresponding relationship with a specific speaker, the navigation voice may be played through the left speaker and the right speaker at the same time.
It should be noted that S121, S122, and S130 in the embodiment shown in fig. 7 are steps selectively executed, and S121, S122, or S130 is executed according to the specific type of the hint tag identified in S110.
In the embodiment of the invention, the prompt language label is associated with the direction characteristic, and a specific mode of playing voice through different speakers according to the prompt language label is provided, so that the voice direction played by the navigation program and the sounding direction of the speakers felt by the user have the same straightness, the indication effect of playing voice by the navigation program is favorably improved, and the intuitive feeling of the user can be improved. In addition, the embodiment of the invention fully utilizes the stereophonic technology configured by the double loudspeakers, applies the technology to a new scene, and improves the application value of the double loudspeakers.
Optionally, in an application scenario of the embodiment of the present invention, a left speaker and a right speaker of a terminal device are external speakers of the terminal device, for example, when the terminal device is a vehicle-mounted navigation device, the left speaker is a speaker installed in a left door (a door beside a cab) of a vehicle, the right speaker is a speaker installed in a right door (a door beside a co-cab) of the vehicle, and for another example, the terminal device is a smart phone or an independent navigation device, and the left speaker and the right speaker may be external speakers electrically connected to the terminal device or connected to bluetooth.
Optionally, in another application scenario of the embodiment of the present invention, the left speaker and the right speaker of the terminal device are speakers disposed on the terminal device, for example, the terminal device is a smartphone or a tablet computer, and the dual speakers are configured at the top and the bottom of the terminal device, and when the terminal device is in the landscape mode, the dual speakers may be regarded as the left speaker and the right speaker. In this application scenario of the embodiment of the present invention, before S120, the method may further include: when the terminal equipment is in a landscape mode, the loudspeaker arranged at the left end of the navigation interface of the terminal equipment is identified as a left loudspeaker, and the loudspeaker arranged at the right end of the navigation interface is identified as a right loudspeaker.
It can be understood that when the terminal device starts the navigation program and executes the navigation task, the terminal device may be placed in a landscape mode to improve the display effect, and generally, the terminal device has a judgment mode of automatic screen selection, so that it can be judged that the terminal device has entered a landscape mode; however, according to different habits of users, the top of the terminal device is placed on the left side by using habits of some users, the bottom of the terminal device is placed on the left side by using habits of some users, and the bottom of the terminal device is placed on the left side by using arbitrary horizontal screen.
The navigation voice processing method provided by the embodiment of the present invention is described below by an application example, and fig. 8 is a schematic flow diagram of another navigation voice processing method provided by the embodiment of the present invention. In this embodiment, a terminal device is taken as an example to show that the method provided in this embodiment may include: s301 to S308:
s301: after a navigation APP is opened, judging whether the smart phone is in a horizontal screen mode; if the smart phone is not in the horizontal screen mode, ending the process; if the smart phone is in the landscape mode, executing S302;
s302, identifying that the loudspeaker arranged at the left end of the navigation interface of the smart phone is the left loudspeaker, and identifying that the loudspeaker arranged at the right end of the navigation interface is the right loudspeaker.
S303, identifying a prompt language label of the voice to be played in the navigation program; the prompt label in this embodiment may also include: a left prompt, a right prompt and a non-directional prompt.
S304, judging whether the identified prompt label is a non-directional prompt or not; if yes, go to S305; if not, go to S306.
S305, playing the voice to be played through a left loudspeaker and a right loudspeaker of the smart phone.
S306, judging whether the recognized prompt label is a left prompt or not; if yes, go to S307; if not, go to S308.
And S307, playing the voice to be played through a left loudspeaker of the smart phone.
And S308, playing the voice to be played through a right loudspeaker of the smart phone.
It should be noted that, the contents in S304 and S306 in the embodiment shown in fig. 8 may be appropriately adjusted, for example, it may be determined whether the prompt label is a right prompt in S306; for another example, if the prompt label is determined to be the left prompt in S304, then it may be determined whether the prompt label is the non-directional prompt or the right prompt in S306; the embodiment shown in fig. 8 is only one possible implementation manner of determining a hint tag, and is not limited to the implementation manner of the navigation voice processing method provided by the embodiment of the present invention.
Compared with the traditional voice navigation, the navigation voice processing method provided by the embodiment of the invention has the advantages that the stereo effect is generated when the navigation voice is played, the direction information of the voice prompt is recognized from the content of the navigation voice, and the direction information of the navigation voice prompt can be sensed from the auditory reception, like a person shouting to navigate in the corresponding direction in front. The prompt effect of navigation voice playing is improved, the impression of the user receiving the prompt voice is more profound, the accuracy is high, and mistakes are avoided.
Although the embodiments of the present invention have been described above, the above description is only for the convenience of understanding the present invention, and is not intended to limit the present invention. It will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (10)

1. A navigation speech processing apparatus, characterized in that the apparatus comprises:
the recognition module is used for recognizing a prompt language tag of voice to be played in the navigation program, wherein the prompt language tag comprises: a left prompt and a right prompt;
and the voice playing module is used for playing the voice to be played by adopting a left side loudspeaker or a right side loudspeaker of the terminal equipment according to the prompt language label of the voice to be played, which is identified by the identification module.
2. The navigation voice processing apparatus according to claim 1, characterized in that the apparatus further comprises:
and the adding module is used for adding a prompt language label to the navigation voice in the navigation program before the recognition module recognizes the prompt language label of the voice to be played in the navigation program.
3. The navigation voice processing apparatus according to claim 1, wherein the voice playing module uses a left speaker or a right speaker of the terminal device to play the voice to be played according to the prompt language tag of the voice to be played, which is identified by the identification module, and includes:
when the prompt language label is the left prompt language, playing the voice to be played through a left side loudspeaker of the terminal equipment;
when the prompt label is the right prompt, the voice to be played is played through a right loudspeaker of the terminal equipment;
the prompt language label also comprises a non-directional prompt language, and the voice playing module is further used for playing the voice to be played through a left side loudspeaker and a right side loudspeaker of the terminal device when the prompt language label is the non-directional prompt language.
4. The navigation voice processing device according to any one of claims 1 to 3, wherein the left side speaker and the right side speaker are speakers provided on the terminal device; or,
the left side loudspeaker and the right side loudspeaker are external loudspeakers of the terminal equipment.
5. The navigation voice processing apparatus according to claim 4, wherein when the left speaker and the right speaker are speakers disposed on the terminal device, the recognition module is further configured to recognize, before the voice playing module plays the voice to be played, that a speaker disposed at a left end of a navigation interface of the terminal device is the left speaker and that a speaker disposed at a right end of the navigation interface is the right speaker when the terminal device is in a landscape mode.
6. A navigation voice processing method is characterized by comprising the following steps:
identifying a prompt language tag of voice to be played in a navigation program, wherein the prompt language tag comprises: a left prompt and a right prompt;
and according to the prompt language label of the voice to be played, playing the voice to be played by adopting a left side loudspeaker or a right side loudspeaker of the terminal equipment.
7. The navigation voice processing method according to claim 6, wherein before the recognition of the prompt language tag of the voice to be played in the navigation program, the method further comprises:
and adding a prompt language tag to the navigation voice in the navigation program.
8. The navigation voice processing method according to claim 6, wherein the playing the voice to be played by using a left speaker or a right speaker of the terminal device according to the prompt language tag of the voice to be played comprises:
when the prompt language label is the left prompt language, playing the voice to be played through a left side loudspeaker of the terminal equipment;
when the prompt label is the right prompt, the voice to be played is played through a right loudspeaker of the terminal equipment;
the prompt label further comprises a non-directional prompt, and the method further comprises:
and when the prompt language label is the non-directional prompt language, playing the voice to be played through a left side loudspeaker and a right side loudspeaker of the terminal equipment.
9. The navigation voice processing method according to any one of claims 6 to 8, wherein the left side speaker and the right side speaker are speakers provided on the terminal device; or,
the left side loudspeaker and the right side loudspeaker are external loudspeakers of the terminal equipment.
10. The navigation voice processing method according to claim 9, wherein when the left speaker and the right speaker are speakers provided on the terminal device, before the voice to be played is played, the method further includes:
when the terminal equipment is in a horizontal screen mode, a loudspeaker arranged at the left end of a navigation interface of the terminal equipment is identified as the left side loudspeaker, and a loudspeaker arranged at the right end of the navigation interface is identified as the right side loudspeaker.
CN201610976685.2A 2016-10-27 2016-10-27 Navigation voice processing device and method Pending CN106534527A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610976685.2A CN106534527A (en) 2016-10-27 2016-10-27 Navigation voice processing device and method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610976685.2A CN106534527A (en) 2016-10-27 2016-10-27 Navigation voice processing device and method

Publications (1)

Publication Number Publication Date
CN106534527A true CN106534527A (en) 2017-03-22

Family

ID=58349816

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610976685.2A Pending CN106534527A (en) 2016-10-27 2016-10-27 Navigation voice processing device and method

Country Status (1)

Country Link
CN (1) CN106534527A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111402879A (en) * 2020-03-19 2020-07-10 斑马网络技术有限公司 Vehicle navigation prompt voice control method, device, equipment and medium
CN112033417A (en) * 2020-09-29 2020-12-04 北京深睿博联科技有限责任公司 Real-time navigation method and device for visually impaired people
CN112911065A (en) * 2021-01-18 2021-06-04 咪咕音乐有限公司 Audio playing method and device for terminal, electronic equipment and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090216441A1 (en) * 2005-03-31 2009-08-27 The University Of Waikato ,A University Method and system for indicating the location of an object
CN101606040A (en) * 2007-02-13 2009-12-16 索尼爱立信移动通讯有限公司 The system and method that is used for broadcasting navigation prompts
US20130191753A1 (en) * 2012-01-25 2013-07-25 Nobukazu Sugiyama Balancing Loudspeakers for Multiple Display Users
EP2720001A1 (en) * 2012-10-15 2014-04-16 GN Store Nord A/S A navigation system with a hearing device
US20150338232A1 (en) * 2014-05-26 2015-11-26 Deutsche Telekom Ag System and method for providing geographically based sound cues to a user of a mobile device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090216441A1 (en) * 2005-03-31 2009-08-27 The University Of Waikato ,A University Method and system for indicating the location of an object
CN101606040A (en) * 2007-02-13 2009-12-16 索尼爱立信移动通讯有限公司 The system and method that is used for broadcasting navigation prompts
US20130191753A1 (en) * 2012-01-25 2013-07-25 Nobukazu Sugiyama Balancing Loudspeakers for Multiple Display Users
EP2720001A1 (en) * 2012-10-15 2014-04-16 GN Store Nord A/S A navigation system with a hearing device
US20150338232A1 (en) * 2014-05-26 2015-11-26 Deutsche Telekom Ag System and method for providing geographically based sound cues to a user of a mobile device

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111402879A (en) * 2020-03-19 2020-07-10 斑马网络技术有限公司 Vehicle navigation prompt voice control method, device, equipment and medium
CN111402879B (en) * 2020-03-19 2023-10-31 斑马网络技术有限公司 Vehicle navigation prompt voice control method, device, equipment and medium
CN112033417A (en) * 2020-09-29 2020-12-04 北京深睿博联科技有限责任公司 Real-time navigation method and device for visually impaired people
CN112033417B (en) * 2020-09-29 2021-08-24 北京深睿博联科技有限责任公司 Real-time navigation method and device for visually impaired people
CN112911065A (en) * 2021-01-18 2021-06-04 咪咕音乐有限公司 Audio playing method and device for terminal, electronic equipment and storage medium
CN112911065B (en) * 2021-01-18 2022-04-26 咪咕音乐有限公司 Audio playing method and device for terminal, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
CN106888158B (en) Instant messaging method and device
CN106598457B (en) A kind of sound control method for playing back, device and terminal
CN106534566B (en) Frameless mobile terminal and information display method thereof
CN104811865B (en) Audio-frequency inputting method, device and system
CN106157970B (en) Audio identification method and terminal
CN106357905B (en) A kind of mobile terminal
CN106547439B (en) Method and device for processing message
CN105353947A (en) Mobile terminal and application display content controlling method
CN106648324B (en) Hidden icon control method and device and terminal
CN106101455A (en) Page jump management method, device and terminal
CN106250130B (en) Mobile terminal and method for responding key operation
CN107018334A (en) A kind of applied program processing method and device based on dual camera
CN106453863B (en) Method and system for controlling terminal and earphone
CN106534527A (en) Navigation voice processing device and method
CN105261054A (en) Device and method for compositing audio GIF image
CN105049916B (en) A kind of video recording method and device
CN107241497B (en) Mobile terminal and loudness output adjusting method
CN107197084B (en) Method for projection between mobile terminals and first mobile terminal
CN106658116B (en) Play control method, device and terminal
CN109275038B (en) Game live broadcast method, terminal and computer readable storage medium
CN105407432B (en) A kind of method and system realizing audio and playing
CN106791042A (en) A kind of Voice Navigation processing method, device and terminal
CN105208193A (en) Button-free device and method
CN105141834A (en) Device and method for controlling picture shooting
CN104850351A (en) Method and device for frameless interaction

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20170322

RJ01 Rejection of invention patent application after publication