throbber
1
`
`VWGoA - Ex. 1007
`Volkswagen Group of America, Inc., Petitioner
`
`

`
`U.S. Patent
`
`May 16, 2000
`
`Sheet 1 of 9
`
`6,064,323
`
`F/5.7
`
`2
`
`

`
`U.S. Patent
`
`May 16, 2000
`
`Sheet 2 of 9
`
`6,064,323
`
`F/5.2
`
`3
`
`

`
`U.S. Patent
`
`May 16, 2000
`
`Sheet 3 of 9
`
`6,064,323
`
`.__________..__________—___
`
`
`
`«Hag_mH:mu_o:_
`
`u:<mama
`
`
`
`muou.mHum_m;Qno.
`
`
`
`mamamu=y_m:o4
`
`Ea._:38.
`
`
`
`:o_m$>:8n_wo
`
`
`
`:o_~_:moommmo_o>aa--x
`
`
`
`H_:=-ulid.¢_m_N.
`
`:o_yumym=o_#m_w=mu
`
`:o_H_mo1u%owam.%¢o__m:w_m
`
`
`«:mmm.m.Hcw.<omu_>
`
`.|||||||nI||I||lu|.II'I.l|l|llI..InI.uI.|||II|luI||I|Il..|nl.I|.InIululllulllllnullllln
`
`uw:_my:oo
`
`.oHmm_>m:
`
`-+_mm
`
`___
`
`NN
`
`>mx
`
`va\»qN:o_Hmm_>mz
`
`m=ym.moa<
`
`~_:=>m_Qm_a
`
`4
`
`
`
`
`

`
`U.S. Patent
`
`May 16, 2000
`
`Sheet 4 of 9
`
`6,064,323
`
`F/5.4
`
`Prefecture
`
`Hokkaido
`
`Other
`Munici‘ Names
`palities
`
`Akabira City
`Asahikawa City
`
`Phonemic Model
`
`A0m0Fi
`Prefecture
`
`Aomori city
`
`Kanagawa
`Prefecture
`
`Ooiso Town
`00iS0*Ch0
`
`Kanagawa
`
`Prefecture
`
`Kawasaki City
`
`What
`
`time is it now?
`
`Present Time
`
`Time
`
`5
`
`

`
`tHCtaP3nu
`
`May 16, 2000
`
`Sheet 5 of 9
`
`6,064,323
`
`
`
`
`
`mu:w_Hm4mu:y_wco4moamzmm:_EaJmmu
`
`o.=uom+m_;
`
`m.b\..\
`
`
`
`>H_om;mx_;mm<
`
`
`
`>~_om__nmx<
`
`ou_mxxoI
`
`6
`
`
`
`
`

`
`U.S. Patent
`
`May 16, 2000
`
`Sheet 6 of 9
`
`6,064,323
`
`Recognize voice
`
`1
`
`'egistereu
`place name
`ecognized7
`
`YES
`
`Output recognized
`
`character data
`
`Convert place name to
`
`latitude /
`
`longitude
`
`Specific
`
`V°gT;é§fiihQ?an
`
`latitude
`Output
`/
`longitude data
`
`Determine code
`corresponding to
`recognized voice
`
`Output determine code
`
`process based on voice
`
`reco:nition
`
`7
`
`

`
`‘nonaD13U
`
`‘nunuup
`
`May 16, 2000
`
`Sheet 7 of 9
`
`6,064,323
`
`
`
`w:~m.maam:o_Hmm_>m::_
`
`
`
`m:_mm¢uo_a>m_am_u
`
`:o_ym.ym_mm.+o
`
`
`
`:o_y_moaumwm_uommmmzy
`
`
`
`+0mymuanscmmm
`
`
`
`;#_;>m_gm_g
`
`mymuawe
`
`m_N
`
`
`
`
`
`xum_#m:_mwme.m_<
`
`:o_w_moa
`
`
`
`uflmcwfimu.8um:Bm_mm_Lt;>m_Qm_m_
`
`NmymuQME
`
`>m_am__
`
`
`
`:o_«m:m_mmu:o_«_moo
`
`:o_ym.mao>mx>3
`
`
`
`m:_«m:w_mmu:o_y_moo
`
`:a:_>__mn_m>mno
`
`
`
`mo_o>mN_mm;~:>m
`
`m#m:_u.oou
`
`
`
`Emum.u:tm:o_
`
`:.:Q:_3_m€m>
`
`
`
`mo_o>mN_mm;#:>m
`
`mwmugas«socmmm
`
`noam_oommmmcw+o
`
`:o_«_moa
`
`‘%
`
`«:mmm.¢
`
`
`
`>m_am_u:o_w_moa
`
`Nance
`
`
`
`#:mmm.a«omymo
`
`:o_«_moa
`
`mwmuems«soummm
`
`u2o2mu+o
`
`:o_«_moa
`
`
`
`;#_;>m_am_a
`
`mymuems
`
`8
`
`
`
`
`
`
`
`
`
`
`
`

`
`U.S. Patent
`
`May 16, 2000
`
`Sheet 8 of 9
`
`6,064,323
`
`F/6.8
`
`Display map by
`
`voice input
`
`401
`
`402
`
` Cursor to
`be moved?
`
`403
`
`Move cursor to display
`
`Dosition
`
`
`
`N0
`
`Destination-
`
`YES
`
`'
`
`404
`
`405
`
`407
`
`
`
`Set route from
`
`Set route from
`
`present position to
`
`residence to cursor
`
`cursor position
`
`-osition
`
`Display destination mark,
`etc. destination-
`
`setting process
`
`408
`
`9
`
`

`
`U.S. Patent
`
`May 16, 2000
`
`Sheet 9 of 9
`
`6,064,323
`
`
`
`
`
`
`
` «HE-- ; 5:]
`
`
`‘-5..
`F0.‘.l"!‘.‘!/E
`
`
`
`
` \£ Suidounsh
`
`
`F/5.9/4
`
`F/5.98
`
`10
`
`

`
`6,064,323
`
`1
`
`NAVIGATION APPARATUS, NAVIGATION
`METHOD AND AUTOMOTIVE VEHICLES
`
`BACKGROUND OF THE INVENTION
`
`1. Field of the Invention
`
`The present invention relates to a navigation apparatus
`suitably mounted on an automotive vehicle for displaying a
`road map or the like, a navigation method and an automotive
`vehicle having such a navigation apparatus mounted
`thereon.
`
`2. Description of the Prior Art
`Various types of navigation apparatuses for use on an
`automotive vehicle are under development. The navigation
`apparatus comprises, for example, a large-capacity storage
`means such as a CD-ROM storing road map data, means for
`detecting the present position, and a display unit for dis-
`playing the road map of the neighborhood of the detected
`present position on the basis of the data read from the data
`storage means. The present position detection means
`includes a positioning system using a positioning earth
`satellite called the GPS (Global Positioning System) or a
`device operated using a self-contained navigation method by
`tracking the change in the present position from the starting
`point on the basis of the information including the vehicle
`running speed and the direction in which the vehicle is
`running.
`the map of the desired position as well as the
`Also,
`neighborhood of the present position can be displayed on the
`display unit by key operation as far as the associated map
`data are available.
`
`In this navigation apparatus for use on automotive
`vehicles, for example, the display unit is generally mounted
`in the vicinity of the driver’s seat so that the driver can check
`the map of the neighborhood of the present position while
`the vehicle is running or waiting for a traffic signal.
`The above-mentioned navigation apparatus is required to
`be operated in a manner not to interfere with the operation
`of the vehicle on which the apparatus is mounted. While the
`vehicle is moving, for example, complex operations of the
`navigation apparatus are prohibited. In other words, this
`navigation apparatus, when installed on the vehicle,
`is
`connected with a running condition detector (such as a
`parking brake switch). In this way, the apparatus can be fully
`operated only when the detector finds that the vehicle is
`stationary, and a complicated key operation is prohibited
`while the vehicle is not stationary (i.e., when the vehicle is
`moving).
`inconvenient that the key operation is
`It
`is, however,
`impossible for switching the display map while the vehicle
`is moving. Demand is high,
`therefore, for a navigation
`apparatus that can be manipulated in sophisticated manner
`without interfering with the vehicle operation.
`SUMMARY OF THE INVENTION
`
`In view of these points, the object of the present invention
`is to provide a navigation apparatus that facilitates a sophis-
`ticated operation of various devices including the navigation
`apparatus without interfering with the vehicle operation.
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`FIG. 1 is a perspective view showing the apparatus
`according to an embodiment built in an automotive vehicle.
`FIG. 2 is a perspective view showing the vicinity of the
`driver’s seat of an automotive vehicle into which the appa-
`ratus according to an embodiment is built in.
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`2
`FIG. 3 is a diagram showing a configuration according to
`an embodiment of the invention.
`
`FIG. 4 is a diagram for explaining a storage area con-
`figuration of a voice recognition memory according to an
`embodiment.
`
`FIG. 5 is a diagram for explaining a storage area con-
`figuration of a latitude/longitude conversion memory
`according to an embodiment.
`FIG. 6 is a flowchart showing the process based on voice
`recognition according to an embodiment.
`FIG. 7 is a flowchart showing the display process of the
`navigation apparatus according to an embodiment.
`FIG. 8 is a flowchart showing the destination-setting
`process according to an embodiment.
`FIG. 9 including FIGS. 9A and 9B is a diagram for
`explaining an example display of the destination according
`to an embodiment.
`
`DESCRIPTION OF THE PREFERRED
`EMBODIMENTS
`
`An embodiment of the present invention will be described
`with reference to the accompanying drawings.
`This embodiment is applied to a navigation apparatus
`mounted on an automotive vehicle. The manner in which the
`
`apparatus according to this embodiment is mounted on the
`vehicle will be explained with reference to FIGS. 1 and 2. As
`shown in FIG. 2, an automotive vehicle 50 has a steering
`wheel 51 mounted on the front of a driver’s seat 52.
`
`the driver seated in the seat 52 operates the
`Basically,
`navigation apparatus. Other occupants of the vehicle 50,
`however, may operate the navigation apparatus. A naviga-
`tion apparatus body 20 and a voice recognition unit 10
`connected to the navigation apparatus body 20 are installed
`in an arbitrary space (in the rear trunk, for example) in the
`vehicle 50, and a positioning signal-receiving antenna 21 is
`mounted outside of the vehicle body (or in the vehicle inside
`of the rear window) as described later.
`As seen from FIG. 2, which shows the neighborhood of
`the driver’s seat, a talk switch 18 and an operating key 27 for
`the navigation apparatus are arranged beside the steering
`wheel 51. The switch and keys are so arranged as not to
`interfere with the operation of the vehicle while running.
`Also, a display unit 40 connected with the navigation
`apparatus is arranged in such a position as not to interfere
`with the forward field of view of the driver. Aspeaker 32 for
`outputting a voice signal synthesized in the navigation
`apparatus 20 is mounted at such a position (beside the
`display unit 40, for example) that the output voice can reach
`the driver.
`
`the navigation apparatus according to this
`Further,
`embodiment is so constructed as to accept a voice input. For
`this purpose, a microphone 11 is mounted on a sun visor 53
`arranged in the upper part of the windshield in a way to pick
`up the speech of the driver seated in the driver’s seat 52.
`Also, the navigation apparatus body 20 according to this
`embodiment is connected with an engine control computer
`54 of the automotive vehicle, which supplies a pulse signal
`proportional to the vehicle speed.
`Now, the internal configuration of the navigation appara-
`tus will be explained with reference to FIG. 3. In this
`embodiment, the voice recognition unit 10, which is con-
`nected with the navigation apparatus 20, is also connected
`with the microphone 11. This microphone 11 preferably has
`a comparatively narrow directivity to pick up only the
`speech of the person seated in the driver’s seat.
`
`11
`
`11
`
`

`
`6,064,323
`
`3
`The voice signal picked up by the microphone 11 is
`supplied to an analog/digital converter 12, where it
`is
`sampled with a signal of a predetermined sampling fre-
`quency and converted into a digital signal. The digital voice
`signal output from the analog/digital converter 12 is applied
`to a digital voice processing circuit 13 including an inte-
`grated circuit called a DSP (Digital Signal Processor). The
`digital voice processing circuit 13 supplies the digital voice
`signal as vector data to the voice recognition circuit 14 by
`such means as bandwidth division or filtering.
`This voice recognition circuit 14 is connected with a voice
`recognition data storage ROM 15 and performs the recog-
`nition operation in accordance with a predetermined voice
`recognition algorithm (such as HMM or Hidden Markov
`Model). Aplurality of candidates are thus selected from the
`voice recognition phonemic models stored in the ROM 15,
`and the character data corresponding to the most coincident
`phonemic model among the candidates is read out.
`Explanation will now be made about the data storage
`condition of the voice recognition data storage ROM 15.
`According to this embodiment, only the place names and the
`words are recognized for giving instructions to operate the
`navigation apparatus. As shown by the setting of the storage
`area in FIG. 4, the place names registered include only the
`names of the prefectures and other municipalities (cities,
`wards, towns and villages) in Japan. In addition, character
`codes of the place names and phonemic models providing
`data for voice recognition of the place names are stored for
`each pair of prefecture and municipality.
`In Japan, for example, there are about 3500 municipalities
`over the whole country, and therefore about 3500 place
`names are stored. In the case of a town which is pronounced
`one of two ways “Machi” and “Cho”, however, two types of
`data including “xx Cho” and “xx Machi” are stored. In
`similar fashion, for the place name of a village which is also
`pronounced one of two ways “Son” and “Mura”, two types
`of names, i.e., “xx Son” and “xx Mura”, are stored for each
`of such villages.
`Also, as regards the municipalities located adjacent to a
`boundary of a prefecture the name of which is liable to be
`called by the wrong name,
`the prefectural name often
`mistaken for it
`is additionally registered. The city of
`“Kawasaki, Kanagawa Prefecture”, for example, is regis-
`tered also as “Kawasaki, Tokyo” including the name of the
`adjacent prefecture.
`Further, various character codes of words representing
`operating instructions to the navigation apparatus and cor-
`responding phonemic models are stored. The words include
`those indicating a display position such as “destination”,
`“starting point”, “intermediate place”, “residence”, and
`those for giving various operating instructions such as
`“What time is it now?” (a command asking about the present
`time), “Where are we now?” (a command asking about the
`present position), “What is the next” (a command asking
`about the next intersection), “How long to go?” (a command
`asking about
`the remaining distance to the destination),
`“What is the speed” (a command asking about the current
`speed), “What is the altitude” (a command asking about the
`altitude), “Which direction should we take?” (a command
`asking about the direction in which to run), and “Command
`list” (a command for displaying a list of recognizable
`commands).
`In the case where a character code corresponding to a
`phonemic model coinciding with the recognition obtained
`through a predetermined voice recognition algorithm from
`an input vector data through the voice recognition circuit 14
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`4
`represents a place name, then the particular character code
`is read from the ROM 15. The character code thus read is
`
`applied to a converter circuit 16. The converter circuit 16 is
`connected with a conversion data storage ROM 17, so that
`the longitude/latitude data and incidental data associated
`with the character data supplied from the voice recognition
`circuit 14 are read out of the ROM 17.
`
`Now, explanation will be made about the data storage
`condition of the conversion data storage ROM 17 according
`to this embodiment. In this embodiment, a storage area is set
`for each character code identical to that of the place name
`stored in the voice recognition data storage ROM 15. As
`shown in FIG. 5,
`the latitude/longitude data and display
`scale data as incidental data of each place name are stored
`for each character code representing a place name. Unlike
`the character code read from the voice recognition data
`storage ROM 15 that is expressed in katakana, the latitude/
`longitude conversion data storage ROM 17 has also stored
`therein character codes for display in kanji, hiragana and
`katakana, in addition to character codes of katakana.
`According to the present embodiment,
`the latitude/
`longitude data of each place name represents the absolute
`position of a municipal oflice (such as city oflice, ward
`oflice, town oflice or village oflice) of the area indicated by
`the particular place name. Also, the incidental data which are
`output together with the latitude/longitude data include the
`display character code and the display scale. The display
`scale data are set in several levels, for example, according to
`the size of the area indicated by each place name.
`The latitude/longitude data and the incidental data read
`from the latitude/longitude conversion data storage ROM 17
`are applied to output terminals 10a, 10b as an output of the
`voice recognition unit 10. The data produced at the output
`terminals 10a, 10b are applied to the navigation apparatus
`20. The voice recognition unit 10 according to this embodi-
`ment includes a talk switch 18 adapted to open and close in
`unlocked state (that is to say, turned on only as long as
`depressed). While this talk switch 18 is depressed, only the
`voice signal picked up by the microphone 11 is processed as
`described above by the circuits all the way from the analog/
`digital converter 12 to the latitude/longitude conversion
`circuit 16.
`
`Now, explanation will be made about a configuration of
`the navigation apparatus 20 connected with the voice rec-
`ognition unit 10. The navigation apparatus 20 comprises a
`GPS antenna 21. The positioning signal received from a
`GPS satellite by the antenna 21 is processed by a present
`position detecting circuit 22. The data thus received is
`analyzed to detect the present position. The present position
`data thus detected include the latitude/longitude data pro-
`viding the prevailing absolute position.
`The present position data thus detected are applied to an
`arithmetic circuit 23. The arithmetic circuit 23 functions as
`
`a system controller for controlling the operation of the
`navigation apparatus 20. The arithmetic circuit 23, in which
`a CD-ROM (optical disk) for storing the road map data is
`set, is connected with a CD-ROM driver 23 for reading the
`data stored in the CD-ROM, a RAM 25 for storing various
`data required for data processing, a vehicle speed sensor 26
`for detecting the behavior of the vehicle on which the
`navigation apparatus is mounted, and an operating key 27. In
`the case where the latitude/longitude coordinate data includ-
`ing the present position is obtained, the CD-ROM drive 24
`is controlled to read the road map data of the neighborhood
`of the position represented by the coordinate. The road map
`data read by the CD-ROM driver 24 is temporarily stored in
`
`12
`
`12
`
`

`
`6,064,323
`
`5
`the RAM 25. Display data for displaying the road map is
`produced by use of the road map data thus stored. In the
`process, the map is displayed from the display data in a scale
`set by the operation of the key 27 arranged at a predeter-
`mined position in the vehicle.
`The display data produced by the arithmetic circuit 23 are
`applied to a video signal producing circuit 28, which gen-
`erates a video signal of a predetermined format on the basis
`of the display data. The resulting video signal is applied to
`an output terminal 20c.
`The video signal output from the output terminal 20c is
`applied to a display unit 40. The display unit 40 performs the
`receiving process on the basis of the video signal and causes
`the road map or the like to be displayed on the display panel
`of the display unit 40.
`In addition to the road map of the neighborhood of the
`present position,
`the road map of an arbitrary position
`designated by the operation of the key 27 can also be
`displayed under the control of the arithmetic circuit 23. Also,
`specific coordinate positions representing “destination”,
`“starting point”, “intermediate place” and “residence” can
`be registered by the operation of the key 27. In the case
`where a specific coordinate position is registered, the coor-
`dinate position data thus registered (latitude/longitude data)
`are stored in the RAM 25.
`
`in the case where the vehicle speed sensor 26
`Also,
`detects that the vehicle is running, the arithmetic circuit 23
`rejects the operation of the key 27 except for comparatively
`minor ones.
`
`The navigation apparatus 20 also comprises a self-
`contained navigator 29 for computing the running speed of
`the vehicle accurately on the basis of a pulse signal corre-
`sponding to the vehicle speed supplied from an automotive
`engine control computer or the like, detecting the direction
`in which the vehicle is running on the basis of the output of
`a gyro sensor in the self-contained navigator 29, and mea-
`suring the present position autonomously from a position
`determined by the speed and the running direction. Under
`the circumstances where the present position is incapable of
`being detected by the present position detection circuit 22,
`for example, the present position is measured by the self-
`contained navigation method from the position last detected
`by the present position detection circuit 22.
`The arithmetic circuit 23 is also connected with a voice
`
`synthesis circuit 31. In the case where the arithmetic circuit
`23 is required to issue some instruction by voice, the voice
`synthesis circuit 31 is caused to synthesize the voice for the
`instruction and to produce the voice from a speaker 32
`connected to the voice synthesis circuit 31. The voice
`instructions include, for example, “We are approaching the
`destination”, “You should proceed to the right”, etc. These
`and various other instructions are issued by voice as required
`as a navigation apparatus. Also, the voice synthesis circuit
`31 synthesizes the voice recognized by the voice recognition
`unit 10 on the basis of the character data supplied thereto,
`and output the synthesized voice from the speaker 32. This
`process will be described later.
`The navigation apparatus 20 comprises input terminals
`20a, 20b supplied with the character code, latitude/longitude
`data and data incidental
`thereto output from the output
`terminals 10a, 10b of the voice recognition unit 10. These
`latitude/longitude data, data incidental thereto and the char-
`acter code data produced from the input terminals 20a, 20b
`are applied to the arithmetic circuit 23. The arithmetic circuit
`23, when supplied with the latitude/longitude and other data
`from the voice recognition unit 10, controls the CD-ROM
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`6
`driver 24 to read the road map data of the neighborhood of
`the particular latitude/longitude from the disk. The road map
`data read by the CD-ROM driver 24 are temporarily stored
`in the RAM 25. By use of the road map data thus stored,
`display data for displaying the road map are produced. The
`display data thus produced are used to display the map
`around the supplied latitude and longitude in the designated
`display scale incidental to the latitude/longitude data.
`On the basis of this display data, the video signal pro-
`ducing circuit 28 generates a video signal, and the display
`unit 40 is caused to display the road map of the coordinate
`point designated by the voice recognition unit 10.
`In the case where a character code representing a verbal
`instruction for operating the navigation apparatus is supplied
`from the output terminal 10b of the voice recognition unit
`10, the arithmetic circuit 23 identifies the verbal character
`code and performs related control operations. Assume that
`the verbal character code specifies a display position such as
`“destination”, “starting point”, “intermediate place” or “resi-
`dence”. It is decided whether the coordinate of the display
`position is registered in the RAM 25 or not. If it is registered
`so, the road map data of the neighborhood of the particular
`position is read from the disk by the CD-ROM driver 24.
`Also, when registering the display position of
`“destination”, “starting point”, “intermediate place” or
`“residence”, the voice of, say, “register destination” can be
`recognized and set. When an instruction is given for regis-
`tration of any of these display positions, the cursor position
`(indicated by a mark at an arbitrary position in the map by
`a predetermined operation of the key 27) on the map
`displayed on the display unit 40 is registered. The arithmetic
`circuit 23 according to this embodiment
`is adapted to
`automatically set the route up to the position of a destination
`or an intermediate place which may be registered. More
`specifically, what is considered the most appropriate route
`from the position registered as the residence or the present
`position detected by the present position detecting circuit 22
`up to the destination or the intermediate place is determined
`automatically by arithmetic operations. In the case where
`information on traffic jam or other road conditions is avail-
`able from an external source, however, the route can be set
`taking such information into account.
`On the other hand, assume that a character code data
`indicating the pronunciation of a recognized voice is sup-
`plied to the arithmetic circuit 23 from the voice recognition
`unit 10. The words represented by the character code are
`synthesized by the voice synthesis circuit 31 and output as
`a voice from the speaker 32 connected with the voice
`synthesis circuit 31. Suppose the voice recognition unit 10
`recognizes the voice as “Bunkyo Ward, Tokyo”,
`for
`example, the voice synthesis circuit 31 performs synthesis in
`a manner to generate a voice signal pronounced “Bunkyo
`Ward, Tokyo” on the basis of the character string data of the
`pronunciation recognized. The voice signal thus generated is
`output by way of the speaker 32.
`In such a case, according to the present embodiment,
`whenever a voice is recognized by the voice recognition unit
`10, the latitude/longitude data are supplied to the terminal
`20a of the navigation apparatus 20 substantially at the same
`time as the character code data representing the recognized
`pronunciation is applied to the terminal 20b. The arithmetic
`circuit 23, however, first causes the voice synthesis circuit
`31 to synthesize the recognized voice, and then causes the
`road map display data to be produced on the basis of the
`latitude/longitude data.
`Now, explanation will be made about displaying the road
`map using the voice recognition unit 10 and the navigation
`
`13
`
`13
`
`

`
`6,064,323
`
`7
`apparatus 20. First, the voice recognition operation of the
`voice recognition unit 10 will be described with reference to
`the flowchart of FIG. 6. The first step decides whether the
`talk switch 18 is on or not (step 101). In the case where the
`decision is that the talk switch 18 is on, the voice signal
`picked up by the microphone 11 during the on-time of the
`talk switch 18 is sampled by the analog/digital converter 12
`and processed by the digital voice processing circuit 13 into
`vector data (step 102). On the basis of this vector data, the
`voice recognition circuit 14 performs the voice recognition
`process (step 103).
`It is decided whether the voice of a place name stored in
`the voice recognition data storage ROM 15 (i.e., a place
`name registered in advance) has been recognized (step 104).
`In the case where the voice of a registered place name has
`been recognized, the character data for pronouncing the
`recognized place name is read out of the ROM 15 and output
`from the output terminal 10b (step 105). At the same time,
`the latitude/longitude data of the recognized place name are
`read from the latitude/longitude conversion data storage
`ROM 17 connected to the latitude/longitude conversion
`circuit 16 (step 106). The place names registered in the ROM
`15 represent domestic prefectures and municipalities, and
`therefore the voices of a place name are recognized in the
`form of “xx City, xx Prefecture”, “xx Ward, xx City” or the
`like (in the case under consideration, the ward name can be
`recognized even if the prefectural name is omitted).
`The latitude/longitude data and incidental data read out on
`the basis of the recognized voice are output from the output
`terminal 10a (step 107).
`In the case where step 104 is unable to recognize the voice
`of a registered place name, it is decided whether a registered
`specific voice other than the place name has been recognized
`or not (step 108). In the case where a registered specific
`voice other than the place name has been recognized, a
`character code corresponding to the recognized voice is
`determined (step 109), and the character code thus identified
`is output from the output terminal 10b (step 110).
`In the case where even a specific registered voice other
`than a place name could not be recognized in step 108, the
`process is terminated. In such a case, an alternative is to
`notify the navigation apparatus 20 that the voice could not
`be recognized and to issue a warning by means of the voice
`synthesized on the voice synthesis circuit 31 or the charac-
`ters displayed on the display unit 40.
`Now, the operation of the navigation apparatus 20 will be
`explained with reference to the flowchart of FIG. 7. First, the
`arithmetic circuit 23 decides whether the present position
`display mode is set or not (step 201). In the case where the
`decision is that the present position display mode is set, the
`present position detecting circuit 22 is caused to execute the
`measurement of the present position (step 202). The road
`map data of the neighborhood of the present position thus
`measured is read from the CD-ROM (step 203). On the basis
`of the road map data thus read out, the process is performed
`for displaying the road map of the corresponding coordinate
`point on the display unit 40 (step 204).
`In the case where the decision in step 201 is that the
`present position display mode is not set, or in the case where
`the process for display of the road map of the present
`position has been completed in step 204 and the road map is
`on display, then it is decided whether the latitude/longitude
`data, etc. are supplied from the voice recognition unit 10
`through the input terminals 20a, 20b (step 205). In the case
`where the decision is that the latitude/longitude data and
`incidental character data or the like are supplied, the pro-
`
`8
`nunciation character code supplied through the terminal 20b
`is supplied to the voice synthesis circuit 31, so that the voice
`recognized by the voice recognition unit 10 is synthesized
`and output from the speaker 32 (step 206). The road map
`data of the neighborhood of the position indicated by the
`latitude/longitude data is then read out of the CD-ROM (step
`207). On the basis of the road map data thus read out, the
`road map display process is performed thereby to display the
`road map of the corresponding coordinate point on the
`display unit 40 (step 208).
`In the case where the decision in step 205 is that the
`latitude/longitude data are not supplied from the voice
`recognition unit 10, or in the case where the process for
`displaying the road map of a designated place name is
`complete in step 208, and the road map is on display, on the
`other hand, it is decided whether the character code directly
`specifying a display position is supplied from the voice
`recognition unit 10 through the input terminal 20b (step
`209). In the case where the decision is that the character code
`is supplied from the terminal 20b, the particular character
`code is supplied to the voice synthesis circuit 31, and the
`voice recognized by the voice recognition unit 10 is output
`from the speaker 32 (step 210). In the case where step 209
`identifies a character code directly specifying the display
`position (i.e., such words as “destination”, “starting point”,
`“intermediate place”, “residence” or the like), it is decided
`whether the coordinate point specified by these characters is
`registered in the RAM 25 or not (step 211). In the case where
`such coordinate point is registered, the road map data of the
`neighborhood of the position indicated by the latitude/
`longitude data representing the registered coordinate point is
`read from the CD-ROM (step 212). The process for display-
`ing the road map is performed on the road map data thus read
`out, and the road map of the corresponding coordinate point
`is displayed on the display unit 40 (step 213). The process
`returns to step 201 while the same road map is on display.
`In the case where step 209 decides that the character code
`directly specifying the display position is not supplied from
`the voice recognition unit 10,
`the arithmetic circuit 23
`decides whether the operation is performed for specifying
`the display position by the operating key 27 (step 214). In
`the case where the operation is performed for specifying the
`display position, it is decided whether the vehicle is moving
`or not on the basis of the data detected by the vehicle speed
`sensor 26 (step 215). In the case where the arithmetic circuit
`23 decides that the vehicle is moving, the operation per-
`formed at that time is invalidated and the process returns to
`step 201 (in which case some alarm may be issued).
`In the case where the decision is that the vehicle is not
`
`moving, on the other hand, the process proceeds to step 211
`for deciding whether a coordinate point is registered or not.
`In the case where a coordinate point
`is registered,
`the
`process is performed for displaying the road map of the
`coordinate point (steps 212, 213), after which the process
`returns to step 201.
`In the case where step 211 decides that the coordinate
`point of the corresponding position such as “destination”,
`“starting point”, “intermediate place” or “residence” is not
`registered, an alarm is issued against the lack of registration
`by voice synthesis through the voice synthesis circuit 31 or
`by character display on the display unit 40 (step 216), and
`the process returns to the decision in step 201.
`The foregoing explanation with reference to the flowchart
`of FIG. 7 refers to the process relating to map display. In the
`case where a character code is supplied from the voice
`recognition unit 10 based on the result of recognizing a voice
`
`10
`
`15
`
`20
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`14
`
`14
`
`

`
`6,064,323
`
`9
`specifying an operation other than map display, however, the
`corresponding process is performed under the control of the
`arithmetic circuit 23. When the character code is supplied
`upon recognition of a voice meaning “What time is it now?”,
`for example, a voice announcing the present time is synthe-
`sized by the voice synthesis circuit 31 and output from the
`speaker 32 under the control of the arithmetic circuit 23.
`Other commands are also processed similarly as a verbal
`reply synthesized by the voice synthesis circuit 31 and
`output from the speaker 32, or by a correspon

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket