`U.S. Patent No. 6,067,502 to Hayashida (“Hayashida”)
`
`I have provided below a claim chart comparing the disclosures of Hayashida to the ’317 Asserted Claims, as well the combination of
`Hayashida and JPH10-197277 to Maruyama et al. (“Maruyama”).
`
`Hayashida was filed August 21, 1997 and issued May 23, 2000. Hayashida therefore qualifies as prior art with regard to the ’317 patent
`under 35 U.S.C. § 102(e) (pre-AIA).
`
`Maruyama was published on July 31, 1998 and therefore qualifies as prior art with regard to the ’317 Patent under at least 35 U.S.C. §
`102(a).
`
`
`Hayashida
`
`U.S. Patent No. 6,748,317
`Claim 1
`1[P]. A portable
`comprising:
`
`terminal,
`
`To the extent the preamble is limiting, Hayashida discloses a portable terminal, as it describes a
`“carrying-type navigation device.”
`
`
`This invention is related to a map display device to guiding and searching a
`movement route of a vehicle based on a map information, especially this
`invention is relate with the improvement of the display of the map information.
`Hayashida at 1:5-8.
`
`
`Then this navigation processing can be also executed by this computer device,
`if the device which can detect the present position by GPS reception device 25
`and this information memory part 37 are connected with the carrying-type
`computer device. Moreover this invention can be applied as the vehicle except
`the car and the navigation device of the shipping, the aircraft and the map
`which is used for the navigation may be a chart and a submarine map and so
`on in addition to the road map. Moreover again this invention may be applied
`to the carrying-type navigation device in addition to the navigation device
`which is attached to the movement bodies such as the car. In other words, this
`invention may be applied to the small navigation device which can be
`
`OMM_US:77143848.1
`
`
`1
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 1 of 42
`
`
`
`[1(a)] a device for getting
`location information denoting
`a resent1 place of said portable
`terminal;2
`
`accompanied by the human and which is used in a cycling, a travel, a
`mountaineering, a hike, a fishing or so on.
`Hayashida at 76:5-20.
`
`Under the Court’s construction of this limitation, Hayashida discloses the function of getting
`location information denoting a present place of said portable terminal using a structure of a
`wireless or cellular antenna, a GPS, a PHS, or the like; a data receiver; and a CPU for analyzing
`received data; or equivalents thereof. For example, Hayashida discloses a device (i.e., CPU 2 and
`present position detector 20, including GPS receiver unit 25 and beacon receiver unit 26) that
`perform the claimed function of getting location information denoting a present place of said
`portable terminal.
`
`
`FIG. 1 illustrates the overall circuitry of the navigation device. A central
`processor 1 controls the operation of the whole navigation device. The central
`processor 1 is comprised with a CPU 2, a flush memory 3, a RAM 5, a ROM 4,
`a sensor input interface 7, a communication interface 8, an image (picture)
`processor 9, a image (picture) memory 10, a voice processor 11 and a clock
`(clock generator) 6. The CPU 2 and the devices through up to the clock 6 are
`connected together through a CPU local bus 15, and the data are exchanged
`among these devices.
`Hayashida at 2:46-55.
`
`
`The sensor input interface 7 comprises an A/D converter circuit or a buffer
`circuit. The sensor input interface 7 receives analog or digital sensor data from
`the sensors 21 to 24 of a present position detector 20. The present position
`
`
`
`OMM_US:77143848.1
`
`
`1 Based on Plaintiff’s Infringement Contentions and subsequent claim limitations referring to a present place, Defendant understands
`this to mean “present.”
`2 The Court construed this element as:
`“Function: getting location information denoting a present place of said portable terminal
`Structure: a wireless or cellular antenna, a GPS, a PHS, or the like; a data receiver; and a CPU for analyzing received data; or
`equivalents thereof”
`
`2
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 2 of 42
`
`
`
`detector 20 includes an absolute direction sensor 21, a relative direction sensor
`22, a distance sensor 23 and a vehicle speed sensor 24.
`Hayashida at 7:24-30.
`
`
`An I/O data bus 28 is connected to the communication interface 8 of the central
`processor 1. To the I/O data bus 28 are connected the GPS receiver unit 25,
`the beacon receiver unit 26 and the data transmitter/receiver unit 27 of the
`present position detector 20. To the I/O data bus 28 are further connected a
`touch switch 34 and a printer 35 of the input/output unit 30, and an information
`memory unit 37. That is, a variety of data are exchanged between the external
`accessory equipment and the CPU local bus 15 through the communication
`interface 8.
`
`The present position detector 20 outputs data for detecting the present position
`of the car. In other words, the absolute direction sensor 21 detects the absolute
`direction. The relative direction sensor 22 detects the relative direction with
`respect to the absolute direction. Furthermore, the distance sensor 23 detects
`the distance travelled. The vehicle speed sensor 24 detects the running speed of
`the car. The GPS receiver unit 25 receives GPS (Global Positioning System)
`signals to detect position data such as longitude and latitude of the car. The
`GPS signals are microwaves transmitted from a plurality of satellites orbiting
`round the earth.
`
`Similarly the beacon receiver unit 26 receives beacon from a data offering
`system such as VICS (Vehicle Information and Communication System) or the
`like, and the received data and the corrected data of GPS are output to the I/O
`data bus 28.
`
`The data transmitter/receiver unit 27 exchanges a variety of information related
`to the present position or the road conditions near the car relative to the bi-
`directional present position information offering system or the ATIS (advanced
`traffic information service), etc. by utilizing a cellular phone, FM multiplex
`signals or a telephone circuit. These information are used as a detecting
`
`3
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 3 of 42
`
`
`
`information of the car position or a support information of movement. The
`beacon receiver unit 26 and the data transmitter/receiver unit 27 may not be
`provided. As for this data sending and the data transmitter/receiver unit 27, a
`radio receiver, a television receiver, a carrying telephone, a pager or the other
`radio communication machine are used.
`Hayashida at 7:50-8:21.
`
`
`Id. at Fig. 1.
`
`
`FIG. 3 illustrates some of a group of data stored in the RAM 5. The present
`position data MP represent the present position of the vehicle and are detected
`by a present position detector 20. The absolute direction data ZD represents
`the south-north direction relying upon the terrestrial magnetism and are found
`based upon the data from an absolute direction sensor 21. The relative
`direction angle data Dθ represent an angle of the direction in which the vehicle
`
`
`
`OMM_US:77143848.1
`
`
`4
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 4 of 42
`
`
`
`is traveling with respect to the absolute position data ZD and are found based
`upon the data from a relative direction sensor 22.
`
`
`
`The traveled distance data ML represent a distance traveled by the vehicle and
`are found based on the data from a distance sensor 23. The present position
`data PI are related to the present position and are input from a beacon receiver
`26 or the data transmitter-receiver 27. The VICS data VD and ATIS data AD
`are input from the beacon receiver 26 or the data transmitter-receiver 27. The
`VICS data VD are used for correcting an error in the position of the vehicle
`detected by a GPS receiver 25. The ATIS data AD are used for determining
`traffic regulations and traffic jamming in the areas.
`Hayashida at 10:55-11:8.
`
`
`Then a processing for detecting the present position (step SA2) and the
`subsequent processing are executed. The processing for detecting the present
`position (step SA2) detects the geographical coordinates (latitude, longitude,
`altitude, etc.) of an overland moving body, i.e., of a vehicle mounting the
`navigation device. That is, a GPS receiver 25 receives signals from a plurality
`of satellites orbiting around the earth, detects coordinate positions of the
`satellites, times at which the electromagnetic waves are emitted from the
`satellites and the time at which the electromagnetic waves are received by the
`GPS receiver 25, and calculates the distances to the satellites. The coordinate
`position of the vehicle is calculated from the distances to the satellites, to detect
`the present position of the vehicle. The thus found geographical coordinate data
`of the vehicle are stored in the RAM 5 as present position data MP. The present
`position data MP are often corrected by the data input through a beacon
`receiver 26 or the data transmitter/receiver 27.
`Hayashida at 13:16-33; see also id. at Fig. 5.
`
`
`Under the Court’s construction of this limitation, Hayashida discloses the function of getting
`direction information denoting an orientation of said portable terminal using a structure of a
`compass, gyroscope, and/or sensor such as a clinometer in conjunction with a CPU, or equivalents
`
`5
`
`[1(b)] a device for getting a
`direction
`information
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 5 of 42
`
`
`
`denoting an orientation of said
`portable terminal;3
`
`
`thereof. Hayashida discloses a device (i.e., present position detector 20, including absolute
`direction sensor 21 (terrestrial magnetism sensor) and relative direction sensor 22 (gyroscope), and
`CPU 2) that performs the claimed function of getting direction information denoting an orientation
`of said portable terminal. A PHOSITA would have understood geomagnetic sensor as taught by
`Hayashida to refer to a compass.
`
`
`
`OMM_US:77143848.1
`
`
`3 The Court construed this element as:
`“Function: getting direction information denoting an orientation of said portable terminal
`Structure: a compass, gyroscope, and/or sensor such as a clinometer in conjunction with a CPU, or equivalents thereof.”
`6
`
`FIG. 1 illustrates the overall circuitry of the navigation device. A central
`processor 1 controls the operation of the whole navigation device. The central
`processor 1 is comprised with a CPU 2, a flush memory 3, a RAM 5, a ROM 4,
`a sensor input interface 7, a communication interface 8, an image (picture)
`processor 9, a image (picture) memory 10, a voice processor 11 and a clock
`(clock generator) 6. The CPU 2 and the devices through up to the clock 6 are
`connected together through a CPU local bus 15, and the data are exchanged
`among these devices.
`Hayashida at 2:46-55.
`
`
`The sensor input interface 7 comprises an A/D converter circuit or a buffer
`circuit. The sensor input interface 7 receives analog or digital sensor data from
`the sensors 21 to 24 of a present position detector 20. The present position
`detector 20 includes an absolute direction sensor 21, a relative direction sensor
`22, a distance sensor 23 and a vehicle speed sensor 24.
`
`
`
`The absolute direction sensor 21 is, for example, a terrestrial magnetism
`sensor and detects terrestrial magnetism. The absolute direction sensor 21
`outputs data indicating a south-and-north direction which serves as the
`absolute direction. The relative direction sensor 22 is, for example, a steering
`angle sensor and detects the steering angle of the wheel based upon a
`gyroscope such as optical fiber gyroscope or piezo-electric vibration
`gyroscope. The relative direction sensor 22 outputs a relative angle of a
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 6 of 42
`
`
`
`direction of progress of the car with respect to the absolute direction detected
`by the absolute direction sensor 21.
`Hayashida at 7:24-42.
`
`
`
`
`Id. at Fig. 1.
`
`
`The present position detector 20 outputs data for detecting the present position
`of the car. In other words, the absolute direction sensor 21 detects the absolute
`direction. The relative direction sensor 22 detects the relative direction with
`respect to the absolute direction. Furthermore, the distance sensor 23 detects
`the distance travelled. The vehicle speed sensor 24 detects the running speed of
`the car. The GPS receiver unit 25 receives GPS (Global Positioning System)
`signals to detect position data such as longitude and latitude of the car. The
`GPS signals are microwaves transmitted from a plurality of satellites orbiting
`round the earth.
`
`OMM_US:77143848.1
`
`
`7
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 7 of 42
`
`
`
`Hayashida at 7:60-8:3.
`
`
`FIG. 3 illustrates some of a group of data stored in the RAM 5. The present
`position data MP represent the present position of the vehicle and are detected
`by a present position detector 20. The absolute direction data ZD represents
`the south-north direction relying upon the terrestrial magnetism and are found
`based upon the data from an absolute direction sensor 21. The relative
`direction angle data Dθ represent an angle of the direction in which the vehicle
`is traveling with respect to the absolute position data ZD and are found based
`upon the data from a relative direction sensor 22.
`Hayashida at 10:55-64.
`
`
`In the processing for detecting the present position (step SA2), furthermore the
`absolute direction data ZD, relative direction angle data D θ and the traveled
`distance data ML are simultaneously found by using an absolute direction
`sensor 21, a relative direction sensor 22 and a distance sensor 23. The absolute
`direction data ZD, relative direction angle data Dθ and traveled distance data
`ML are operated to determine the position of the vehicle. The thus determined
`position of the vehicle is collated with map data stored in a data 38c of the
`information memory unit 37, and the present position on the map screen is
`corrected and is indicated more correctly. Therefore, the present position of the
`vehicle is correctly indicated even when the GPS signals are not received such
`as traveling through a tunnel.
`Hayashida at 13:34-47; see also id. at Fig. 5.
`
`Hayashida discloses an input device (e.g., input/output device 30) for inputting a destination.
`
`
`The input/output device 30 comprises a display 33, a transparent touch panel
`34, a printer 35 and a speaker 13. The display 33 displays guide data during
`the navigation operation. The touch panel 34 is constituted by a plurality of
`transparent touch switches that are arranged in the form of a matrix on a plane,
`and is adhered onto the screen of the display 33. By using the touch panel 34,
`
`[1(c)] an input device for
`inputting a destination; and
`
`OMM_US:77143848.1
`
`
`8
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 8 of 42
`
`
`
`data necessary for setting the destination, such as start point, destination,
`passing points, drop-in places, etc. are input to the navigation device.
`Hayashida at 8:22-31.
`
`
`Id. at Fig. 1, see also id. at Figs. 68-72.
`
`
`The inputted destination data TP are related to the coordinate positions and
`names of the destinations and are input by the user. The start point of route
`data SP are map coordinate data of a point from where the navigation
`operation starts. The end point of route data ED are map coordinate data of a
`point where the navigation operation ends.
`Hayashida at 11:9-14.
`
`
`
`
`OMM_US:77143848.1
`
`
`9
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 9 of 42
`
`
`
`Then a present position processing (step SA2), a destination-setting processing
`(step SA3), a route search processing (step SA4), a guide/display processing
`(step SA5) and a other processing (step SA6) are cyclically executed.
`Hayashida at 13:7-10.
`
`
`In setting the destination(step SA3), the geographical coordinates of the
`destination desired by the user are stored as inputted destination data TP. For
`example, a coordinate position is specified by the user on a road map or on a
`house map shown on a display 33. The destination is selected by operator from
`a list of destinations shown on the display 33. Thereafter, a central processing
`unit 1 stores the data related to the geographical coordinates of the destination
`in the RAM 5 as inputted destination data TP.
`Hayashida at 13:48-56; see also id. at Fig. 5.
`
`
`The destination setting processing may be implemented as follows. First a area
`is specified by the postcode. Next a specific street is chosen from all the streets
`in the area. Moreover when genre is specified, the facilities which correspond
`to the specified genre and is around the choice street are put on the list. Lastly
`the requested destination is chosen from the facility name that is put on this list.
`Therefore the required destination can be retrieved from the street and
`neighborhood and the specific genre.
`
`Moreover the destination setting processing may be executed as follows. A area
`is specified by the postcode. Next a specific street is chosen from the streets in
`the area. The facilities along the chosen street are put on the list. Lastly a
`required destination is chosen from the facility name that is put on this list.
`Therefore a required destination is extracted along the specific street in the
`specific area.
`
`Moreover again the destination setting processing may be executed as follows.
`First a area is specified by the postcode RDN. After this, a specific genre is
`chosen. Next the street name where each of the facilities corresponding to the
`specified identical genre exist is enumerated. Then facilities are chosen based
`
`OMM_US:77143848.1
`
`
`10
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 10 of 42
`
`
`
`[1(d)] a display, wherein
`
`
`
`on this street name list. When a destination is set, it can be retrieved in order
`of the area, the genre and the street.
`Hayashida at 67:62-68:18; see also id. at 61:62-63:30 (describing Destination-setting
`Processing); 64:66-66:39 (describing Display of Street Name and Destination Setting
`Subroutine); 66:42-67:59 (describing Genre List Display and Destination Setting
`Subroutine).
`
`Hayashida discloses a display, which could be display element 33.
`
`
`The input/output device 30 comprises a display 33, a transparent touch panel
`34, a printer 35 and a speaker 13. The display 33 displays guide data during
`the navigation operation. The touch panel 34 is constituted by a plurality of
`transparent touch switches that are arranged in the form of a matrix on a plane,
`and is adhered onto the screen of the display 33. By using the touch panel 34,
`data necessary for setting the destination, such as start point, destination,
`passing points, drop-in places, etc. are input to the navigation device.
`Hayashida at 8:22-31.
`
`The display 33 may be a CRT, a liquid crystal display or a plasma display, and
`displays picture. Desirably, however, the liquid crystal display is preferred as
`a display 33 because it consumes small amounts of electric power, it can be
`seen highly clearly, and is light in weight. In this embodiment (invention), this
`display 33 is liquid crystal having more wide screen.
`
`Furthermore this display 33 may be composed of more than 2 liquid crystal
`displays which can separate. Then each liquid crystal display is connected with
`a image processor 9 respectively with an independent video signal cable and is
`placed in a same or a different position each other.
`Hayashida at 8:37-49.
`
`
`OMM_US:77143848.1
`
`
`11
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 11 of 42
`
`
`
`
`Id. at Fig. 1; see also id. at Figs. 9, 10, 14, 16, 17, 20, 23-26, 28-33, 40-46, 48, 50-52, 69, 71, and
`73.
`
`Hayashida’s display displays positions of said destination and said present place, and a relation of
`said direction and a direction from said present place to said destination. Hayashida discloses a
`number of ways to present the user’s current direction and direction to the destination, including:
`(1) a route line and a pointer symbol indicating the user’s current direction, along with a compass
`graphic showing north (e.g., Fig. 16); (2) an arrow pointing to the destination and a pointer symbol
`indicating the user’s current direction, along with a compass graphic showing north (e.g., Fig. 17);
`and (3) a route line and a pointer symbol indicating the user’s current direction, along with text
`describing distance to the destination (e.g., Fig. 32-33). Hayashida also discloses a “head up”
`orientation for the map.
`
`
`A navigation system has a display screen which can be divided into a plurality
`of windows for simultaneously displaying different classifications of routes in
`
`12
`
`[1(e)] said display displays
`positions of said destination
`and said present place, and a
`relation of said direction and a
`direction from said present
`place to said destination, and
`
`
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 12 of 42
`
`
`
`the same scale, or different routes from a present position or starting point to a
`destination or return to a route from which a car has deviated.
`Hayashida at Abstract.
`
`
`Also when the car approaches a right or left turning point, a geographical
`information of the right or left turning point on the identified route is shown in
`the 1st screen. Moreover at this time, a detailed map is displayed in the 2nd
`screen. At this detailed map, a information quantity which is displayed
`according to a traveling condition of the car is adjusted. Therefore the user can
`know a direction which the car should progress and a geographical information
`around the car.
`
`Moreover a north up map, i.e. a map which above (upper) of the map is north
`is displayed in the 1st screen, a head up map, i.e. a map which above (upper)
`of the map is the direction of the progress is displayed in the 2nd screen.
`Hayashida at 2:1-12.
`
`
`To solve the above problem, in this invention, a screen which displays a map is
`divided into more than one, according to a traveling condition of the car, a
`simple map which shows geographical relation between a present position of
`the car and a destination is shown in a 1st screen, detailed map information is
`shown in a 2nd screen. Therefore a seeing person can know a direction of the
`destination immediately and can also know detailed geographical information
`around the present position of the car at the same time.
`Hayashida at 1:58-67.
`
`(1) When a single screen (a 1st screen) is divided (step SC4), a map of head up
`or a map at north up is shown in one of the screens (a 2nd screen) which were
`divided (Steps SC20, SC22). A present position of a car and distance to a
`direction, a destination of a destination etc. are shown in the other which
`divided screen (a 3rd screen) (step SC24). This is a simple map where necessary
`and minimum information is displayed. After this, an other guidance/display
`processing is executed (step SC18).
`
`OMM_US:77143848.1
`
`
`13
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 13 of 42
`
`
`
`Hayashida at 5:66-6:7; see also id. at 3:9-16; 6:8-16.
`
`The road map data and the house map data are changed over depending upon
`the conditions such as a distance from the present position to a guide point
`(destination, place to be dropped in or intersection), a traveling speed of the
`vehicle, the size of the area that can be displayed, or the switching operation
`by the user. Near the guide point (destination, place to be dropped in or
`intersection), the map is shown on an enlarged scale on the display 33. Instead
`of displaying the road map, a minimum amount of necessary data only may be
`displayed, such as the guide route, direction of the destination or the place to
`be dropped in and the present position, but omitting geographical data.
`Hayashida at 14:15-26.
`
`
`This guide/display processing is the processing to report the information to
`make travel the car along the guide route. In other words, a guide route is
`searched and this searched guide route is shown in display 33. This guide route
`is displayed for the car to become always a screen center and an information
`about points which should turn right or left in the guide route is reported at any
`time. In this guide/display processing, various information is reported and is
`shown at any time for the car to travel a guide route smoothly.
`Hayashida at 14:56-65.
`
`
`FIG. 16 shows the 2nd screen 108 where the guide route from crossing point
`CSP to destination 152 is displayed by the display processing of this whole
`route (fore). In this way, in the display processing of the whole route (fore), a
`whole guide route in front of the crossing point CSP which is a end of the guide
`route of the 3rd screen is shown in the 2nd screen.
`Hayashida at 21:57-63; see also id. at 21:7-56.
`
`
`OMM_US:77143848.1
`
`
`14
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 14 of 42
`
`
`
`
`
`Hayashida at Fig. 16; see also id. at Figs. 6-15, 14.
`
`
`FIG. 17 shows the simple map which is displayed in the screen of display 33.
`A symbol 120 which designates the absolute direction (a north of a terrestrial
`magnetism), a symbol 126 which designates the present position of the car,
`an arrow 122 which designates a direction to the destination, and characters
`124 which designates a distance to the destination are shown in screen 104.
`Such a simple map may be shown in either of the 1st, the 2nd and/or the 3rd
`screen.
`
`Furthermore, this simple map isn't limited to the one of FIG. 17, e.g. a traveling
`direction of the car may be always point up in the screen 104. In this case, a
`symbol 120 which designates the absolute direction is turned on screen 104
`according to relative azimuth data D θ.
`Hayashida at 22:32-44; see also id. at 21:64-22:31.
`15
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 15 of 42
`
`
`
`
`
`
`
`OMM_US:77143848.1
`
`
`16
`
`Hayashida at Fig. 17.
`
`
`
`
`FIG. 20 shows a divided screen of this 2nd embodiment. A head up map is
`shown in the 1st screen before the division as shown in FIG. 9. When
`instructing to divide the screen, the north up map is displayed in the 2nd screen
`108, in this 2nd screen 108, a symbol 120 which shows an absolute azimuth
`points up and a symbol 100 which shows the direction of the traveling by the
`car points to the left.
`
`Also the head up map is displayed in the 3rd screen 110 and a symbol 100 which
`designates the direction of the traveling by the car points up in the 3rd screen
`and a symbol 122 which designates a northern direction points to the right in
`the 3rd screen.
`Hayashida at 24:9-20.
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 16 of 42
`
`
`
`Hayashida at Fig. 20; see also id. at Figs. 19 and 21.
`
`
`
`
`FIG. 27 shows a flow chart of the guide/display processing of the 5th
`embodiment. When the car deviate from the guide route during the guide, at the
`guide/display processing of this 5th embodiment, a return route which returns
`to the guide route and another route which is searched newly or the previous
`route are shown respectively at the divided screen.
`Hayashida at 30:50-55.
`
`
`FIG. 32 shows the other example of FIG. 24 and FIG. 25 of the 4th embodiment
`and FIG. 30 of the 5th embodiment. In this example, a distance and time
`information 186 are shown with the guide route. This distance and time
`information indicate a distance and required time to the end point of guide
`route. For example this distance and time information are the distance
`
`OMM_US:77143848.1
`
`
`17
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 17 of 42
`
`
`
`according to the route and the required time to travel this distance by the
`prescribed speed, when traveling guide route 182 to the end point of guide
`route in FIG. 32.
`
`Travelling distance to the end point of the guide route 184 and traveling time
`is also shown in the 3rd screen 110 of FIG. 32 as the distance and time
`information 186. FIG. 33 shows the guide route which is changed from FIG. 32
`when the step SL26 etc. of FIG. 22 is executed. The distance and time
`information 186 of this FIG. 33 indicate the travelling distance and the required
`time to the end point of the guide route 188. Furthermore the 3rd screen 110 of
`FIG. 33 is copied from the 2nd screen of FIG. 32.
`
`Further in the FIGS. 32 and 33 the distance and time information 186 may
`include the VICS information (construction information or information at the
`congestion road in the route etc.) which is received by data sending and the
`data transmitter/receiver unit 27 and may include the number of times of the
`turning right or left in the this route and so on. This VICS information can
`include congestion information at a parking lot around the route and so on.
`Hayashida at 34:49-35:8.
`
`
`
`OMM_US:77143848.1
`
`
`18
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 18 of 42
`
`
`
`Hayashida at Figs. 32 and 33.
`
`
`
`
`FIG. 34 shows the flow chart of the guide/display processing of the 6th
`embodiment. At the 6th embodiment, the screen is divided when the car deviates
`from the first guide route. The road map which includes the part of the deviating
`guide route and the present position of the car is shown at the divided 2nd
`screen. The previous guide-route which is displayed before the division or the
`new re-searched whole guide route to the destination is shown at the divided
`3rd screen. The previous guide route from which the car deviates is searched
`by the route search processing (step SA4) of FIG. 5 or the whole route
`researching (step SR20) which is described later.
`19
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 19 of 42
`
`
`
`Hayashida at 35:36-47.
`
` FIG. 46 shows the application example of the display of the 2nd screen. For
`example, the return route is shown by arrow 268. Further when the return route
`isn't displayed in the 2nd screen, only the confluence which the return route and
`the previous route cross may be shown by mark 262. Further the confluence
`may be also shown by mark 262 when the return route is displayed. Moreover,
`the direction of the destination in the previous guide route may be shown by a
`direction mark 264 of the destination.
`Hayashida at 49:3-11.
`
`
`Id. at Fig. 46; see also id. at Figs. 38 and 39.
`
`
`20
`
`
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 20 of 42
`
`
`
`Next the processing to display the direction mark 264 of the destination which
`shows the direction of the destination in the computed intersection point CP is
`executed. For example, this processing is executed as follows. The character
`pattern data of the arrow which is used for the direction mark of the destination
`is stored in the information memory unit 37 or the ROM 4. Then the character
`pattern of the arrow is turned to show the above direction of the destination
`according to the above intersection point data CP. This character pattern data
`is sent to image processor 9 and is changed into the image data. Therefore the
`direction mark of the destination 264 is shown in the coordinate which is
`designated by the intersection point data CP. Actually as the tip of the arrow
`agree with the coordinate of the intersection point data CP, A deviation to the
`intersection point data CP is provided for the coordinate of the direction mark
`of the destination.
`Hayashida at 49:44-60.
`
`(8) In addition to the above matter, the map which is the displayed in the 3rd
`screen may be the map information which shows around the car, the start point,
`the destination, the dropping in place or the chosen optional part. Then the map
`which connects with one azimuth of all the azimuths of the map which is
`displayed in this 3rd screen may be displayed in the 2nd screen. In other words,
`a map around the destination (the dropping in place, the specification place or
`so on) is displayed in the 3rd screen. Moreover the map which is connected
`with the map of this 3rd screen and contains the guide route which connects
`from the start point to the destination may be displayed in the 2nd screen.
`Hayashida at 74:41-52.
`
`Moreover the map which connects with the 3rd screen and is displayed in the
`2nd screen may be the map which connects on the running prolongation of the
`car. In other words, a map around the destination is displayed in the 3