throbber
Invalidity of U.S. Patent No. 6,748,317 by
`U.S. Patent No. 6,067,502 to Hayashida (“Hayashida”)
`
`I have provided below a claim chart comparing the disclosures of Hayashida to the ’317 Asserted Claims, as well the combination of
`Hayashida and JPH10-197277 to Maruyama et al. (“Maruyama”).
`
`Hayashida was filed August 21, 1997 and issued May 23, 2000. Hayashida therefore qualifies as prior art with regard to the ’317 patent
`under 35 U.S.C. § 102(e) (pre-AIA).
`
`Maruyama was published on July 31, 1998 and therefore qualifies as prior art with regard to the ’317 Patent under at least 35 U.S.C. §
`102(a).
`
`
`Hayashida
`
`U.S. Patent No. 6,748,317
`Claim 1
`1[P]. A portable
`comprising:
`
`terminal,
`
`To the extent the preamble is limiting, Hayashida discloses a portable terminal, as it describes a
`“carrying-type navigation device.”
`
`
`This invention is related to a map display device to guiding and searching a
`movement route of a vehicle based on a map information, especially this
`invention is relate with the improvement of the display of the map information.
`Hayashida at 1:5-8.
`
`
`Then this navigation processing can be also executed by this computer device,
`if the device which can detect the present position by GPS reception device 25
`and this information memory part 37 are connected with the carrying-type
`computer device. Moreover this invention can be applied as the vehicle except
`the car and the navigation device of the shipping, the aircraft and the map
`which is used for the navigation may be a chart and a submarine map and so
`on in addition to the road map. Moreover again this invention may be applied
`to the carrying-type navigation device in addition to the navigation device
`which is attached to the movement bodies such as the car. In other words, this
`invention may be applied to the small navigation device which can be
`
`OMM_US:77143848.1
`
`
`1
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 1 of 42
`
`

`

`[1(a)] a device for getting
`location information denoting
`a resent1 place of said portable
`terminal;2
`
`accompanied by the human and which is used in a cycling, a travel, a
`mountaineering, a hike, a fishing or so on.
`Hayashida at 76:5-20.
`
`Under the Court’s construction of this limitation, Hayashida discloses the function of getting
`location information denoting a present place of said portable terminal using a structure of a
`wireless or cellular antenna, a GPS, a PHS, or the like; a data receiver; and a CPU for analyzing
`received data; or equivalents thereof. For example, Hayashida discloses a device (i.e., CPU 2 and
`present position detector 20, including GPS receiver unit 25 and beacon receiver unit 26) that
`perform the claimed function of getting location information denoting a present place of said
`portable terminal.
`
`
`FIG. 1 illustrates the overall circuitry of the navigation device. A central
`processor 1 controls the operation of the whole navigation device. The central
`processor 1 is comprised with a CPU 2, a flush memory 3, a RAM 5, a ROM 4,
`a sensor input interface 7, a communication interface 8, an image (picture)
`processor 9, a image (picture) memory 10, a voice processor 11 and a clock
`(clock generator) 6. The CPU 2 and the devices through up to the clock 6 are
`connected together through a CPU local bus 15, and the data are exchanged
`among these devices.
`Hayashida at 2:46-55.
`
`
`The sensor input interface 7 comprises an A/D converter circuit or a buffer
`circuit. The sensor input interface 7 receives analog or digital sensor data from
`the sensors 21 to 24 of a present position detector 20. The present position
`
`
`
`OMM_US:77143848.1
`
`
`1 Based on Plaintiff’s Infringement Contentions and subsequent claim limitations referring to a present place, Defendant understands
`this to mean “present.”
`2 The Court construed this element as:
`“Function: getting location information denoting a present place of said portable terminal
`Structure: a wireless or cellular antenna, a GPS, a PHS, or the like; a data receiver; and a CPU for analyzing received data; or
`equivalents thereof”
`
`2
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 2 of 42
`
`

`

`detector 20 includes an absolute direction sensor 21, a relative direction sensor
`22, a distance sensor 23 and a vehicle speed sensor 24.
`Hayashida at 7:24-30.
`
`
`An I/O data bus 28 is connected to the communication interface 8 of the central
`processor 1. To the I/O data bus 28 are connected the GPS receiver unit 25,
`the beacon receiver unit 26 and the data transmitter/receiver unit 27 of the
`present position detector 20. To the I/O data bus 28 are further connected a
`touch switch 34 and a printer 35 of the input/output unit 30, and an information
`memory unit 37. That is, a variety of data are exchanged between the external
`accessory equipment and the CPU local bus 15 through the communication
`interface 8.
`
`The present position detector 20 outputs data for detecting the present position
`of the car. In other words, the absolute direction sensor 21 detects the absolute
`direction. The relative direction sensor 22 detects the relative direction with
`respect to the absolute direction. Furthermore, the distance sensor 23 detects
`the distance travelled. The vehicle speed sensor 24 detects the running speed of
`the car. The GPS receiver unit 25 receives GPS (Global Positioning System)
`signals to detect position data such as longitude and latitude of the car. The
`GPS signals are microwaves transmitted from a plurality of satellites orbiting
`round the earth.
`
`Similarly the beacon receiver unit 26 receives beacon from a data offering
`system such as VICS (Vehicle Information and Communication System) or the
`like, and the received data and the corrected data of GPS are output to the I/O
`data bus 28.
`
`The data transmitter/receiver unit 27 exchanges a variety of information related
`to the present position or the road conditions near the car relative to the bi-
`directional present position information offering system or the ATIS (advanced
`traffic information service), etc. by utilizing a cellular phone, FM multiplex
`signals or a telephone circuit. These information are used as a detecting
`
`3
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 3 of 42
`
`

`

`information of the car position or a support information of movement. The
`beacon receiver unit 26 and the data transmitter/receiver unit 27 may not be
`provided. As for this data sending and the data transmitter/receiver unit 27, a
`radio receiver, a television receiver, a carrying telephone, a pager or the other
`radio communication machine are used.
`Hayashida at 7:50-8:21.
`
`
`Id. at Fig. 1.
`
`
`FIG. 3 illustrates some of a group of data stored in the RAM 5. The present
`position data MP represent the present position of the vehicle and are detected
`by a present position detector 20. The absolute direction data ZD represents
`the south-north direction relying upon the terrestrial magnetism and are found
`based upon the data from an absolute direction sensor 21. The relative
`direction angle data Dθ represent an angle of the direction in which the vehicle
`
`
`
`OMM_US:77143848.1
`
`
`4
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 4 of 42
`
`

`

`is traveling with respect to the absolute position data ZD and are found based
`upon the data from a relative direction sensor 22.
`
`
`
`The traveled distance data ML represent a distance traveled by the vehicle and
`are found based on the data from a distance sensor 23. The present position
`data PI are related to the present position and are input from a beacon receiver
`26 or the data transmitter-receiver 27. The VICS data VD and ATIS data AD
`are input from the beacon receiver 26 or the data transmitter-receiver 27. The
`VICS data VD are used for correcting an error in the position of the vehicle
`detected by a GPS receiver 25. The ATIS data AD are used for determining
`traffic regulations and traffic jamming in the areas.
`Hayashida at 10:55-11:8.
`
`
`Then a processing for detecting the present position (step SA2) and the
`subsequent processing are executed. The processing for detecting the present
`position (step SA2) detects the geographical coordinates (latitude, longitude,
`altitude, etc.) of an overland moving body, i.e., of a vehicle mounting the
`navigation device. That is, a GPS receiver 25 receives signals from a plurality
`of satellites orbiting around the earth, detects coordinate positions of the
`satellites, times at which the electromagnetic waves are emitted from the
`satellites and the time at which the electromagnetic waves are received by the
`GPS receiver 25, and calculates the distances to the satellites. The coordinate
`position of the vehicle is calculated from the distances to the satellites, to detect
`the present position of the vehicle. The thus found geographical coordinate data
`of the vehicle are stored in the RAM 5 as present position data MP. The present
`position data MP are often corrected by the data input through a beacon
`receiver 26 or the data transmitter/receiver 27.
`Hayashida at 13:16-33; see also id. at Fig. 5.
`
`
`Under the Court’s construction of this limitation, Hayashida discloses the function of getting
`direction information denoting an orientation of said portable terminal using a structure of a
`compass, gyroscope, and/or sensor such as a clinometer in conjunction with a CPU, or equivalents
`
`5
`
`[1(b)] a device for getting a
`direction
`information
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 5 of 42
`
`

`

`denoting an orientation of said
`portable terminal;3
`
`
`thereof. Hayashida discloses a device (i.e., present position detector 20, including absolute
`direction sensor 21 (terrestrial magnetism sensor) and relative direction sensor 22 (gyroscope), and
`CPU 2) that performs the claimed function of getting direction information denoting an orientation
`of said portable terminal. A PHOSITA would have understood geomagnetic sensor as taught by
`Hayashida to refer to a compass.
`
`
`
`OMM_US:77143848.1
`
`
`3 The Court construed this element as:
`“Function: getting direction information denoting an orientation of said portable terminal
`Structure: a compass, gyroscope, and/or sensor such as a clinometer in conjunction with a CPU, or equivalents thereof.”
`6
`
`FIG. 1 illustrates the overall circuitry of the navigation device. A central
`processor 1 controls the operation of the whole navigation device. The central
`processor 1 is comprised with a CPU 2, a flush memory 3, a RAM 5, a ROM 4,
`a sensor input interface 7, a communication interface 8, an image (picture)
`processor 9, a image (picture) memory 10, a voice processor 11 and a clock
`(clock generator) 6. The CPU 2 and the devices through up to the clock 6 are
`connected together through a CPU local bus 15, and the data are exchanged
`among these devices.
`Hayashida at 2:46-55.
`
`
`The sensor input interface 7 comprises an A/D converter circuit or a buffer
`circuit. The sensor input interface 7 receives analog or digital sensor data from
`the sensors 21 to 24 of a present position detector 20. The present position
`detector 20 includes an absolute direction sensor 21, a relative direction sensor
`22, a distance sensor 23 and a vehicle speed sensor 24.
`
`
`
`The absolute direction sensor 21 is, for example, a terrestrial magnetism
`sensor and detects terrestrial magnetism. The absolute direction sensor 21
`outputs data indicating a south-and-north direction which serves as the
`absolute direction. The relative direction sensor 22 is, for example, a steering
`angle sensor and detects the steering angle of the wheel based upon a
`gyroscope such as optical fiber gyroscope or piezo-electric vibration
`gyroscope. The relative direction sensor 22 outputs a relative angle of a
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 6 of 42
`
`

`

`direction of progress of the car with respect to the absolute direction detected
`by the absolute direction sensor 21.
`Hayashida at 7:24-42.
`
`
`
`
`Id. at Fig. 1.
`
`
`The present position detector 20 outputs data for detecting the present position
`of the car. In other words, the absolute direction sensor 21 detects the absolute
`direction. The relative direction sensor 22 detects the relative direction with
`respect to the absolute direction. Furthermore, the distance sensor 23 detects
`the distance travelled. The vehicle speed sensor 24 detects the running speed of
`the car. The GPS receiver unit 25 receives GPS (Global Positioning System)
`signals to detect position data such as longitude and latitude of the car. The
`GPS signals are microwaves transmitted from a plurality of satellites orbiting
`round the earth.
`
`OMM_US:77143848.1
`
`
`7
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 7 of 42
`
`

`

`Hayashida at 7:60-8:3.
`
`
`FIG. 3 illustrates some of a group of data stored in the RAM 5. The present
`position data MP represent the present position of the vehicle and are detected
`by a present position detector 20. The absolute direction data ZD represents
`the south-north direction relying upon the terrestrial magnetism and are found
`based upon the data from an absolute direction sensor 21. The relative
`direction angle data Dθ represent an angle of the direction in which the vehicle
`is traveling with respect to the absolute position data ZD and are found based
`upon the data from a relative direction sensor 22.
`Hayashida at 10:55-64.
`
`
`In the processing for detecting the present position (step SA2), furthermore the
`absolute direction data ZD, relative direction angle data D θ and the traveled
`distance data ML are simultaneously found by using an absolute direction
`sensor 21, a relative direction sensor 22 and a distance sensor 23. The absolute
`direction data ZD, relative direction angle data Dθ and traveled distance data
`ML are operated to determine the position of the vehicle. The thus determined
`position of the vehicle is collated with map data stored in a data 38c of the
`information memory unit 37, and the present position on the map screen is
`corrected and is indicated more correctly. Therefore, the present position of the
`vehicle is correctly indicated even when the GPS signals are not received such
`as traveling through a tunnel.
`Hayashida at 13:34-47; see also id. at Fig. 5.
`
`Hayashida discloses an input device (e.g., input/output device 30) for inputting a destination.
`
`
`The input/output device 30 comprises a display 33, a transparent touch panel
`34, a printer 35 and a speaker 13. The display 33 displays guide data during
`the navigation operation. The touch panel 34 is constituted by a plurality of
`transparent touch switches that are arranged in the form of a matrix on a plane,
`and is adhered onto the screen of the display 33. By using the touch panel 34,
`
`[1(c)] an input device for
`inputting a destination; and
`
`OMM_US:77143848.1
`
`
`8
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 8 of 42
`
`

`

`data necessary for setting the destination, such as start point, destination,
`passing points, drop-in places, etc. are input to the navigation device.
`Hayashida at 8:22-31.
`
`
`Id. at Fig. 1, see also id. at Figs. 68-72.
`
`
`The inputted destination data TP are related to the coordinate positions and
`names of the destinations and are input by the user. The start point of route
`data SP are map coordinate data of a point from where the navigation
`operation starts. The end point of route data ED are map coordinate data of a
`point where the navigation operation ends.
`Hayashida at 11:9-14.
`
`
`
`
`OMM_US:77143848.1
`
`
`9
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 9 of 42
`
`

`

`Then a present position processing (step SA2), a destination-setting processing
`(step SA3), a route search processing (step SA4), a guide/display processing
`(step SA5) and a other processing (step SA6) are cyclically executed.
`Hayashida at 13:7-10.
`
`
`In setting the destination(step SA3), the geographical coordinates of the
`destination desired by the user are stored as inputted destination data TP. For
`example, a coordinate position is specified by the user on a road map or on a
`house map shown on a display 33. The destination is selected by operator from
`a list of destinations shown on the display 33. Thereafter, a central processing
`unit 1 stores the data related to the geographical coordinates of the destination
`in the RAM 5 as inputted destination data TP.
`Hayashida at 13:48-56; see also id. at Fig. 5.
`
`
`The destination setting processing may be implemented as follows. First a area
`is specified by the postcode. Next a specific street is chosen from all the streets
`in the area. Moreover when genre is specified, the facilities which correspond
`to the specified genre and is around the choice street are put on the list. Lastly
`the requested destination is chosen from the facility name that is put on this list.
`Therefore the required destination can be retrieved from the street and
`neighborhood and the specific genre.
`
`Moreover the destination setting processing may be executed as follows. A area
`is specified by the postcode. Next a specific street is chosen from the streets in
`the area. The facilities along the chosen street are put on the list. Lastly a
`required destination is chosen from the facility name that is put on this list.
`Therefore a required destination is extracted along the specific street in the
`specific area.
`
`Moreover again the destination setting processing may be executed as follows.
`First a area is specified by the postcode RDN. After this, a specific genre is
`chosen. Next the street name where each of the facilities corresponding to the
`specified identical genre exist is enumerated. Then facilities are chosen based
`
`OMM_US:77143848.1
`
`
`10
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 10 of 42
`
`

`

`[1(d)] a display, wherein
`
`
`
`on this street name list. When a destination is set, it can be retrieved in order
`of the area, the genre and the street.
`Hayashida at 67:62-68:18; see also id. at 61:62-63:30 (describing Destination-setting
`Processing); 64:66-66:39 (describing Display of Street Name and Destination Setting
`Subroutine); 66:42-67:59 (describing Genre List Display and Destination Setting
`Subroutine).
`
`Hayashida discloses a display, which could be display element 33.
`
`
`The input/output device 30 comprises a display 33, a transparent touch panel
`34, a printer 35 and a speaker 13. The display 33 displays guide data during
`the navigation operation. The touch panel 34 is constituted by a plurality of
`transparent touch switches that are arranged in the form of a matrix on a plane,
`and is adhered onto the screen of the display 33. By using the touch panel 34,
`data necessary for setting the destination, such as start point, destination,
`passing points, drop-in places, etc. are input to the navigation device.
`Hayashida at 8:22-31.
`
`The display 33 may be a CRT, a liquid crystal display or a plasma display, and
`displays picture. Desirably, however, the liquid crystal display is preferred as
`a display 33 because it consumes small amounts of electric power, it can be
`seen highly clearly, and is light in weight. In this embodiment (invention), this
`display 33 is liquid crystal having more wide screen.
`
`Furthermore this display 33 may be composed of more than 2 liquid crystal
`displays which can separate. Then each liquid crystal display is connected with
`a image processor 9 respectively with an independent video signal cable and is
`placed in a same or a different position each other.
`Hayashida at 8:37-49.
`
`
`OMM_US:77143848.1
`
`
`11
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 11 of 42
`
`

`

`
`Id. at Fig. 1; see also id. at Figs. 9, 10, 14, 16, 17, 20, 23-26, 28-33, 40-46, 48, 50-52, 69, 71, and
`73.
`
`Hayashida’s display displays positions of said destination and said present place, and a relation of
`said direction and a direction from said present place to said destination. Hayashida discloses a
`number of ways to present the user’s current direction and direction to the destination, including:
`(1) a route line and a pointer symbol indicating the user’s current direction, along with a compass
`graphic showing north (e.g., Fig. 16); (2) an arrow pointing to the destination and a pointer symbol
`indicating the user’s current direction, along with a compass graphic showing north (e.g., Fig. 17);
`and (3) a route line and a pointer symbol indicating the user’s current direction, along with text
`describing distance to the destination (e.g., Fig. 32-33). Hayashida also discloses a “head up”
`orientation for the map.
`
`
`A navigation system has a display screen which can be divided into a plurality
`of windows for simultaneously displaying different classifications of routes in
`
`12
`
`[1(e)] said display displays
`positions of said destination
`and said present place, and a
`relation of said direction and a
`direction from said present
`place to said destination, and
`
`
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 12 of 42
`
`

`

`the same scale, or different routes from a present position or starting point to a
`destination or return to a route from which a car has deviated.
`Hayashida at Abstract.
`
`
`Also when the car approaches a right or left turning point, a geographical
`information of the right or left turning point on the identified route is shown in
`the 1st screen. Moreover at this time, a detailed map is displayed in the 2nd
`screen. At this detailed map, a information quantity which is displayed
`according to a traveling condition of the car is adjusted. Therefore the user can
`know a direction which the car should progress and a geographical information
`around the car.
`
`Moreover a north up map, i.e. a map which above (upper) of the map is north
`is displayed in the 1st screen, a head up map, i.e. a map which above (upper)
`of the map is the direction of the progress is displayed in the 2nd screen.
`Hayashida at 2:1-12.
`
`
`To solve the above problem, in this invention, a screen which displays a map is
`divided into more than one, according to a traveling condition of the car, a
`simple map which shows geographical relation between a present position of
`the car and a destination is shown in a 1st screen, detailed map information is
`shown in a 2nd screen. Therefore a seeing person can know a direction of the
`destination immediately and can also know detailed geographical information
`around the present position of the car at the same time.
`Hayashida at 1:58-67.
`
`(1) When a single screen (a 1st screen) is divided (step SC4), a map of head up
`or a map at north up is shown in one of the screens (a 2nd screen) which were
`divided (Steps SC20, SC22). A present position of a car and distance to a
`direction, a destination of a destination etc. are shown in the other which
`divided screen (a 3rd screen) (step SC24). This is a simple map where necessary
`and minimum information is displayed. After this, an other guidance/display
`processing is executed (step SC18).
`
`OMM_US:77143848.1
`
`
`13
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 13 of 42
`
`

`

`Hayashida at 5:66-6:7; see also id. at 3:9-16; 6:8-16.
`
`The road map data and the house map data are changed over depending upon
`the conditions such as a distance from the present position to a guide point
`(destination, place to be dropped in or intersection), a traveling speed of the
`vehicle, the size of the area that can be displayed, or the switching operation
`by the user. Near the guide point (destination, place to be dropped in or
`intersection), the map is shown on an enlarged scale on the display 33. Instead
`of displaying the road map, a minimum amount of necessary data only may be
`displayed, such as the guide route, direction of the destination or the place to
`be dropped in and the present position, but omitting geographical data.
`Hayashida at 14:15-26.
`
`
`This guide/display processing is the processing to report the information to
`make travel the car along the guide route. In other words, a guide route is
`searched and this searched guide route is shown in display 33. This guide route
`is displayed for the car to become always a screen center and an information
`about points which should turn right or left in the guide route is reported at any
`time. In this guide/display processing, various information is reported and is
`shown at any time for the car to travel a guide route smoothly.
`Hayashida at 14:56-65.
`
`
`FIG. 16 shows the 2nd screen 108 where the guide route from crossing point
`CSP to destination 152 is displayed by the display processing of this whole
`route (fore). In this way, in the display processing of the whole route (fore), a
`whole guide route in front of the crossing point CSP which is a end of the guide
`route of the 3rd screen is shown in the 2nd screen.
`Hayashida at 21:57-63; see also id. at 21:7-56.
`
`
`OMM_US:77143848.1
`
`
`14
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 14 of 42
`
`

`

`
`
`Hayashida at Fig. 16; see also id. at Figs. 6-15, 14.
`
`
`FIG. 17 shows the simple map which is displayed in the screen of display 33.
`A symbol 120 which designates the absolute direction (a north of a terrestrial
`magnetism), a symbol 126 which designates the present position of the car,
`an arrow 122 which designates a direction to the destination, and characters
`124 which designates a distance to the destination are shown in screen 104.
`Such a simple map may be shown in either of the 1st, the 2nd and/or the 3rd
`screen.
`
`Furthermore, this simple map isn't limited to the one of FIG. 17, e.g. a traveling
`direction of the car may be always point up in the screen 104. In this case, a
`symbol 120 which designates the absolute direction is turned on screen 104
`according to relative azimuth data D θ.
`Hayashida at 22:32-44; see also id. at 21:64-22:31.
`15
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 15 of 42
`
`

`

`
`
`
`
`OMM_US:77143848.1
`
`
`16
`
`Hayashida at Fig. 17.
`
`
`
`
`FIG. 20 shows a divided screen of this 2nd embodiment. A head up map is
`shown in the 1st screen before the division as shown in FIG. 9. When
`instructing to divide the screen, the north up map is displayed in the 2nd screen
`108, in this 2nd screen 108, a symbol 120 which shows an absolute azimuth
`points up and a symbol 100 which shows the direction of the traveling by the
`car points to the left.
`
`Also the head up map is displayed in the 3rd screen 110 and a symbol 100 which
`designates the direction of the traveling by the car points up in the 3rd screen
`and a symbol 122 which designates a northern direction points to the right in
`the 3rd screen.
`Hayashida at 24:9-20.
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 16 of 42
`
`

`

`Hayashida at Fig. 20; see also id. at Figs. 19 and 21.
`
`
`
`
`FIG. 27 shows a flow chart of the guide/display processing of the 5th
`embodiment. When the car deviate from the guide route during the guide, at the
`guide/display processing of this 5th embodiment, a return route which returns
`to the guide route and another route which is searched newly or the previous
`route are shown respectively at the divided screen.
`Hayashida at 30:50-55.
`
`
`FIG. 32 shows the other example of FIG. 24 and FIG. 25 of the 4th embodiment
`and FIG. 30 of the 5th embodiment. In this example, a distance and time
`information 186 are shown with the guide route. This distance and time
`information indicate a distance and required time to the end point of guide
`route. For example this distance and time information are the distance
`
`OMM_US:77143848.1
`
`
`17
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 17 of 42
`
`

`

`according to the route and the required time to travel this distance by the
`prescribed speed, when traveling guide route 182 to the end point of guide
`route in FIG. 32.
`
`Travelling distance to the end point of the guide route 184 and traveling time
`is also shown in the 3rd screen 110 of FIG. 32 as the distance and time
`information 186. FIG. 33 shows the guide route which is changed from FIG. 32
`when the step SL26 etc. of FIG. 22 is executed. The distance and time
`information 186 of this FIG. 33 indicate the travelling distance and the required
`time to the end point of the guide route 188. Furthermore the 3rd screen 110 of
`FIG. 33 is copied from the 2nd screen of FIG. 32.
`
`Further in the FIGS. 32 and 33 the distance and time information 186 may
`include the VICS information (construction information or information at the
`congestion road in the route etc.) which is received by data sending and the
`data transmitter/receiver unit 27 and may include the number of times of the
`turning right or left in the this route and so on. This VICS information can
`include congestion information at a parking lot around the route and so on.
`Hayashida at 34:49-35:8.
`
`
`
`OMM_US:77143848.1
`
`
`18
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 18 of 42
`
`

`

`Hayashida at Figs. 32 and 33.
`
`
`
`
`FIG. 34 shows the flow chart of the guide/display processing of the 6th
`embodiment. At the 6th embodiment, the screen is divided when the car deviates
`from the first guide route. The road map which includes the part of the deviating
`guide route and the present position of the car is shown at the divided 2nd
`screen. The previous guide-route which is displayed before the division or the
`new re-searched whole guide route to the destination is shown at the divided
`3rd screen. The previous guide route from which the car deviates is searched
`by the route search processing (step SA4) of FIG. 5 or the whole route
`researching (step SR20) which is described later.
`19
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 19 of 42
`
`

`

`Hayashida at 35:36-47.
`
` FIG. 46 shows the application example of the display of the 2nd screen. For
`example, the return route is shown by arrow 268. Further when the return route
`isn't displayed in the 2nd screen, only the confluence which the return route and
`the previous route cross may be shown by mark 262. Further the confluence
`may be also shown by mark 262 when the return route is displayed. Moreover,
`the direction of the destination in the previous guide route may be shown by a
`direction mark 264 of the destination.
`Hayashida at 49:3-11.
`
`
`Id. at Fig. 46; see also id. at Figs. 38 and 39.
`
`
`20
`
`
`
`OMM_US:77143848.1
`
`
`Apple v. Maxell
`IPR2020-00407
`Maxell Ex. 2008
`
`Page 20 of 42
`
`

`

`Next the processing to display the direction mark 264 of the destination which
`shows the direction of the destination in the computed intersection point CP is
`executed. For example, this processing is executed as follows. The character
`pattern data of the arrow which is used for the direction mark of the destination
`is stored in the information memory unit 37 or the ROM 4. Then the character
`pattern of the arrow is turned to show the above direction of the destination
`according to the above intersection point data CP. This character pattern data
`is sent to image processor 9 and is changed into the image data. Therefore the
`direction mark of the destination 264 is shown in the coordinate which is
`designated by the intersection point data CP. Actually as the tip of the arrow
`agree with the coordinate of the intersection point data CP, A deviation to the
`intersection point data CP is provided for the coordinate of the direction mark
`of the destination.
`Hayashida at 49:44-60.
`
`(8) In addition to the above matter, the map which is the displayed in the 3rd
`screen may be the map information which shows around the car, the start point,
`the destination, the dropping in place or the chosen optional part. Then the map
`which connects with one azimuth of all the azimuths of the map which is
`displayed in this 3rd screen may be displayed in the 2nd screen. In other words,
`a map around the destination (the dropping in place, the specification place or
`so on) is displayed in the 3rd screen. Moreover the map which is connected
`with the map of this 3rd screen and contains the guide route which connects
`from the start point to the destination may be displayed in the 2nd screen.
`Hayashida at 74:41-52.
`
`Moreover the map which connects with the 3rd screen and is displayed in the
`2nd screen may be the map which connects on the running prolongation of the
`car. In other words, a map around the destination is displayed in the 3

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket