Professional Documents
Culture Documents
Patents: Image Data Processing System
Patents: Image Data Processing System
Patents JPH05191902A
Abstract
JPH05191902A
PURPOSE:To process enormous volume of input image data such as front image data of a traveling
Japan
vehicle in a short time. CONSTITUTION:Partially hatched images are cut out from the image in each
frame outputted from a television camera located at a position in front of a vehicle. The region of the
partial image is set such that the image data in front of the vehicle can be obtained continuously based Find Prior Art Similar
on traveling speed of vehicle, frame frequency of television camera, view angle of the television
camera, and the like. Thus cut out partial image is then cut off at ceiling parts S, S' and subjected to Other languages: Japanese
conversion such that a left ceiling part A1, left side face part B, oor part C, right side face part D, and
Inventor: Hiroshi Sasama, 宏 笹間
right ceiling part Aa are arranged linearly thus creating a wall face developing image 2. Since only
partial images are processed, volume of image data to be processed can be reduced drastically.
Worldwide applications
1991 JP
[0006]
In order to solve the above problems, the present invention provides an image data
processing system for processing image data in front of a moving body traveling on a
predetermined route, which is output from an image input device. The image in front
of the moving body is stored in the image memory, and from the image stored in the
image memory, a frame-shaped partial image having a predetermined width in the
horizontal direction and the vertical direction at a predetermined position in the
screen is cut out, The partial image was cut at a predetermined position of the frame
to generate linearly expanded wall surface image data.
Further, according to the present invention, in the image data processing method for
processing the image data in front of the moving body traveling on the determined
route, the wall surface expanded image is formed on the inner surface based on the
wall surface expanded image data created from the partial images. Create the pixel
data of the wall surface expansion image at each position of the tubular tunnel and
the inner surface of the tunnel having the, set the viewpoint and its viewing angle at a
predetermined position in the tunnel or in front of the tunnel, and intersect the
viewing angle and the tunnel. Create a screen at the position where the straight line
connecting the viewpoint and any pixel on the screen intersects the tunnel, and
obtain the pixel data of the reconstructed image using the pixel data of the wall
surface expansion image in the intersecting region. I did it.
[0008]
According to the present invention, since the image data processing system is
con gured as described above, the image input device stores the image in front of
the moving body in the image memory, and the image processing device stores each
image stored in the image memory. Wall-expanded image data obtained by cutting
out a frame-shaped partial image having a predetermined width in the horizontal and
vertical directions at a predetermined position on the screen, cutting the partial image
at a predetermined position of the frame, and expanding it linearly. To create.
Further, according to the present invention, the image processing device is based on
the wall surface developed image data created from the partial images,
Create the pixel data of the wall surface expansion image at each position of the
cylindrical tunnel and the tunnel inner surface having the wall surface expansion
image on the inner surface, set the viewpoint and its viewing angle at a
predetermined position in the tunnel or in front of the tunnel, and Create a screen at
the position where the tunnel and the tunnel intersect, nd the region where the
straight line connecting the viewpoint and any pixel on the screen intersects the
tunnel, and reconstruct it using the pixel data of the wall surface expansion image in
the intersecting region. Get image pixel data.
[0010]
Embodiments of the present invention will now be described in detail with reference
to the drawings. FIG. 1 is an explanatory diagram of an embodiment in which the
image data processing method according to the present invention is applied to the
processing of image data in front of a railway vehicle. In the gure, (a) is an image of
the (i + 1) th frame output from a television camera provided at a predetermined
position of a railway vehicle (not shown; hereinafter, referred to as a vehicle), and (b)
is an image of the i-th frame. , (C) are images of the (i-1) th frame. The hatched
portion in the image of each frame is the partial image 1 cut out in each frame.
Further, (d) is a demapping image formed by demapping the cut out partial image,
and (e) is a reconstructed image reconstructed by remapping the demapping image.
Next, the cut out partial image 1i-1 is cut at the ceiling portions S and S ', and as
shown in FIG.
1, the left side wall surface portion B, the oor portion C, the right side wall surface
portion D, and the right side ceiling portion A2 are converted so as to be aligned
linearly, and the wall surface developed image 2i-1
To create. Note that when the width of the partial image 1 is set to a plurality of lines,
the number of pixels inside the partial image 1 becomes smaller than the number of
pixels outside, so pixels are created by interpolation processing.
Next, in the image of the i-th frame shown in (b), the image inside the frame-shaped
portion inside the partial image 1i-1 of the i-1th frame comes to the outermost side of
the screen, so Image 1i is cut out and a wall surface developed image 2i is created.
Hereinafter, the partial images are sequentially cut out from the images of the
respective frames starting from the (i + 1) th frame shown in (a), and are linearly
arranged as shown in (d) to create the wall surface developed image 2.
The above is the demapping process. As shown in (e), the remapping process is a
reverse process of the demapping process, and is a process of reconstructing an
image at an arbitrary position and viewpoint from the wall surface developed image 2
created by the demapping process. For details, refer to FIG. 2 and FIG.
explain.
FIG. 2 shows a cross section of the tunnel cut vertically along the longitudinal
direction. Therefore, 3 is a ceiling part and 4 is a oor part. Further, 6 is a screen set in
a tunnel section perpendicular to the longitudinal direction of the tunnel.
The scales on the wall surface and the screen indicate the pixel unit of the image on
each surface. That is, here
The width of the partial image in the demapping process is 1 pixel.
Next, in step S3, one viewpoint position for which a reconstructed image is to be
created is selected in the viewpoint moving path in the rectangular tunnel. Next, in
step S4, a screen for creating an image when looking forward from the viewpoint at a
viewing angle θ is set. The position and size of this screen are arbitrary, but it is
generally desirable to extend the viewing angle θ so as to intersect the rectangular
tunnel wall surface, as shown in FIG. It is easy to assume that this screen is
transparent.
Next, in step S5, one pixel on the screen (screen pixel) is selected, the straight line
connecting the viewpoint and the grid points at the four corners thereof is extended,
and the intersection with the rectangular tunnel wall surface is obtained. The area
surrounded by these four intersections is the projection area of the screen pixel.
Next, as step S6,
The wall surface demapping image data in the screen pixel projection area is used to
determine the pixel data on the screen. Although various methods are conceivable as
the algorithm for determining the pixel data in this case, it is generally preferable to
use the average value of the wall surface demapping image data in the pixel
projection area.
Next, the electronic computer 13 cuts out the outermost portion of the screen from
the digitized image signal stored in the image memory 12 as a partial image 1 in FIG.
Convert to image data. Here, this conversion process will be described in more detail.
Now, assuming that the image signal output from the image input device 11 is 512
dots × 512 dots in length × width, the number of pixels of the partial image cut out by
the electronic computer 13 is 2048 dots in each frame. As shown in FIG. 5A, the
electronic computer 13 sequentially samples the pixels stored in the image memory
12 counterclockwise from the point S at the upper center of the screen, and as
shown in FIG. S is 0
Then, the address is converted so that S ′ becomes 2047. If this conversion
processing is performed for each frame of the digitized image signal stored in the
image memory 12, the wall surface expanded image 2
To obtain a digitized image signal. The address-converted image signal is stored in a
main storage memory (not shown) of the electronic computer 13, and when the
amount exceeds the storage capacity of the main storage memory, the auxiliary
storage device 14
Remember.
The expanded wall image 2 is created by the auxiliary storage device 1.
This can be performed by reading the digitized image signal stored in No. 4 and
converting it into an analog signal and outputting it by the image output device 15.
Next, the remapping operation of the image data processing system in the
embodiment of the present invention will be described with reference to FIGS.
First, a tubular tunnel having a rectangular cross section in FIG. 2 is created, a screen
6 is set, and a viewpoint 5 is set by using a three-dimensional computer graphics
technique. Here, the wall surface developed image 2 at each position in the tubular
tunnel created by using the three-dimensional computer graphics technology 2
The pixel data (luminance level, etc.) of the
It is stored in the main memory within 3.
Next, the straight line connecting the viewpoint 5 and any pixel on the screen 6 is
extended to nd the coordinates of the intersection with the tubular tunnel. Then, the
pixel data of the wall surface developed image at the intersection is read from the
main memory and used as the reconstructed image data of the pixel. If this process
is performed for all the pixels on the screen 6, the reconstructed image data seen
from the viewpoint 5 is obtained, and if it is displayed or printed by the image output
device 15, the reconstructed image can be seen. it can.
The reconstructed image in the present invention can be created at any position,
viewing angle, and time interval (speed different from that at the time of image
capturing). In addition, even if the image input by the image input device is an image
on a straight track, by changing the shape of the tubular tunnel with a rectangular
cross section having a wall surface expansion image created on the basis of it, It is
possible to create a curved tunnel image, an uphill tunnel image, etc. according to the
shape. FIG. 6 shows this state.
In the above embodiment, the image is stored and processed in the image memory
frame by frame. However, when one frame is composed of two elds as in the
standard television scanning system, the eld is stored. Each may be stored in the
image memory and processed.
[0029]
As described above in detail, according to the present invention, the image of each
frame stored in the image memory has a predetermined width in the horizontal and
vertical directions at a predetermined position in the screen. Since the frame-shaped
partial image is cut out and the wall surface expanded image data is created, the
amount of image data to be processed is greatly reduced without losing image
information (for example, in the above-described embodiment, 512 × 512 for each
frame).
Of the dots, only 512 × 4 dots are processed, so
The throughput can be 1/128). Therefore, it becomes possible to e ciently process
a large amount of image data such as the image data in front of the traveling vehicle,
and it is possible to contribute to the realization of a self-excited inspection system
by image processing instead of the track patrol inspection in railways. Be expected.
Further, according to the present invention, since the reconstructed image can be
created at an arbitrary position, viewing angle, time interval, etc., the reconstructed
image can be obtained under free conditions different from reality. Therefore, it is
effective for creating an image in a driving simulator or the like.
[Explanation of symbols]
1 Partial image 2 Wall expansion image 3,4 Cylindrical tunnel 5 Viewpoint 6 Screen
11 Image input device 12 Image memory 13 Electronic computer 14 Auxiliary storage
device 15 Image output device
Cited By (6)
JP2009295107A * 2008-06-09 2009-12-17 Kajima Corp Guidance system and guidance method
JP2013073504A * 2011-09-28 2013-04-22 Casio Image transition device, image transition method, and program
Comput Co
Ltd
WO2014162852A1 * 2013-04-04 2014-10-09 ソニー株式 Image processing device, image processing method and program
会社
US10093332B2 2014-06-25 2018-10-09 Mitsubishi Device for creating construction gauge measurement diagram, device for creating construction
Electric gauge measurement diagram data, and method for creating construction gauge measurement
Corporation diagram
Similar Documents
KR100453530B1 2004-10-20 Apparatus and method for drawing three dimensional graphics by converting two dimensional polygon data to three dimensional polygon
data
CN111582080A 2020-08-25 Method and device for realizing 360-degree all-round monitoring of vehicle
CN101006469A 2007-07-25 System and method for creating a panoramic view of a volumetric image
JPH0850469A 1996-02-20 Magni ed display device for large screen with high resolution
JP2004220102A 2004-08-05 Image processing device and method and vehicle monitoring system
JP3267107B2 2002-03-18 Background image creation method and image processing apparatus
JP3465005B2 2003-11-10 Gaze point detection method and gaze point detection device
JP2019154018A 2019-09-12 Vehicle periphery display system, display control unit, and computer program
WO1994018639A1 1994-08-18 Method and apparatus for generating well-registered subtraction images for digital subtraction angiography
JP2000348196A 2000-12-15 Three-dimensional picture generation device and environment map generation method
Legal Events
1996-06-04 A01 Written decision to grant a patent or to grant a registration (utility model) Free format text: JAPANESE INTERMEDIATE CODE: A01
Concepts
machine-extracted
Download Filter table