Page 1

Resolution-enhanced three-dimension / two-

dimension convertible display based on integral

imaging

Jae-Hyeung Park, Joohwan Kim, Yunhee Kim and Byoungho Lee

Optical Engineering and Quantum Electronics Laboratory

School of Electrical Engineering, Seoul National University, Kwanak-Gu Shinlim-Dong,

Seoul 151-744, Korea

byoungho@snu.ac.kr

http://oeqelab.snu.ac.kr

Abstract: A scheme for the resolution-enhancement of a three-

dimension/two-dimension convertible display based on integral imaging is

proposed. The proposed method uses an additional lens array, located

between the conventional lens array and a collimating lens. Using the

additional lens array, the number of the point light sources is increased far

beyond the number of the elemental lenses constituting the lens array, and,

consequently, the resolution of the generated 3D image is enhanced. The

principle of the proposed method is described and verified experimentally.

©2005 Optical Society of America

OCIS codes: (110.2990) Image formation theory, (100.6890) Three-dimensional image

processing, (220.2740) Geometrical optics, optical design

References and links

1.

G. Lippmann, “La photographie integrale,” Comptes-Rendus Acad. Sci. 146, 446-451 (1908).

2.

F. Okano, H. Hoshino, J. Arai, and I. Yuyama, “Real-time pickup method for a three-dimensional image

based on integral photography,” Appl. Opt. 36, 1598-1603 (1997).

3.

S. Manolache, A. Aggoun, M. McCormick, N. Davies, and S. Y. Kung, “Analytical model of a three-

dimensional integral image recording system that uses circular and hexagonal-based spherical surface

microlenses,” J. Opt. Soc. Am. A. 18, 1814-1821 (2001).

4.

T. Naemura, T. Yoshida, and H. Harashima, “3-D computer graphics based on integral photography,” Opt.

Express 8, 255-262 (2001), http://www.opticsexpress.org/abstract.cfm?URI=OPEX-8-4-255.

5.

J.-H. Park, Y. Kim, J. Kim, S.-W. Min, and B. Lee, "Three-dimensional display scheme based on integral

imaging with three-dimensional information processing," Opt. Express 12, 6020-6032 (2004),

http://www.opticsexpress.org/abstract.cfm?URI=OPEX-12-24-6020.

6.

S.-H. Shin and B. Javidi, “Speckle reduced three-dimensional volume holographic display using integral

imaging,” Appl. Opt. 41, 2644–2649 (2002).

7.

S.-H. Hong, J.-S. Jang, and B. Javidi, "Three-dimensional volumetric object reconstruction using

computational integral imaging," Opt. Express 12, 483-491 (2004),

http://www.opticsexpress.org/abstract.cfm?URI=OPEX-12-3-483.

8.

L. Erdmann and K. J. Gabriel, “High-resolution digital integral photography by use of a scanning microlens

array,” Appl. Opt. 40, 5592-5599 (2001).

9.

H. Liao, M. Iwahara, N. Hata, and T. Dohi, "High-quality integral videography using a multiprojector,"

Opt. Express 12, 1067-1076 (2004), http://www.opticsexpress.org/abstract.cfm?URI=OPEX-12-6-1067

10. B. Lee, S. Jung, and J. -H. Park, “Viewing-angle-enhanced integral imaging using lens switching,” Opt.

Lett. 27, 818-820 (2002).

11. J. S. Jang, Y.-S. Oh, and B. Javidi, “Spatiotemporally multiplexed integral imaging projector for large-scale

high resolution three-dimensional display,” Opt. Express 12, 557-563 (2004),

http://www.opticsexpress.org/abstract.cfm?URI=OPEX-12-4-557

12. Y. Kim, J.-H. Park, S.-W. Min, S. Jung, H. Choi, and B. Lee, "Wide-viewing-angle integral three-

dimensional imaging system by curving a screen and a lens array," Appl. Opt. 44, 546-552 (2005).

13. J. S. Jang, and B. Javidi, “Three dimensional synthetic aperture integral imaging,” Opt. Lett. 27, 1144-1146

(2002).

(C) 2005 OSA 21 March 2005 / Vol. 13, No. 6 / OPTICS EXPRESS 1875

#6287 - $15.00 USReceived 11 January 2005; revised 24 February 2005; accepted 2 March 2005

Page 2

14. J. S. Jang, and B. Javidi, “Improved viewing resolution of 3-D integral imaging with nonstationary micro-

optics,” Opt. Lett. 27, 324-326 (2002).

15. J. Hong, J.-H. Park, S. Jung and B. Lee, "A depth-enhanced integral imaging by use of optical path

control," Opt. Lett. 29, 1790-1792 (2004).

16. J. S. Jang, and B. Javidi, “Large depth-of-focus time-multiplexed three-dimensional integral imaging by use

of lenslets with nonuniform focal lengths and aperture sizes,” Opt. Lett. 28, 1924-1926 (2003).

17. J.-H. Park, H.-R. Kim, Y. Kim, J. Kim, J. Hong, S.-D. Lee, and B. Lee, "Depth-enhanced three-

dimensional-two-dimensional convertible display based on modified integral imaging," Opt. Lett. 29, 2734-

2736 (2004).

1. Introduction

Integral imaging is a technique for displaying three-dimensional (3D) images using a lens

array. The feature of the lens array that spatially samples and captures the light field from the

objects is exploited in the integral imaging to capture and display 3D images [1]. Full color,

full parallax real-time 3D images provided by the integral imaging make it attractive [2-7] and

considerable efforts have been made to enhance its viewing parameters [8-16]. A novel

integral imaging scheme with enhanced depth and 3D / two-dimension (2D) convertibility was

recently reported [17]. In this approach, the lens array, which is located behind the

transmission-type spatial light modulator (SLM) and a polymer-dispersed liquid crystal

(PDLC) attached to the lens array, enhance the expressible depth range substantially and

enable 3D/2D conversion. With these superior features, integral imaging acquires a much

wider range of applications, and approaches the level of commercialization. This is because

only 3D/2D convertible display techniques can infiltrate into commercial markets for 3D TV.

Our previous work [17] using PDLC was the first demonstration of a 3D/2D convertible

integral imaging technique (that has both horizontal and vertical parallaxes unlike the

lenticular or parallax barrier technique). The remaining problem is resolution. In fact, depth

enhancement and 3D/2D convertibility can be achieved, but at the expense of the resolution.

Therefore it is necessary to develop a method that compensates for resolution degradation

while retaining the useful features, i.e., depth enhancement and 3D/2D convertibility.

In this paper, we report on a resolution enhancement scheme for a depth-enhanced 3D/2D

convertible display based on integral imaging. The proposed method utilizes an additional

lens array to generate an excess of point light sources, thus permitting the resolution of the

generated 3D images to be enhanced.

Light

source

Collimating

lens

Lens arraySLM

PDLC

(diffuse)

2D image2f

Light

source

Collimating

lens

Lens array

SLM

PDLC

(transparent)

3D image

2f

Point light

sources

Light

source

(a) (b)

Fig. 1. Schematic diagram of 3D/2D convertible integral imaging (a) 2D mode (b) 3D mode

(C) 2005 OSA21 March 2005 / Vol. 13, No. 6 / OPTICS EXPRESS 1876

#6287 - $15.00 US Received 11 January 2005; revised 24 February 2005; accepted 2 March 2005

Page 3

2. 3D/2D convertible integral imaging and resolution limitation

Figure 1 shows the concept of the 3D/2D convertible integral imaging. The 3D/2D convertible

integral imaging consists of a collimated incoherent light source, a PDLC, a lens array and an

SLM. Conversion between the 3D and 2D modes is achieved by controlling the diffusing rate

of the PDLC. In the 2D mode, the PDLC is set to be diffusive. The collimated light is

scattered by the PDLC and this scattered field is then relayed to the SLM. Each pixel in the

SLM is then illuminated effectively from all directions, and consequently, 2D images are

observed on the SLM plane with full resolution and viewing angle. In the 3D mode, the PDLC

is set to be transparent. In this case, the collimated light rays pass through the PDLC without

scattering and are imaged into an array of point light sources at the focal plane of the lens

array. The light rays from each point light source are modulated according to the direction of

their propagation, and finally integrated into 3D images.

In the 3D mode, the resolution is limited by the number of point light sources. Figure 2

shows this point. As shown in Fig. 2, the light rays from each point light source are modulated

by the SLM. Hence, a unit comprised of one elemental lens, a point light source, and the

corresponding region on the SLM is effectively one pixel that emits light rays of different

colors or intensities according to the observation directions and, as a result, acts as one pixel

in the generated 3D images. Since the role of the elemental lens and the SLM is merely the

formation and modulation of the point light sources, the key parameter for determining the

resolution of the generated 3D images is the number of point light sources. That is, the

resolution of the generated 3D images is the same as the number of point light sources. In

order to enhance the resolution of generated 3D images, it is necessary to generate more point

light sources with a smaller spacing between them. One straightforward method for achieving

this end is to use a lens array with more elemental lenses and small elemental lens pitches.

Such a lens array should reach the desired size of the display panel (e.g. 12-inch or 15-inch)

with uniform elemental lenses whose pitch is comparable to the pixel pitch of conventional

2D displays (e.g. 200 um) and f-number is sufficiently small to preserve a reasonable viewing

angle, which is difficult to realize at this time. This high requirement stimulates research for

an alternate resolution compensation method that utilizes lens arrays that can be fabricated

more readily.

Point light

source plane

ff

1 elemental

image region

ϕ

Integrated 3D image

act as one pixel

in 3D image

θ

f

f

ϕ

Fig. 2. Limitations in resolution of the 3D/2D convertible integral imaging

(C) 2005 OSA 21 March 2005 / Vol. 13, No. 6 / OPTICS EXPRESS 1877

#6287 - $15.00 US Received 11 January 2005; revised 24 February 2005; accepted 2 March 2005

Page 4

3. Configuration of the proposed method

The proposed method enhances the resolution of the generated 3D images using an additional

lens array. A schematic diagram of the proposed method is shown in Fig. 3. In addition to the

lens array between the PDLC and the SLM, an additional lens array is inserted between the

collimating lens and the PDLC. In the 3D operation mode, collimated light rays from the light

source are imaged by the first (additional) lens array into the first array of point light sources

at the focal plane of the first lens array. Each elemental lens of the second (original) lens array

then forms an image of this first point light source array again, and, consequently, an array of

point light sources with far more point light sources than the number of the elemental lenses in

the second lens array is generated at the image plane of the second lens array. Each one in this

abundant number of second point light sources serves as one pixel in the generated 3D images,

as explained above, and, hence, the resolution is dramatically enhanced, compared to the

previous configuration shown in Fig. 1. One significant difference between the proposed

method and our previous method [17] is that each elemental lens produces multiple point light

sources in the proposed method, while only one point light source is produced in our previous

method. Therefore, with the lens array, which can be readily fabricated, a high density of

point light sources is easily obtained over a large display area, resulting in high resolution 3D

images.

In the proposed method, the number of elemental images should be same as that of the

second point light sources and each one should be generated with reference to the position of

the corresponding point light source. Since much more point light sources are generated in the

proposed method than in the previous method, the number of elemental images should be

increased while the size of each one is decreased, which may result in a coarser modulation of

the light rays from each point light source than in the previous method if the pixel pitch of the

SLM is fixed.

The number of second point light sources, the spacing between them, and the diverging

angle of the light rays from each point light source are determined by the specifications and

locations of the two lens arrays in the proposed method. The geometry for this configuration is

shown in Fig. 4. The first lens array with focal length f1 and elemental lens pitch ϕ1 forms the

first point light source array at its focal plane. Each elemental lens with pitch ϕ2 and focal

length f2 in the second lens array images these first point light sources at its image plane. Note

Light

source

Collimating

lens

Lens array

SLM

PDLC

Integrated

image

Second

point light

sources

Additional

lens array

First point

light sources

Fig. 3. Schematic diagram of the proposed method

(C) 2005 OSA 21 March 2005 / Vol. 13, No. 6 / OPTICS EXPRESS 1878

#6287 - $15.00 US Received 11 January 2005; revised 24 February 2005; accepted 2 March 2005

Page 5

that not all of the first point light sources are imaged by each elemental lens in the second lens

array because the diverging angle of the light rays from each first point light source is limited.

The diverging angle of each first point light source, ψ1, is given by

⎟⎟

⎠

⎞

⎜⎜

⎝

⎛

=

−

1

1

f

1

1

2

tan2

ϕ

ψ

. (1)

Since the position of the first point light source generated by k-th elemental lens in the first

lens array, y1,k, can be represented by y1,k=kϕ1, it illuminates the second lens array’s elemental

lenses that satisfy

⎟

⎠

⎞

⎜

⎝

⎛

+<<

⎟

⎠

⎞

⎜

⎝

⎛

−

2

tan

2

tan

1

12

1

1

ψ

ϕ

k

ϕ

q

ψ

ϕ

kll

, (2)

where l is the distance between the second lens array and the focal plane of the first lens array,

and q is the index of the elemental lenses in the second lens array. Or equivalently, the q-th

elemental lens in the second lens array is illuminated by the first point light sources whose

index k satisfies kl≤k≤kh where kl and kh are given by

ψ

ϕ

qlk

l

⎟

⎠

⎞

⎜

⎝

⎛

−==

⎟

⎠

⎞

⎜

⎝

⎛

+

2

tan

2

tan

1

12

1

1

ψ

ϕ

h

ϕ

lk

. (3)

From Eqs. (1) and (3), kl and kh can also be represented by

11

2

2 f

l

q

ϕ

kl

−=

ϕ

, (4)

11

2

2 f

l

q

ϕ

kh

+=

ϕ

. (5)

Therefore the first point light sources satisfying (kl≤k≤kh) are imaged into the second point

light sources by the q-th elemental lens in the second lens array at

l

gk

l

g

qy

qk

1

2, , 2

1

ϕ

ϕ

−

⎟

⎠

⎞

⎜

⎝

⎛+=

, (kl≤k≤kh) (6)

where g is the location of the image plane of the second lens array which is given by

2

2

fl

lf

−

g

=

. (7)

The diverging angle of each second point light source which mainly determines the viewing

angle of the generated 3D images is written by

⎟⎟

⎠

⎞

⎜⎜

⎝

⎛

=

−

g

2

tan2

2

1

2

ϕ

ψ

. (8)

Since g is slightly larger than f2 as indicated by Eq. (7), the viewing angle of the proposed

method can be somewhat narrower than conventional method where g=f. Note that the

diverging directions of the second point light sources are, in general, not parallel but vary

(C) 2005 OSA21 March 2005 / Vol. 13, No. 6 / OPTICS EXPRESS 1879

#6287 - $15.00 US Received 11 January 2005; revised 24 February 2005; accepted 2 March 2005