Next Article in Journal
Reconstruction of Radio Environment Map Based on Multi-Source Domain Adaptive of Graph Neural Network for Regression
Previous Article in Journal
LSTM Short-Term Wind Power Prediction Method Based on Data Preprocessing and Variational Modal Decomposition for Soft Sensors
Previous Article in Special Issue
Determination of the Accuracy of the Straight Bevel Gear Profiles by a Novel Optical Coaxial Multi-Ring Measurement Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mind the Exit Pupil Gap: Revisiting the Intrinsics of a Standard Plenoptic Camera

Department of Computer Science, Kiel University, 24118 Kiel, Germany
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(8), 2522; https://doi.org/10.3390/s24082522
Submission received: 22 February 2024 / Revised: 29 March 2024 / Accepted: 5 April 2024 / Published: 15 April 2024
(This article belongs to the Special Issue Short-Range Optical 3D Scanning and 3D Data Processing)

Abstract

:
Among the common applications of plenoptic cameras are depth reconstruction and post-shot refocusing. These require a calibration relating the camera-side light field to that of the scene. Numerous methods with this goal have been developed based on thin lens models for the plenoptic camera’s main lens and microlenses. Our work addresses the often-overlooked role of the main lens exit pupil in these models, specifically in the decoding process of standard plenoptic camera (SPC) images. We formally deduce the connection between the refocusing distance and the resampling parameter for the decoded light field and provide an analysis of the errors that arise when the exit pupil is not considered. In addition, previous work is revisited with respect to the exit pupil’s role, and all theoretical results are validated through a ray tracing-based simulation. With the public release of the evaluated SPC designs alongside our simulation and experimental data, we aim to contribute to a more accurate and nuanced understanding of plenoptic camera optics.

1. Introduction

Plenoptic cameras as initially described by Lippmann [1] and Ives [2] combine a traditional camera with an additional microlens array (MLA) located between the main lens and the sensor. Over the years, two primary designs have been extensively studied and brought to market, the standard plenoptic camera (SPC) [3,4] and the focused plenoptic camera (FPC) [5,6], which mainly differ in the microlens focus distance. Due to their earlier commercialization, larger angular resolution, and simpler decoding process, SPCs remain popular despite certain disadvantages in terms of spatial resolution and depth of field when compared to the multi-focus variant of FPCs [6]. Classical applications for SPCs include depth reconstruction [3] and post-capture refocusing from single shots [4]; as a first step to achieve these, the raw 2D image of a plenoptic camera is usually de-multiplexed and resampled into a 4D light field [7], as shown in Figure 1. For this reparametrization procedure, knowledge about the exact position of each microlens image center (MIC) is crucial, as any inaccuracies in their locations can result in computational errors affecting the quality of the refocused images [8]. Furthermore, a formal connection between the MICs and the plenoptic camera optical setup is required to relate the light field within the camera to the optical reality outside the camera, e.g., to find the correct refocusing parameters for the desired object distance [7,8].
Over the past two decades, a number of studies have delved into the topic of processing plenoptic camera images; however, these have often considered the MICs to be determined by the main lens center or its principal planes as a consequence of reducing the main lens to a simple thin lens. This assumption oversimplifies the actual optics involved. A more accurate representation acknowledges the role of the exit pupil in determining these image centers, as observed in studies by Hahne et al. [8,9]. Despite these advancements, the exit pupil is often ignored in studies relating the light field within the camera to the 3D scene in front of the camera.
Figure 1. Exemplary pipeline for SPC post-shot refocusing. A scene is captured by a virtual SPC, shown here without the housing. The resulting raw image consists of a large number of microlens images and is subsequently decoded into a 4D light field representation, which can be visualized by a subset of the sub-aperture images [7]. By resampling the light field, a refocused image can be created [4]. The correctly focused images shown here were created based on parameters considering the exit pupil as described in Section 2, while the slightly defocused image are results from the directly calculated parameters based on [10] without any exit pupil consideration.
Figure 1. Exemplary pipeline for SPC post-shot refocusing. A scene is captured by a virtual SPC, shown here without the housing. The resulting raw image consists of a large number of microlens images and is subsequently decoded into a 4D light field representation, which can be visualized by a subset of the sub-aperture images [7]. By resampling the light field, a refocused image can be created [4]. The correctly focused images shown here were created based on parameters considering the exit pupil as described in Section 2, while the slightly defocused image are results from the directly calculated parameters based on [10] without any exit pupil consideration.
Sensors 24 02522 g001
In this context, our work aims to again highlight the importance of the exit pupil. To this end, a paraxial model of the SPC under consideration of the exit pupil is first described, directly relating the refocusing shift [4] to the object distance. The expected errors of models ignoring the exit pupil are formally analyzed and later verified through a ray tracing-based simulation of various plenoptic cameras in Blender [11] using real lens data. Subsequently, multiple works in the domain of plenoptic camera calibration are revisited and examined with respect to the need for a more complex lens model. More specifically, the popular work of Dansereau et al. [7] is first revisited, along with the works of Zhang et al. [12] and Monteiro et al. [13] building upon Dansereau’s ideas. In these cases, it can be concluded that the parameters of the respective calibration models are sufficiently general to permit the simplicity of a main lens model without considering the exit pupil. However, this only holds true because these works do not require a specific interpretation of the model parameters. On the other hand, for the work of Pertuz et al. [10], which also employs the decoding from [7] for metric distance measurement, it is shown that the oversimplified main lens model leads to an incorrect interpretation of the metric refocusing model parameters. In summary, our contributions are:
  • A formal deduction of the connection between object distance and sub-aperture image shift considering the exit pupil.
  • A model for the errors resulting from ignoring the exit pupil in this relation.
  • An analysis of the exit pupil’s role in popular works on SPC calibration [7,10,12,13,14].
  • Publicly available SPC designs [15] and a camera simulation framework [16] based on Blender [11] supporting a large database of lens designs and enabling quick generation of new plenoptic camera setups.

1.1. Related Work

Plenoptic Cameras: There are two primary design concepts of plenoptic cameras which have been extensively studied and brought to market. The first, known as the standard plenoptic camera (SPC), was described by Adelson and Wang [3] and later commercialized by Ng [4]. It requires the microlenses to be focused at infinity, i.e., the MLA-to-sensor separation must match the microlens focal length. Consequently, for a scene object placed at the SPC’s focus distance, all sensor pixels behind a single microlens effectively capture a nearly identical segment of this object, albeit from slightly varied perspectives. This results in large angular resolution but low spatial resolution.
Later, the focused plenoptic camera (FPC) was presented by Lumsdaine and Georgiev [17] and extended by Perwass and Wietzke [6] to feature multifocal MLAs for extended depth of field. For an FPC, instead of directly dissecting the scene into its directional components, the microlenses are focused at the scene’s virtual image inside the camera. This arrangement is advantageous in maintaining a greater portion of the conventional camera’s spatial resolution, albeit at the expense of a decrease in angular resolution compared to SPCs.
Despite the advantages of FPCs, especially when combined with multifocal MLAs, in this work we chose to focus on SPCs due to their straightforward image processing pipeline. More specifically, in order to show the effect of the exit pupil, the application of post-shot refocusing is used throughout this work. While this is a possible application for an FPC, the process involves direct [18] or indirect [19] depth estimation, which can introduce artifacts that affect subsequent processing steps. In contrast, for a well-configured SPC the post-shot refocusing only comprises a demultiplexing step [7] for the raw image and a subsequent shift-and-sum procedure for the resulting sub-aperture images. The simplicity of this pipeline reduces the quantity of artifacts resulting from complex interpolation and optimization steps, allowing for more direct analysis of the exit pupil’s effects. It should be noted, however, that this work is intended to be a starting point for further research which will analyze and improve FPC calibration algorithms suffering from the same model flaws as their SPC pendants.
SPC Calibration: The calibration of plenoptic cameras plays a crucial role in relating the captured light field within the camera to the 3D world in front of the camera. To this end, Dansereau et al. [7] presented a method for demultiplexing the raw image of an SPC into a 2D array of sub-aperture images and using these for the geometric calibration. Due to its popularity and accessibility in the form of a Matlab toolbox, this work is still being used as a base for publications concerning the processing of SPC data. With respect to calibration models, Zhang et al. [12] and Monteiro et al. [13] modified the ideas of [7] to associate a plenoptic camera with an equivalent multi-camera array. Both of these works made direct use of the decoding process proposed by Dansereau et al. [7]. Pertuz et al. [10] followed the same approach in proposing a focus-based metric for depth estimation.
Due to the popularity of the demultiplexing process from [7], the revision of the previous literature in Section 4 focuses on this method and the approaches based on it. Nevertheless, there has been further work, which in part also reduces the main lens to a thin lens, ignoring the effects of the exit pupil. One such calibration approach, which directly uses the line features in the microlens images, is the one presented by Bok et al. [20]. Using a similar model, Zhao et al. [21] performed SPC calibration based on the plenoptic disc features. Both Thomasen et al. [22] and Suliga and Wrona [23] directly estimated the MLA pose and microlens pitch; however, neither related the captured light field to the scene-side light field. Similar to the approaches related to [7], these works assume the microlens images centers to be projections of the main lens center, i.e., the center of the camera-side principal plane. On the other hand, Hahne et al. [8,9] described the refocusing distance based on known main lens and MLA parameters under consideration of the exit pupil. Further improvements to aspects of the calibration pipeline which acknowledge the exit pupil have been presented by Schambach et al. [24], who increased the MIC detection accuracy, and Mignard-Debise and Ihrke [25], who analyzed the effect of vignetting on calibration models.
Of these works, Hahne et al. [9] considered the exit pupil and its connection to the microlens image geometry in a similar fashion to this work; however, they did not situate this within the direct context of pre-existing calibration methods by a comparative evaluation, nor provide an analysis of the expected errors resulting from oversimplified lens models. In order to further validate our model, which establishes the refocusing distance in terms of the two-plane parameterization, its equivalence to the chief ray intersection model in [9] is formally proven in Section 4.1.
FPC Calibration: Despite this work focusing on SPCs, as explained above, there is relevant work to be found in the related field of FPC calibration. Johannsen et al. [26] described a metric reprojection model for FPCs incorporating a radial distortion model. This was further enhanced by Heinze et al. [27], who included the tilt and shift of the main lens as well as multi-focus MLAs. Further improvements to the distortion model were presented by Zeller et al. [28]. All of these approaches are based on the reconstruction of the virtual scene between the MLA and the main lens followed by association of these virtual 3D points with known scene points. In contrast, Noury et al. [29] proposed an approach that works directly on the microlens images, i.e., by associating the scene points with their projections on the sensor without the intermediate step of calculating virtual depths. This method, however, is limited to single-focus FPCs. and models the microlenses as simple pinholes. Nousias et al. [30], on the other hand, featured a more complete microlens model in their work and directly included the estimation of multiple microlens focal lengths in their approach. Wang et al. [31] presented a two-step model consisting of a forward projection from the scene into the camera and a second projection from the virtual image to the sensor. More recently, Labussiere et al. [32] proposed simultaneous calibration of the different microlens types in a multi-focus plenoptic camera by incorporating defocus blur into the features used for parameter optimization.
None of the listed methods for FPC calibration directly consider the exit pupil, and while most of these works, all of which require the identification of MICs, incorporate a scaling between the grid of microlens centers and the grid of MICs, this is usually a result of projecting the main lens center, i.e., the center of the camera-side principal plane, through the microlens centers. However, as observed by Hahne et al. [8,9] for SPCs and confirmed in Section 5, the MICs actually result from a projection of the exit pupil’s center. Thus, using the distance between the simplified main lens plane and the MLA for both the image formation model and the calculation of MICs could inadvertently reduce the degrees of freedom of the model. While this might be desirable in terms of increased stability during parameter optimization, such reduction should be analyzed for FPC models. For reasons of clarity and comprehensibility, we decided against including the topic of FPC calibration in this work, and leave this for future efforts.
Lens Models and Simulation: In the domain of ray tracing-based camera simulation, realistic main lens models which consider all lens components and their respective properties have been used for over two decades, either explicitly by direct modeling, as in Kolb et al. [33] and Wu et al. [34], or implicitly via learned black-box lens models, as proposed in Zheng et al. [35]. Regarding plenoptic cameras, most previous works have used oversimplified models for rendering, such as pinhole cameras or multi-camera arrays modeling the MLA without a model for the main lens [36,37,38]. More recently, Nürnberg et al. [39] as well as our own group [40] have provided simulations of plenoptic cameras without oversimplifying the main lens. Due to this familiarity, we extended our previous work for our synthetic experiments.

1.2. Organization

In Section 2, the general lens model and two-plane parameterization are explained first, before deducing the refocusing model under consideration of the exit pupil. Subsequently, Section 3 provides a formal analysis of the expected errors when dismissing the exit pupil. In Section 4, previous works are revisited with a focus on the need for more complex lens models. Finally, our deductions are validated with synthetic experiments in Section 5.

2. SPC Optics

2.1. Preliminaries—Lens Models

The thin lens model describes a lens by assuming it as infinitely thin and only refracting light at a single lens plane. The relation between the real scene and the lens image in this model is described by the equation
1 f M = 1 o + 1 i ,
where f M is the focal length of the lens, o is the object distance, and i is the image distance, with both measured from the refraction plane. This concept can be extended to a thick lens model by expanding the refraction plane into two principal planes, H scene and H cam , between which a traced light ray is considered to run parallel to the optical axis [41,42]. Furthermore, a combination of thick lenses, such as the main lens of a plenoptic camera, can again be represented as a single thick lens [42]. As visualized in Figure 2, the object distances o and image distance i are then measured based on the positions of the principal planes, and Equation (1) remains valid.
In addition to this model, it is possible to consider the exit pupil, i.e., the image of the aperture stop viewed towards the image plane. It defines the size and location of the virtual aperture in the optical system [42] and, as pointed out by Hahne et al. [8,9], determines the positions of the microlens image centers (MIC) on the sensor. As empirically shown in Section 3, the exit pupil and H cam rarely coincide; accordingly, a systematic error can be introduced when a plenoptic camera image is de-multiplexed based on MICs that are incorrectly estimated under the premise that the main lens follows the thin lens model without considering the exit pupil.

2.2. Preliminaries—Light Field Parametrization

Despite being a standard tool when working with light field data, the two-plane parameterization as described by Levoy and Hanrahan [43] and used in various popular works, including the work of Dansereau et al. [7] and Ng et al. [4] is reiterated in this section for two reasons. First, the previous descriptions do not consider the exit pupil; second, the literature is not consistent in terms of the underlying data representation. While [7] used the raw camera image indexed by integer pixel coordinates as the basis for their description, [4] assumed known metric coordinates for every pixel. We follow the approach of [7] in order to facilitate the reproduction of our results.
For a given SPC following the thick lens model with an exit pupil as visualized in Figure 2, the light field inside this camera can be parameterized using two planes: the MLA, which serves as virtual sensor plane, and the exit pupil plane, which can be interpreted as a virtual lens plane. By following the decoding process of Dansereau et al. [7], the 4D light field can be parameterized as L F ( i , j , k , l ) , with integer indices ( k , l ) for the uniformly sampled sub-aperture image and ( i , j ) for the pixel coordinates in that image.
The corresponding metric parameterization L ˜ F ( s , t , u , v ) describes the intensity of light captured at the MLA plane point ( s , t , d ) coming from the exit plane point ( u , v , X ) . In accordance with Ng et al. [4,44], the distance between these two parameterization planes is denoted as F : = d X . Note, however, that Ng et al. [4] implicitly assume X = 0 , as the main lens in their work is modeled by a thin lens. While this is approximately correct for the two tested main lenses in [4], a Zeiss Planar T* 2/80 with X = 7.8 mm = 0.098 · f M and a Zeiss Sonnar T* 2.8/140 with X = 4 mm = 0.029 · f M (compare [45,46]). Section 3 shows that this assumption does not hold in general.
To calculate the metric parameterization from a given integer parameterization, note that the pixel pitch S T of the virtual sensor, i.e., the step size in the S T -plane, corresponds to the microlens pitch, i.e., S T = d ML ; as shown in Figure 3, the step size in the virtual lens plane, i.e., the U V -plane, can be calculated by means of the triangle equality as U V = s p x · F f m , where s p x and f m respectively denote the pixel size and the microlens focal length. With these step sizes, the light field parameterized in metric coordinates ( s , t , u , v ) is provided by
L ˜ F ( s , t , u , v ) = L F s S T , t S T , u U V , v U V .
Note that the metric coordinates ( s , t , u , v ) might not be integer multiples of their respective step sizes; accordingly, querying the corresponding values from the integer parameterization L F could require additional interpolation steps.
With the described light field parameterization, we can now reproduce the resampling steps necessary to refocus the image by moving the virtual sensor plane while considering the position of the exit pupil.

2.3. Light Field Refocusing with Exit Pupil

In order to refocus the virtual sensor image onto an object at a distance o measured from H scene , the virtual sensor needs to be placed at a distance i measured from H cam according to the thin lens Equation (1). This corresponds to a distance F : = i X between the U V -plane (exit pupil) and the virtual sensor, as visualized in Figure 4. By defining the refocusing parameter α = F F as in [4], the thin lens equation can be applied to deduce
α = F F = o · f M o f M X d X = o · ( f M X ) + f M · X ( o f M ) ( d X ) .
For a given integer 4D light field L F ( i , j , k , l ) based on sub-aperture images, as in [7], the relationship between the virtual sensor movement specified by α and the resulting disparity at the original S T -plane is described by the following. While the general deduction is similar to [4], for reproducibility the following calculations are based on integer indexing.
As shown in Figure 4, the metric light field value L ˜ F ( s , t , u , v ) for the modified sensor plane placed at a distance F from the exit pupil can be calculated as follows:
L ˜ F ( s , t , u , v ) = L ˜ F u + s u α , v + t v α , u , v = L ˜ F u α 1 α + s α , v α 1 α + t α , u , v .
Ignoring the image magnification introduced by the movement of the virtual sensor, i.e., setting the step size for the S T -plane to α · S T and defining : = U V S T , it can be deduced that the integer parameterization L F ( i , j , k , l ) for the modified sensor plane corresponds to
L F k · 1 1 α + i , l · 1 1 α + j , k , l .
At this point, the pixel shift S between neighboring sub-aperture images required to refocus onto the desired distance o can be calculated for a given value α as follows:
S ( α ) = · 1 1 α .
Plugging Equation (3) into Equation (5) then yields the direct relation between the object distance o and the disparity S as
S ( o ) = · o ( f M d ) + f M · d o ( f M X ) + f M · X .
This model can be easily reverted to calculate the object or refocusing distance based on a given sub-aperture image shift via
o ( S ) = f M · ( d · S · X ) S · ( f M X ) ( f M d ) .

3. Error Analysis

In the following, we analyze the error that can be anticipated when ignoring the exit pupil, i.e., when setting X = 0 . First, we define the scaling between the S T and U V planes in under this assumption as
˜ : = X = 0 = s p x · d f m · d ML .
We then calculate the pixel disparity based on Equation (6) as
S ˜ ( o ) : = S X = 0 ( o ) = ˜ · o ( f M d ) + f M · d o · f M
along with the object distance, which can be simplified to
o ˜ ( S ) : = o X = 0 ( S ) = f M · d · ˜ S · f M ˜ ( f M d ) .
The relative error of the shift is then calculated by
E R R S ( o ) : = S ˜ ( o ) S ( o ) S ( o ) = X · d · f M o ( d f M ) o · f M ( d X ) .
By describing o as a multiple of the focus distance o f , i.e., o = λ · o f , we obtain
E R R S ( λ ) : = E R R S ( λ · o f ) = X ( λ 1 ) λ o f · X d 1 .
For the error of the object or refocusing distance, two cases are analyzed. First, it is assumed that the correct shift S corresponding to the ground truth o is given; in a second step, the oversimplified model of Equation (10) is used to calculate the object distance. This error can be found in applications measuring the correct shift, e.g., by repeatedly refocusing an image and subsequently using the incorrect object distance calculation in order to estimate the associated metric distances in the scene. This error can be formulated as
E R R o ˜ ( o ) : = o ˜ ( S ( o ) ) o o = X ( d f M ) + o o f X ( f M d ) + X f M d 1 o f 1 o f M ( d X ) + o o f X ( f M d ) + X f M d 1 o f .
Using o = λ · o f again, we obtain
E R R o ˜ ( λ ) : = E R R o ˜ ( λ o f ) = X ( λ 1 ) 2 λ o f 1 X d X ( λ 1 ) λ .
The second case assumes an incorrectly calculated shift S based on Equation (9), which is subsequently used to refocus an image with a refocusing algorithm complying with the correct object distance estimation in Equation (7). This type of error is provided by
E R R S ˜ ( o ) : = o ( S ˜ ( o ) ) o o = f M f M X ˜ 1 o 1 o f ˜ 1 o o f ( f M X ) + f M o o f 1 .
After substituting o = λ · o f , the error can be formulated by
E R R S ˜ ( λ ) : = E R R S ˜ ( λ o f ) = X ( λ 1 ) 2 λ o f X f M 1 λ 2 X .
Now, assuming a camera with a focal length of f M = 100 mm focused at a finite distance, Figure 5 shows exemplary error values for different values of X relative to the focal length.
The visualization shows that all errors diverge for λ 0 with a rate depending on the positional relationship between the exit pupil and the principal plane H cam . Beyond the focus distance at λ = 1 , the object errors again diverge, while the shift error converges according to E R R S ( λ ) λ X o f ( X / d 1 ) . Note that these graphs present an ideal refocusing case free of aliasing artifacts and limiting optical properties such as the depth of field; hence, our later experiments only verify a section of these results within the respective physical and image processing limits.
In summary, these examples show a large deviation between the estimated refocusing distances in models with and without consideration of the exit pupil whenever there is a non-zero distance X between the exit pupil and H cam . This leads to the question of how prevalent a significant X 0 is in off-the-shelf main lenses. To answer this question, the data of 866 DSLR lenses listed by Claff [47] were collected, then X and f M for each lens were calculated via paraxial ray tracing. The resulting data in Figure 6 show a nearly linear connection between the focal length of a lens and the distance X, with a Pearson correlation coefficient of 0.8994 . Fitting a linear model to these data results in the non-zero function X ( f M ) = 0.7108 · f M 56.5546 with coefficient of determination R 2 = 0.8089 . Further examination shows that only a small subset, 62 of the 866 lenses, exhibits values for X below 5% of the focal length, i.e., | X | < 0.05 f M . On the other hand, for the other 627 lenses the deviation is larger than | X | > 0.25 f M , and 444 lenses even have values | X | > 0.5 f M . Overall, these data show that the assumption of X 0 is usually not met in reality. Therefore, the exit pupil should be considered when relating the camera-side light field to the scene’s light field.

4. Revisiting SPC Methods

In this section, several previous works are examined with respect to the exit pupil’s role in the respective model deductions.

4.1. Equivalent Ray Model

First, the equivalence between the refocusing model in Equation (7) and the ray intersection model presented by Hahne et al. [9] is proven. Instead of basing the model on the decoding scheme of Dansereau et al. [7] in [9], an approach building upon the intersection of chief rays is presented in order to calculate the refocusing distance for a resampling of the raw plenoptic camera image. A comprehensive notation transfer into our setup is provided in Appendix B.1.
As depicted in Figure 7, the basic idea of [9] is to select two pixels on the sensor that show scene points from the desired focus plane, then trace rays from these through the respective microlens centers. The resulting camera-side intersection determines the distance of the virtual image inside the camera from the main lens; accordingly, the thin lens equation can be applied in order to calculate the corresponding object or refocusing distance. Without loss of generality, the following calculations assume an MLA with one microlens center located on the optical axis of the main lens.
For a given a sub-aperture image shift S in a pixel, as depicted in Section 2.3, this translates to a metric pixel disparity S ^ on the sensor by
S ^ = s p x S ,
with the sign flip resulting from the difference in conventions between this work and [9]. Under the premise of a well-configured plenoptic camera with a regular microlens grid, any two pixels from neighboring microlenses with a disparity of S ^ can be chosen to calculate the image distance. To simplify the calculations, the first ray is chosen to run along the optical axis, as shown in Figure 7, and the second ray is based on the pixel at sensor position d MLI + S ^ passing through the neighboring microlens center. According to [9] (compare Appendix B.1) this leads to two ray functions
f ( z ) = 0 and f ˜ ( z ) = d ML ( d MLI + S ^ ) f m · z + d ML
which intersect at
z i = d ML · f m d ML ( d MLI + S ^ ) .
The crucial element that sets [9] apart from the works reviewed in the following sections is the correct calculation of the microlens image center distance d MLI based on the exit pupil (compare the calculation of u c , j in Table A5) via
d MLI = d ML d X · f m + d ML ,
which yields
z i = d ML f m d ML ( d ML d X · f m + d ML + S ^ ) = ( d X ) d ML f m ( d X ) S ^ + f m d ML .
This intersection results in the image distance i = d z i measured from H cam , and can be used to calculate the object distance by the thin lens equation via
o = 1 f M 1 i 1 = 1 f M 1 d z i 1 = f M · d f M · z i ( d f M ) z i = f M · d · ( ( d X ) S ^ + f m d ML ) f M · ( d X ) d ML f m ( d f M ) · ( ( d X ) S ^ + f m d ML ) ( d X ) d ML f m = f M · d · d X f m d ML S ^ + f M X ( d f M ) · d X f m d ML S ^ f M + X = f M · ( d · S · X ) S · ( f M X ) ( f M d ) ,
where Equation (17) and the definition : = U V S T = s p x ( d X ) f m d ML from Section 2 are used in the last step. This equation equals the previously deduced Equation (7), proving the equivalence of both models.

4.2. Light Field Decoding and SPC Calibration

This section revisits the popular decoding and calibration theme presented by Dansereau et al. [7]. In that work, the raw plenoptic camera image is first de-multiplexed into an integer-indexed two-plane parameterization L ( i , j , k , l ) . These indices are then transformed into metric rays and propagated through the main lens. The combination of these steps yields an intrinsics matrix
s t u v 1 = H 1 , 1 0 H 1 , 3 0 H 1 , 5 0 H 2 , 2 0 H 2 , 4 H 2 , 5 H 3 , 1 0 H 3 , 3 0 H 3 , 5 0 H 4 , 2 0 H 4 , 4 H 4 , 5 0 0 0 0 1 · i j k l 1
associating the integer indices directly with metric coordinates ( s , t , u , v ) for the scene-side light field. Note that these do not correspond to the equally named coordinates in Section 2 which describe the camera-side light field coordinates before propagating them through the main lens.
The relevant step with respect to the gap between the exit pupil and the principal plane in this process is the division of the integer indices by the respective spatial frequencies of the pixels and microlenses via the matrix H a b s θ . As explained in Section 2.1, the grid of MICs corresponds to the scaled grid of microlens centers. Accordingly, the sampling rate for the microlens plane has to be scaled down, or equivalently, the pixel sampling rate has to be scaled up by the inverse factor. Dansereau et al. [7] acknowledged this fact and chose the second option by introducing a scaling factor, which in our notation (compare Appendix B.2) corresponds to
M proj = 1 + f m d 1 .
This scaling, however, assumes a projection center at the main lens principal plane. Using the exit pupil instead, the correct rescaling is provided by
M proj = 1 + f m d X 1 ,
as visualized in Figure 8.
Fortunately, due to the overall formulation of the intrinsics as an end-to-end ray transformation, this slight change is not relevant for the calibration results, as neither f m nor d is directly estimated. Instead, the factor M proj contributes to the intrinsic variables H 1 , 1 to H 4 , 4 , and repeating the deduction of H with the correct scaling leads to the same general form for the intrinsics matrix as in Equation (23).
Similar cases of general parameters compensating for the model inaccuracies have been presented by Monteiro et al. [13] and Zhang et al. [12]. Both made use of the same decoding process as [7] and built upon the idea of directly relating the camera-side and scene light field. Monteiro et al. [13] slightly reduced the intrinsics matrix shown above, and subsequently used it to create an equivalent array of cameras for the scene-side light field. Zhang et al. [12] followed a similar approach by first relating the de-multiplexed light field in the form of sub-aperture images to the scene-side light field metric, which they then based all of their further calculations on. In both of these works, the interpretation of the intrinsic matrix parameters is irrelevant and they are not directly used to reconstruct the main lens properties; accordingly, these methods do not require any reformulation using a more complex main lens model.
Despite this, it is important to point out the inaccuracy in [7], as several related works only make use of the proposed decoding process and assume that a two-plane parameterization is received with a plane distance of d instead of d X , as shown by the examples in the following section.

4.3. Depth Reconstruction

One such case is presented by Pertuz et al. [10] and repeated in the follow-up work by Van Duong et al. [14]. In this work, a model relating the sub-aperture image shift to the object distance was deduced, which translates into our notation (compare Appendix B.3) as
o ( ρ ) = o f 1 a 0 · ρ 1 a 1 · ρ
with system-dependent parameters
a 0 = f m · d ML s p x · d and a 1 = o f · a 0 f M
and shift parameter
ρ = s p x · ( F F ) f m · d ML .
There are two problems in the deduction of this model. First, the shift parameter ρ is not correctly deduced under the premise of light field data decoded by the method of Dansereau et al. [7]; second, the exit pupil is ignored. In the following, a corrected version is presented, which additionally explains how these two problems nearly neutralize each other and lead to the same general model, albeit with different parameter interpretations.
First, the incorrect shift parameter ρ is examined. In [10], this parameter is described as the pixel disparity between neighboring sub-aperture images obtained via the decoding process from Dansereau et al. [7], and as such should correspond to our shift parameter S ˜ , which also ignores the exit pupil. However, due to different conventions, ρ is positive when focusing to a distance larger than the focus distance, whereas S ˜ is negative in that case (compare Equation (5) for α < 1 ). Accordingly, ρ should equal S ˜ ; however, transforming these parameters into a common notation leads to
S ˜ ( o ) = ˜ · o ( f M d ) + f M · d o · f M = s p x · d f m · d ML · o ( f M d ) + f M · d o · f M s p x · d f m · d ML · o ( f M d ) + f M · d d ( f M o ) = s p x · ( d o · f M o f M ) f m · d ML = s p x · ( F F ) f m · d ML = ρ .
The reason for this discrepancy is the implicit incorrect assumption in [10] that the grid of microlens image centers equals the grid of microlens centers, i.e., d ML = d MLI . While a light field in general could be reparametrized with this step size in the S T plane, this requires exact knowledge of the camera and MLA geometry, which is usually unknown and not considered in the decoding process of [7]. Using the correct shift parameter for light field data by following [7] instead and rearranging the corresponding Equation (10) results in
o ˜ ( S ˜ ) = f M · d · ˜ S ˜ · f M ˜ ( f M d ) = o f ( d f M ) · ˜ S ˜ · f M ˜ ( f M d ) = o f 1 1 + f M ˜ ( d f M ) · S ˜ = o f 1 1 o f · f m · d ML d 2 · s p x · ( S ˜ ) .
While this correction of Equation (26) continues to ignore the exit pupil, just as [10] does, it is considerably simpler than the original model, as only a single system-dependent parameter o f · f m · d ML d 2 · s p x is present instead of the two parameters a 0 and a 1 in Equation (26).
Finally, by introducing a non-zero distance X and thereby using o and S instead of o ˜ and S ˜ , the full model can be deduced from Equation (7) as
o ( S ) = f M · ( d · S · X ) S · ( f M X ) ( f M d ) = o f ( d f M ) · ( f M · d · f M · X · ( S ) ) ( f M · d ) · ( ( d f M ) ( X f M ) · S ) = o f f M · d · f M · X · S f M · d · ( X f M d f M · S ) = o f 1 X · d · S 1 X f M · ( d f M ) · S .
To align this model with the inverted shift direction of [10], we define
o f 1 X · d · S 1 X f M · ( d f M ) · S = o f 1 X · d · ( S ) 1 f M X · ( d f M ) · ( S ) = : o f 1 a 0 · ( S ) 1 a 1 · ( S ) .
This model has the same general form as Equation (26) proposed by Pertuz et al. [10], which explains the reasonable experimental results in that work. Nevertheless, the interpretation of the system-dependent parameters a 0 and a 1 as provided in [10] and repeated in [14] is incorrect, which is verified by experiment (V) in the following section. This different interpretation could lead to problems when the model needs to be fitted to data and the initial parameter values are based on the incorrect direct calculation.
Overall, the results of Pertuz et al. [10] are not entirely incorrect; however, the light field representation based on [7] simply does not match the implicit assumptions used for the model deduction. More specifically, the main problem of [10] is the definition of the shift parameter ρ for light field data, which is decoded similarly to [7] but is based on known microlens centers instead of MICs. While the light field could theoretically be reparametrized using the parallel projections of the microlens centers onto the sensor, this would require exact knowledge of the SPC intrinsics, namely, the parameters of the MLA and its placement relative to the main lens and sensor.

5. Evaluation

5.1. Simulation Environment

Because SPCs with exchangeable lenses were not commercially available at the time of writing and custom-built solutions are costly as well as prone to misalignment of the optical components, we resorted to synthetic experiments via extension of the ray tracing solution we provided in [40]. Our publicly available [16], updated version of the Blender [11] add-on expands the original simulation in the following aspects:
  • Simulation of aspherical lenses and zoom lenses.
  • Configurable MLA pose, thickness, and IOR.
  • Automatic focusing with lens group movement based on paraxial approximations.
  • Integration of Claff’s lens collection [47] and a collection of sensor presets.
  • Assisted plenoptic camera (SPC and FPC) configuration based on the ideas of [48].
This simulation setup facilitates quick generation of a broad range of plenoptic cameras, such as the example shown in Figure 9, and is used in the following to validate the formal analysis from Section 2.3, Section 3, and Section 4.3.

5.2. Experiments

For the validation, the five lenses listed in Table 1 were selected from the database [47]. While the first lens presents the ideal case of X 0 mm, i.e., with the exit pupil coinciding with the camera-side principal plane, the remaining four lenses present interesting cases with varying relationships between X and the focal lengths f M of the lenses.
Each of these lenses was used in two SPC configurations: one with a finite focus distance o f < and another focused at infinity o f = . To this end, the MLA placement with respect to the main lens, i.e., the distance d between the MLA and the camera-side principal plane of the main lens, was calculated using the thin lens Equation (1), as d = f M · o f o f f M for the finite case, and simply set to d = f M for o f = . The remaining microlens parameters were automatically initialized using the main lens and sensor properties so as to fulfill the following two constraints. First, the microlens f-number needs to match that of the main lens in order to optimally cover the sensor area [4]; second, a predefined number of 129 × 129 microlens images should be visible on the sensor in order to guarantee this resolution for the sub-aperture images. The resulting parameters were then fine-tuned by hand to accommodate the approximating nature of the f-number constraint and guarantee that the MICs coincided with the centers of the sensor pixels. This optimal MIC positioning has two effects. First, it renders the resampling during the decoding process of [7] unnecessary. After normalizing the raw images with white images to get rid of vignetting effects [50], the sub-aperture images in this setup can be directly extracted by combining the same relative sensor pixels from each microlens image [4]. Second, the evaluation can then concentrate on validating the refocusing itself instead of additionally dealing with compensating for interpolation artifacts from the decoding process. The full experimental setups can be found in Appendix D. Five experiments were performed with each of the ten following setups.
(I) MICs and Exit Pupil: The exit pupil as origin of the MICs is verified by first tracing ray bundles from the main lens aperture center through the main lens and MLA onto the sensor. This results in a set of sensor hits for every microlens. Due to the small variance within such a set, the mean is considered to represent the ground truth position of the microlens image center. In a second step, rays are traced from these sensor positions through the corresponding microlens centers and the convergence location of the resulting ray bundle is calculated in two ways: first by performing a line search along the optical axis for the minimum blur spot position of the ray bundle, and second based on the rays’ intersections with the optical axis. For these intersections, the mean and variance are calculated and presented alongside the minimum blur spot position. This whole process is visualized in Figure 10.
(II) o and S : A calibration pattern (more specifically, a Siemens star with four spokes) is placed at various distances in front of the camera. After demultiplexing the plenoptic camera image, the sub-aperture image shift S required to focus onto the given target distance o is measured. This is accomplished via line search, i.e., by repeatedly refocusing the image with a simple shift-and-sum algorithm [4] and calculating the sharpness of the refocused image. Here, the variance of the Laplacian [51] is used as the metric for image sharpness and the shift value with the highest score is considered the optimum. This procedure results in tuples of ground truth distances o and measured shifts S, which are then used to verify the connection S ( o ) as formally described in Equation (6).
To validate the inverted connection o ( S ) in Equation (7), all images are refocused for a given set of shift values. For each of these shift values, the object distance associated with the best focused image is considered the measured object distance for the respective shift value. The resulting tuples of preset shifts and measured distances are then used to verify o ( S ) .
(III) ERR o ˜ Validation: The data from experiment (II) are further used to verify the error model E R R o ˜ . In detail, the measured shift S for a known target distance o is used along with Equation (10) to approximate o ˜ ( S ) and calculate the measured relative error according to Equation (13). This error is then compared to the expected error obtained by directly calculating S based on the camera’s properties instead of measuring it.
(IV) ERR S ˜ Validation: In addition, the images of (II) can be used to verify the error model E R R S ˜ presented in Section 3. First, for every target distance o , the incorrect shift S ˜ ( o ) is calculated based on the assumption X = 0 as in Equation (9), and the images of the patterns at different positions are all refocused with this parameter. The target distance corresponding to the sharpest of the refocused images approximates o ( S ˜ ) , and is then used to measure E R R S ˜ as in Equation (15). Again, the measured values are compared to the errors obtained by directly calculating o ( S ˜ ) .
Note that instead of verifying E R R o ˜ and E R R S ˜ , Equations (9) and (10) and the shift error model in Equation (11) could be directly validated with the data measured and calculated in experiments (III) and (IV). However, these equations do not include a comparison between the incorrect estimations and the ground truth refocusing distances. Therefore, the indirect validation of those models by means of the resulting refocusing errors was preferred.
(V) Validation of Section 4.3: As analyzed in that section, the overall formulation of the model presented by Pertuz et al. [10] is correct; however, the model parameters have a different interpretation under the assumption of light field data decoded following the method from [7]. To verify the corrected model, the parameters a 0 and a 1 were first calculated based on the formula of Pertuz [10], then according to our model from Section 4.3, and finally fitted to the data of (II), i.e., the set of shift–distance pairs with measured shifts and ground truth target distances. For this parameter fitting, a grid search for the best parameters by means of the RMSE was performed with a grid explicitly containing both directly calculated parameter sets.

5.3. Results and Discussion

(I) As shown in Figure 11, the ray bundle consisting of rays running from the calculated ground truth MICs through the microlens centers generally converge towards the exit pupil in all setups. The closer the ray origins are to the optical axis (for example, compare the inner 25% of rays shown in Figure 11), the closer the minimum blur spot is located to the exit pupil. The deviations for larger sets of rays including the outermost MICs, as especially visible for the Canon setups, can be explained by the dependence of the exit pupil’s location on the viewing angle. Similar to the curved focal plane in lenses with a significant non-zero Petzval field curvature, the exit pupil cannot be well approximated by a plane in certain setups, and consequently affects the MIC grid on the sensor in a nonlinear fashion [42]. This is a clear limitation of our model, which is built upon paraxial approximations. Nevertheless, considering that the origin of the MICs is located close to the exit pupil plane even in extreme cases such as the outermost microlenses in the Canon setups, these results again confirm the observation of Hahne et al. [8], and justify the recommendation to examine the necessity of including the exit pupil in the lens model.
(II) The connection between the focus distance o and the shift S as described by Equation (6) is verified by the measurements presented in Figure 12. The mean of the absolute differences | S ( o ) m e a s u r e d S ( o ) G T | between the measured and directly calculated ground truth shift values across all ten setups is 0.008 px, with a variance of 0.0026 px. The worst single setup is the finitely focused Olympus setup, with a mean of absolute differences of 0.04 px and a variance of 0.025 px. These values are in the range of expected inaccuracies resulting from the image processing methods involved; in particular, the interpolation steps required by the shift-and-sum refocusing [4], as well as the rather simple (de)focus measure, which is prone to interference errors, are limiting factors that prevent higher accuracy.
A similar situation is observable in the inverse case, i.e., for the model o ( S ) from Equation (7), as presented in Figure 13. Due to the wide range of target distances, in this case the relative absolute differences | o ( S ) m e a s u r e d o ( S ) G T | o ( S ) G T are calculated to quantify the results from Figure 13. For the finite setups, the mean of these relative absolute differences is 0.13%, with a variance of 2 · 10 5 %, while in the infinite cases the overall mean is 0.72% with a variance of 0.01%. Here, the infinitely focused Zeiss setup has the worst performance, with a relative absolute difference mean of 0.9% and a variance of 0.004%. The worse performance of the infinitely focused setups is a consequence of the smaller range of shift values representing a larger refocusing range (compare Figure 13), resulting in greater susceptibility to small shift changes. Nevertheless, the overall performance again confirms the model o ( S ) within the constraints posed by the involved image processing steps.
(III) The results of the E R R o ˜ verification experiment are shown in Figure 14. The graphs again show the error based on measurements compared to the directly calculated ground truth error based on the error models presented in Section 3.
The mean of the absolute differences between the measured and calculated error values is 0.003, with a variance of 10 5 , which validates our models within the limitations of only the accuracy of the image processing methods and the optical properties of the chosen lenses. More specifically, the small deviations from the expected values, which are present even in the baseline case for the Rodenstock lens, can be explained as due to the interpolation operations required in Experiment (II) by the shift-and-sum refocusing algorithm [4] for non-integer shift values. The larger deviation visible at close range for the Olympus lens with finite focus distance is a result of the optical limits of this lens, which can be explained as follows. A single sub-aperture image contains one pixel per microlens; hence, such an image can be considered sharp if the calibration target is imaged onto the MLA with all blur spot sizes being smaller than the microlens diameter d ML . For the Olympus lens, however, these blur spot sizes increase more drastically at close range than those of the other lenses, leading to severe defocus blur even in the sub-aperture images. This can produce interference artifacts during the refocusing, leading in turn to fluctuating contrast measurements. Because the refocus distance o is determined by these measurements, this affects the error calculated by o ˜ ( S ( o ) ) o o in Equation (13).
(IV) Experiment (III) validated the error model in the case of a correctly estimated shift combined with an oversimplified distance estimator. Figure 15 shows that the inverse problem of an incorrectly calculated shift with the shift-and-sum refocusing is modeled correctly as well. The mean of absolute differences between the measured and directly calculated error values is 0.004, with a variance of 2.47 · 10 5 .
Overall, Experiments (III) and (IV) confirm the formal deductions in Section 3, and again justify our warning to mind the exit pupil when modeling a standard plenoptic camera. However, the results also hint at further minor optical or algorithmic aspects not being accounted for. In all cases, the mean of the absolute differences between the measured values and the ground truth model is up to two orders of magnitude larger than the respective variance. This is the result of nearly constant over- or underestimation, and could indicate a systematic error, which could be caused by, e.g., the refocusing model being limited to paraxial calculations.
(V) Regarding the correction of the model from [10] in Section 4.3, the results presented in Table 2 show that the corrected model appropriately describes the connection between the refocusing distance and the sub-aperture image shift. In all setups, the RMSE of our directly calculated model was within 1.75 mm of the RMSE of the fitted model. In addition, the fitted parameters are approximated well by the direct calculation with our model. On the other hand, the model in [10] can be regarded as incorrect for this light field parameterization, with one exception: the parameter a 1 in the setups focused at infinity is in general correct, which can be explained as follows. Neither the original model of [10] nor our correction directly considers the case o f = ; instead, a large focus distance of o f = 10 6   m is used as an approximation in both cases. For such a focus distance, the distance d between the main lens and the MLA is close to the focal distance of the main lens, i.e., d = f M + ϵ for some small value ϵ > 0 . With this formulation, the correctness of a 1 in the infinite cases can formally be explained by
a 1 , Pertuz a 1 , our = o f · a 0 , Pertuz f M · ( d f M ) f M X = o f · f m · d ML s p x · d f M · s p x · ( d X ) f m · d ML · ( d f M ) f M X = o f · d f M d f M = 1 / o f · d X f M X = f M + ϵ X f M X = 1 + ϵ f M X 1 .

6. Conclusion and Limitations

Overall, this work shows that the exit pupil can play a crucial role when modeling the relation between the camera-side and scene-side light fields. The connection between sub-aperture image shift and refocusing distance is derived analogously to previous work but with the additional consideration of an exit pupil that does not coincide with the principal plane of the main lens. Based on this deduction, two error models for the relative refocusing distance are created and validated. A subsequent review of previous work shows that a sufficiently general formulation of the SPC calibration model in most methods absorbs these errors, albeit leading to an incorrect interpretation of the model parameters. As an example, a correction of the work of Pertuz et al. [10] is presented and validated.
Nevertheless, despite the good evaluation results, there are several limitations to this work. First of all, the experiments were performed on simulated data. While the ray tracing-based lens simulation has been verified to exhibit the optical properties stated in the respective lens patents, i.e., aberration and distortion measurement results from the patents could be reproduced, there still is a gap between simulation and reality. On one hand, the specified lens parameters could differ from the final production lenses due to manufacturing inaccuracies or even deliberate parameter obfuscation by the lens manufacturer to hide specific lens details. On the other hand, the used framework provided by Blender [11] does not include wave optic effects such as diffraction [52,53]. Without these effects, the simulated optics are not diffraction-limited, and might produce images that are sharper than their real pendants.
Further limitations concern the formal lens model used for the error deduction. First, the microlenses in our model are still formally regarded as thin lenses. While Hahne et al. [9] used a thick lens model with explicit microlens principal planes, it was decided to leave this aspect out of our theoretical discussion for reasons of clarity and comprehensibility. However, the microlens thickness was factored in while performing the experiments, and Appendix C shows that an extended model does not change the equations deduced in Section 2 and Section 3.
Another limitation, which does affect the validity of these equations, is the restriction to paraxial models, more specifically the repeated use of the thin lens equation 1 in various calculations as well as fixed positions for the principal planes and exit pupils. The thin lens equation describes the relation between the object distance, the image distance, and the focal length of the lens, and is usually only valid along the optical axis. As the distance from this axis grows, third-order aberrations such as Petzval field curvature, i.e., a curved focus surface, can affect the refocusing distance [42]. Furthermore, as seen for the Canon lens evaluated in Experiment (I), the position of the exit pupil might vary depending on the viewing angle, leading to a reduction in the applicability of the deduced models.
Further work on the listed limitations is not expected to significantly alter the presented results, as these are already within the expected accuracy bounds set by the involved image processing steps. Instead, future work could target the second type of plenoptic cameras, namely, FPCs, for which a multitude of different calibration methods exists, and which also differ in their assumed lens models and could benefit from minding the gap between the principal plane and exit pupil.

Author Contributions

Conceptualization, T.M.; methodology, T.M.; software, T.M.; validation, T.M.; formal analysis, T.M. and D.M.; investigation, T.M. and D.M.; resources, T.M.; data curation, T.M.; writing—original draft preparation, T.M.; writing—review and editing, T.M., D.M. and R.K.; visualization, T.M.; supervision, R.K.; project administration, R.K. and T.M.; funding acquisition, R.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research is in parts funded through the Project "OP der Zukunft" within the funding program by Europäischen Fonds für Regionale Entwicklung (REACT-EU).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available in two publicly accessible repositories The data presented in this study are openly available at https://gitlab.com/ungetym/SPC-revisited, accessed on 22 February 2024, and the camera simulation is available at https://gitlab.com/ungetym/blender-camera-generator, accessed on 22 February 2024.

Acknowledgments

We would like to express our thanks to William ’Bill’ Claff, who supported us with his lens database [47].

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study, in the collection, analysis, or interpretation of data, in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
FPCFocused Plenoptic Camera
MIMicrolens Image
MICMicrolens Image Center
MLMicrolens
MLAMicrolens Array
SPCStandard Plenoptic Camera

Appendix A. Notation

Table A1. Camera Parameter notation.
Table A1. Camera Parameter notation.
ParameterDescription
f M Main lens focal length
H scene Scene-side principal plane of the main lens
H cam Camera-side principal plane of the main lens
XSigned distance between H cam and the exit pupil
measured along the optical axis
f m Microlens focal length
H scene ML Scene-side principal plane of a microlens
H cam ML Camera-side principal plane of a microlens
o Object distance measured from H scene
i Image distance measured from H cam
o f Focus distance
dDistance between the MLA’s H scene ML
and the main lens H cam
d ML Microlens pitch
s p x Pixel pitch, i.e., edge length of a square sensor pixel
Table A2. Light field parametrization notation.
Table A2. Light field parametrization notation.
ParameterDescription
FDistance between U V and S T plane
U V Step size in the virtual lens plane
S T Pixel pitch on the virtual sensor
Scaling between the U V and S T step sizes, i.e., = U V / S T
L F Integer indexed 4D light field
i , j , k , l Integer light field coordinates
L ˜ F Metric parameterization of the 4D light field
s , t , u , v Metric light field coordinates
Table A3. Refocusing and error analysis notation.
Table A3. Refocusing and error analysis notation.
ParameterDescription
F Distance between U V and S T plane for refocusing
α Refocusing parameter defined by α = F / F
L ˜ F L ˜ F , but for S T plane
SSub-aperture image shift in pixels
S ( o ) S for a given refocusing distance o
o ( S ) Refocusing distance o for a given shift S
˜ Quotient based on the assumption X = 0
S ˜ Sub-aperture image shift in pixels based on the assumption X = 0
o ˜ Refocusing distance based on the assumption X = 0
λ Quotient o / o f
E R R o ˜ Relative refocusing distance error
for incorrect target distance model
E R R S ˜ Relative refocusing distance error
for incorrect shift model

Appendix B. Revisited Literature—Notation Transfer

In the following sections, the notations of [7,9,10] are transferred into the notation used in this work.

Appendix B.1. Hahne et al. [9]

Table A4. General notation in Hahne et al. [9] and our work.
Table A4. General notation in Hahne et al. [9] and our work.
DescriptionHahne et al. [9]Our
Main lens focal length f U f M
Distance H cam to MLA b U d
Distance H cam to virtual image b U i
Distance H scene to scene object a U o
Distance MLA to exit pupil d A d X
Distance MLA to virtual image d a d i
ML focal length f s f m
ML diameter s d ML
Pixel size u s p x
In addition to the general camera properties listed in Table A4, Hahne et al. [9] introduced the notation in Table A5 to exemplarily describe the camera-side rays for a horizontal cross-section of the plenoptic camera model.
Table A5. Further notation in Hahne et al. [9] without direct equivalents in our setup.
Table A5. Further notation in Hahne et al. [9] without direct equivalents in our setup.
DescriptionHahne et al. [9]
Horizontal number of microlensesJ
Microlens indexj
Microlens center of lens j s j = j J 1 2 · s
Microlens image center of lens j u c , j = s j d A · f s + s j
Position of i-th neighbor pixel of MIC u c + i , j = u c , j + i · u
Slope of ray from i-th
neighbor through ML center
m c + i , j = s j u c + i , j f s
With these, the linear ray function originating in a pixel at the sensor position u c + i , j and running through the corresponding microlens center s j can be described via
f c + i , j ( z ) = m c + i , j · z + s j = s j u c + i , j f s · z + s j = s j ( u c , j + i · u ) f s · z + s j = s j s j d A · f s + s j + i · u f s · z + s j .
By assuming an MLA with a microlens center located at the main lens optical axis, the ray originating from the center of the central microlens image going through the microlens center s 0 = 0 can be described by
f ( z ) : = f c , 0 ( z ) = s 0 s 0 d X · f m + s 0 + 0 · s p x f m · z + s 0 = 0 .
Given a second pixel position at distance d MLI + S ^ from the central pixel, this pixel’s index in the microlens image j = 1 is provided by i = S ^ s p x . Accordingly, the ray originating from this pixel running through the microlens center s 1 = d ML is provided by
f ˜ ( z ) : = f c + i , 1 ( z ) = s 1 s 1 d X · f m + s 1 + S ^ s p x · s p x f m · z + s 1 = d ML d ML d X · f m + d ML + S ^ f m · z + d ML = d ML d MLI + S ^ f m · z + d ML .

Appendix B.2. Dansereau et al. [7]

Table A6. General notation in Danserau et al. [7] and our work.
Table A6. General notation in Danserau et al. [7] and our work.
DescriptionDansereau et al. [7]Our
Distance H cam to MLA d M d
Distance sensor to MLA
(equals ML focal length for SPCs)
d μ f m
ML focal length f μ f m
Apart from the intrinsic parameters listed in Table A6, many other aspects of the notation in Danserau et al. [7] are dedicated to the light field outside the camera, which is not directly modeled in our work.

Appendix B.3. Pertuz et al. [10]

Using the notation provided in Table A7, the deduction steps for the model in [10] can be translated as stated in Table A8.
Table A7. General notation in Pertuz et al. [10] and our work.
Table A7. General notation in Pertuz et al. [10] and our work.
DescriptionPertuz et al. [10]Our
Main lens focal lengthf f M
Focus distance (measured from H scene ) z 0 o f
Corresponding image distance
(equals distance MLA to H cam )
x 0 d
Target distancez o
Corresponding image distancex i
Distance MLA to sensor (equals ML focal length) β f m
ML diameterD d ML
Pixel size μ s p x
Separation between real and synthetic focal plane x F F
Refocusing parameter (sub-aperture image shift) ρ S ˜
UV plane step size u U V
However, note that Table A8 presents a direct notation transfer, especially regarding Equations (5) and (8) of [10]. As shown in Section 4, the shift parameter ρ is already based on an incorrect assumption, and as such does not directly correspond to S ˜ .
Table A8. Crucial equations from Pertuz et al. [10] and the corresponding notation in our work.
Table A8. Crucial equations from Pertuz et al. [10] and the corresponding notation in our work.
Equation Number in [10]Pertuz et al. [10]Our Notation
(1) 1 f = 1 x 0 + 1 z 0 1 f M = 1 d + 1 o f
(4) u = μ β D x u U V = s p x f m d ML ( F F ) · u
(5) ρ = μ β D x S ˜ = s p x f m d ML ( F F )
(7) z = f ( x 0 x ) x 0 x f o = f M ( d ( F F ) ) d ( F F ) f M
(8) z = z 0 1 a 0 ρ 1 a 1 ρ o = o f 1 a 0 ( S ˜ ) 1 a 1 ( S ˜ )
Addition to (8) a 0 = β D μ x 0 a 0 = f m · d ML s p x · d
Addition to (8) a 1 = z 0 a 0 / f a 1 = o f · a 0 f M

Appendix C. Thick Microlenses

In this appendix, the thin microlenses are extended by a model featuring separate principal planes H cam ML and H scene ML , i.e., a non-zero thickness. The separation between the main lens camera-side principal plane H cam and the MLA d is then provided by the distance between H cam and H scene ML . On the other side, the distance between the MLA and the sensor, which equals the focal length of the microlenses f m , now describes the distance between H cam ML and the sensor plane, as shown in Figure A1. Because a ray entering H scene ML at a certain height leaves the microlens at the same height at H cam ML , the deductions from Section 2 and Section 3 remain valid.
Figure A1. Extension of the plenoptic camera model shown in Figure 2 to a non-zero microlens thickness.
Figure A1. Extension of the plenoptic camera model shown in Figure 2 to a non-zero microlens thickness.
Sensors 24 02522 g0a1

Appendix D. Evaluation Setups

The following Table A9, Table A10 and Table A11 list the specific lens models along with their associated patents and SPC configurations. Note that both H cam and the exit pupil location are measured from the main lens aperture position, and have a positive sign indicating that these planes are located on the sensor side of the aperture. On the other hand, H scene is measured from the aperture, with a positive sign if located on the scene-side of the aperture. X is measured from the position of H cam , as shown in Figure 2, and all remaining properties are assumed to be unsigned distance values. All mentioned properties were measured using a 2D ray tracer, which we also integrated into the Blender add-on to facilitate fast automatic plenoptic camera reconfigurations.
Table A9. Model names of the simulated main lenses and the associated patents specifying the lens element parameters. DE, JP, and US specify the country of the patent application [47].
Table A9. Model names of the simulated main lenses and the associated patents specifying the lens element parameters. DE, JP, and US specify the country of the patent application [47].
ManufacturerModelPatent
RodenstockSironar-N 100 mm F5.6DE 2729831 Example 1
ZeissBatis 85 mm F1.8JP 2015-096915 Example 2
Ricohsmc Pentax-A 200 mm F4 Macro EDUS 4,666,260 Example 1
CanonEF 85 mm F1.8 USMJP 1993-157964 Example 1
OlympusZuiko Auto-Zoom 85–250 mm F5US 4,025,167 Example 2
Table A10. SPC properties for the evaluation setups with finite focus distance.
Table A10. SPC properties for the evaluation setups with finite focus distance.
Finite Focus
PropertyRodenstockZeissRicohCanonOlympus
Focal Length f M (mm)99.99882.047167.99484.99885.120
H cam (mm)−1.395−30.395−118.018−2.30354.527
H scene (mm)−0.98136.19283.127−27.3171.026
Exit Pupil Loc. (mm)−1.20110.257−18.109−31.241−5.690
Exit Pupil Radius (mm)8.6172.4233.8226.5122.917
X (mm)0.19440.65299.909−28.938−60.219
f-number5.6008.5008.0007.3306.590
Focus Distance F (mm)500500500300300
ML Pitch ( μ m)178.158173.703176.246177.856110.567
ML Diam ( μ m)178.158173.703176.285177.856110.567
ML Focal Length f m (mm)1.2902.0843.2611.7790.884
MLA Thickness (mm)0.1000.1000.1000.1000.100
MLA-Sensor Dist (mm)1.2802.0843.2611.7790.874
Aperture-Sensor Dist (mm)124.70269.944138.341118.176174.140
Sensor Width (mm)23.22023.22023.22023.2207.222
Sensor Height (mm)23.22023.22023.22023.2207.222
Table A11. SPC properties for the evaluation setups with infinite focus distance.
Table A11. SPC properties for the evaluation setups with infinite focus distance.
Infinite Focus
PropertyRodenstockZeissRicohCanonOlympus
Focal Length f M (mm)99.99882.860173.11584.99885.004
H cam (mm)−1.395−29.315−109.963−2.30354.617
H scene (mm)−0.98134.02375.005−27.3171.155
Exit Pupil Loc. (mm)−1.20110.257−18.109−31.241−5.690
Exit Pupil Radius (mm)8.6173.8177.0936.5129.290
X (mm)0.19439.57291.525−28.938−60.308
f-number5.6005.8709.1008.0007.660
Focus Distance F (mm)
ML Pitch ( μ m)178.158175.944177.164177.840171.492
ML Diam ( μ m)178.158175.944177.164177.840171.492
ML Focal Length f m (mm)1.0320.9981.2301.3831.297
MLA Thickness (mm)0.1000.1000.1000.1000.100
MLA-Sensor Dist (mm)1.0220.9871.2201.3731.287
Aperture-Sensor Dist (mm)99.74554.65964.52184.189141.035
Sensor Width (mm)23.22023.22023.20023.22022.32
Sensor Height (mm)23.22023.22023.20023.22022.32

References

  1. Lippmann, G. La photographie integrale. Comptes-Rendus 1908, 146, 446–451. [Google Scholar]
  2. Ives, H.E. A camera for making parallax panoramagrams. Josa 1928, 17, 435–439. [Google Scholar] [CrossRef]
  3. Adelson, E.H.; Wang, J.Y. Single lens stereo with a plenoptic camera. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 99–106. [Google Scholar] [CrossRef]
  4. Ng, R.; Levoy, M.; Brédif, M.; Duval, G.; Horowitz, M.; Hanrahan, P. Light Field Photography with a Hand-Held Plenoptic Camera. Ph.D. Thesis, Stanford University, Stanford, CA, USA, 2005. [Google Scholar]
  5. Georgiev, T.; Intwala, C. Light field camera design for integral view photography. In Technical Report; Adobe System, Inc.: San Jose, CA, USA, 2006; p. 1. [Google Scholar]
  6. Perwass, C.; Wietzke, L. Single lens 3D-camera with extended depth-of-field. In Proceedings of the Human Vision and Electronic Imaging XVII, SPIE, Burlingame, CA, USA, 22 January 2012; Volume 8291, pp. 45–59. [Google Scholar]
  7. Dansereau, D.G.; Pizarro, O.; Williams, S.B. Decoding, calibration and rectification for lenselet-based plenoptic cameras. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; pp. 1027–1034. [Google Scholar]
  8. Hahne, C.; Aggoun, A.; Velisavljevic, V.; Fiebig, S.; Pesch, M. Baseline and triangulation geometry in a standard plenoptic camera. Int. J. Comput. Vis. 2018, 126, 21–35. [Google Scholar] [CrossRef]
  9. Hahne, C.; Aggoun, A.; Velisavljevic, V.; Fiebig, S.; Pesch, M. Refocusing distance of a standard plenoptic camera. Opt. Express 2016, 24, 21521–21540. [Google Scholar] [CrossRef]
  10. Pertuz, S.; Pulido-Herrera, E.; Kamarainen, J.K. Focus model for metric depth estimation in standard plenoptic cameras. ISPRS J. Photogramm. Remote Sens. 2018, 144, 38–47. [Google Scholar] [CrossRef]
  11. Blender Foundation. Available online: https://www.blender.org/ (accessed on 12 January 2024).
  12. Zhang, Q.; Zhang, C.; Ling, J.; Wang, Q.; Yu, J. A generic multi-projection-center model and calibration method for light field cameras. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2539–2552. [Google Scholar] [CrossRef]
  13. Monteiro, N.B.; Barreto, J.P.; Gaspar, J.A. Standard plenoptic cameras mapping to camera arrays and calibration based on DLT. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 4090–4099. [Google Scholar] [CrossRef]
  14. Van Duong, V.; Canh, T.N.; Huu, T.N.; Jeon, B. Focal stack based light field coding for refocusing applications. In Proceedings of the 2019 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB), Jeju, Republic of Korea, 5–7 June 2019; pp. 1–4. [Google Scholar]
  15. Michels, T. SPC Designs. Available online: https://gitlab.com/ungetym/SPC-revisited (accessed on 22 February 2024).
  16. Michels, T. Blender Camera Generator Add-On. Available online: https://gitlab.com/ungetym/blender-camera-generator (accessed on 22 February 2024).
  17. Lumsdaine, A.; Georgiev, T. The focused plenoptic camera. In Proceedings of the 2009 IEEE International Conference on Computational Photography (ICCP), San Francisco, CA, USA, 16–17 April 2009; pp. 1–8. [Google Scholar]
  18. Wanner, S.; Goldluecke, B. Globally consistent depth labeling of 4D light fields. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 41–48. [Google Scholar]
  19. Georgiev, T.; Lumsdaine, A. Focused plenoptic camera and rendering. J. Electron. Imaging 2010, 19, 021106. [Google Scholar]
  20. Bok, Y.; Jeon, H.G.; Kweon, I.S. Geometric calibration of micro-lens-based light field cameras using line features. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 287–300. [Google Scholar] [CrossRef]
  21. Zhao, Y.; Li, H.; Mei, D.; Shi, S. Metric calibration of unfocused plenoptic cameras for three-dimensional shape measurement. Opt. Eng. 2020, 59, 073104. [Google Scholar] [CrossRef]
  22. Thomason, C.M.; Thurow, B.S.; Fahringer, T.W. Calibration of a microlens array for a plenoptic camera. In Proceedings of the 52nd Aerospace Sciences Meeting, National Harbor, MD, USA, 13–17 January 2014; p. 396. [Google Scholar]
  23. Suliga, P.; Wrona, T. Microlens array calibration method for a light field camera. In Proceedings of the 2018 19th International Carpathian Control Conference (ICCC), Szilvasvarad, Hungary, 28–31 May 2018; pp. 19–22. [Google Scholar]
  24. Schambach, M.; León, F.P. Microlens array grid estimation, light field decoding, and calibration. IEEE Trans. Comput. Imaging 2020, 6, 591–603. [Google Scholar] [CrossRef]
  25. Mignard-Debise, L.; Ihrke, I. A vignetting model for light field cameras with an application to light field microscopy. IEEE Trans. Comput. Imaging 2019, 5, 585–595. [Google Scholar] [CrossRef]
  26. Johannsen, O.; Heinze, C.; Goldluecke, B.; Perwaß, C. On the calibration of focused plenoptic cameras. In Time-of-Flight and Depth Imaging. Sensors, Algorithms, and Applications: Dagstuhl 2012 Seminar on Time-of-Flight Imaging and GCPR 2013 Workshop on Imaging New Modalities; Springer: Berlin/Heidelberg, Germany, 2013; pp. 302–317. [Google Scholar]
  27. Heinze, C.; Spyropoulos, S.; Hussmann, S.; Perwaß, C. Automated robust metric calibration algorithm for multifocus plenoptic cameras. IEEE Trans. Instrum. Meas. 2016, 65, 1197–1205. [Google Scholar] [CrossRef]
  28. Zeller, N.; Noury, C.; Quint, F.; Teulière, C.; Stilla, U.; Dhome, M. Metric calibration of a focused plenoptic camera based on a 3D calibration target. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 3, 449–456. [Google Scholar] [CrossRef]
  29. Noury, C.A.; Teulière, C.; Dhome, M. Light-field camera calibration from raw images. In Proceedings of the 2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Sydney, Australia, 29 November–1 December 2017; pp. 1–8. [Google Scholar]
  30. Nousias, S.; Chadebecq, F.; Pichat, J.; Keane, P.; Ourselin, S.; Bergeles, C. Corner-based geometric calibration of multi-focus plenoptic cameras. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 957–965. [Google Scholar]
  31. Wang, Y.; Qiu, J.; Liu, C.; He, D.; Kang, X.; Li, J.; Shi, L. Virtual image points based geometrical parameters’ calibration for focused light field camera. IEEE Access 2018, 6, 71317–71326. [Google Scholar] [CrossRef]
  32. Labussière, M.; Teulière, C.; Bernardin, F.; Ait-Aider, O. Leveraging blur information for plenoptic camera calibration. Int. J. Comput. Vis. 2022, 130, 1655–1677. [Google Scholar] [CrossRef]
  33. Kolb, C.; Mitchell, D.; Hanrahan, P. A realistic camera model for computer graphics. In Proceedings of the 22nd Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 6–11 August 1995; pp. 317–324. [Google Scholar]
  34. Wu, J.; Zheng, C.; Hu, X.; Li, C. An accurate and practical camera lens model for rendering realistic lens effects. In Proceedings of the 2011 12th International Conference on Computer-Aided Design and Computer Graphics, Jinan, China, 15–17 September 2011; pp. 63–70. [Google Scholar]
  35. Zheng, Q.; Zheng, C. NeuroLens: Data-Driven Camera Lens Simulation Using Neural Networks. In Proceedings of the Computer Graphics Forum, Wiley Online Library, Lyon, France, 24–26 April 2017; Volume 36, pp. 390–401. [Google Scholar]
  36. Fleischmann, O.; Koch, R. Lens-based depth estimation for multi-focus plenoptic cameras. In Proceedings of the German Conference on Pattern Recognition. Springer, Münster, Germany, 2–5 September 2014; pp. 410–420. [Google Scholar]
  37. Zhang, R.; Liu, P.; Liu, D.; Su, G. Reconstruction of refocusing and all-in-focus images based on forward simulation model of plenoptic camera. Opt. Commun. 2015, 357, 1–6. [Google Scholar] [CrossRef]
  38. Liang, C.K.; Ramamoorthi, R. A light transport framework for lenslet light field cameras. ACM Trans. Graph. (TOG) 2015, 34, 16. [Google Scholar] [CrossRef]
  39. Nürnberg, T.; Schambach, M.; Uhlig, D.; Heizmann, M.; León, F.P. A simulation framework for the design and evaluation of computational cameras. In Proceedings of the Automated Visual Inspection and Machine Vision III, International Society for Optics and Photonics, Munich, Germany, 24–27 June 2019; Volume 11061, p. 1106102. [Google Scholar]
  40. Michels, T.; Petersen, A.; Palmieri, L.; Koch, R. Simulation of plenoptic cameras. In Proceedings of the 2018-3DTV-Conference: The True Vision-Capture, Transmission and Display of 3D Video (3DTV-CON), Helsinki, Finland, 3–5 June 2018; pp. 1–4. [Google Scholar]
  41. Born, M.; Wolf, E. Principles of Optics: Electromagnetic Theory of Propagation, Interference and Diffraction of Light; Elsevier: Amsterdam, The Netherlands, 2013. [Google Scholar]
  42. Smith, W.J. Modern Optical Engineering: The Design of Optical Systems; McGraw-Hill Education: New York, NY, USA, 2008. [Google Scholar]
  43. Levoy, M.; Hanrahan, P. Light field rendering. In Computer Graphics Proceedings, Annual Conference Series; Association for Computing Machinery SIGGRAPH: New Orleans, LA, USA, 1996; pp. 31–42. [Google Scholar]
  44. Ng, R. Fourier slice photography. In ACM Siggraph 2005 Papers; Association for Computing Machinery: New York, NY, USA, 2005; pp. 735–744. [Google Scholar]
  45. Carl Zeiss AG, Datasheet: Zeiss Planar T* 2/80 Lens Datasheet. Available online: https://www.zeiss.de/content/dam/consumer-products/downloads/historical-products/photography/contax-645/de/datasheet-zeiss-planar-280-de.pdf (accessed on 12 January 2024).
  46. Carl Zeiss AG, Datasheet: Zeiss Zeiss Sonnar T* 2.8/140. Available online: https://www.zeiss.de/content/dam/consumer-products/downloads/historical-products/photography/contax-645/de/datasheet-zeiss-sonnar-28140-de.pdf (accessed on 12 January 2024).
  47. Claff, W.J. Available online: https://www.photonstophotos.net (accessed on 12 January 2024).
  48. Michels, T.; Koch, R. Ray tracing-guided design of plenoptic cameras. In Proceedings of the 2021 International Conference on 3D Vision (3DV), London, UK, 1–3 December 2021; pp. 1125–1133. [Google Scholar]
  49. ISO 12233 Pattern, Cornell University. Available online: https://www.graphics.cornell.edu/~westin/misc/res-chart.html (accessed on 1 April 2024).
  50. Yu, W. Practical anti-vignetting methods for digital cameras. IEEE Trans. Consum. Electron. 2004, 50, 975–983. [Google Scholar]
  51. Pertuz, S.; Puig, D.; Garcia, M.A. Analysis of focus measure operators for shape-from-focus. Pattern Recognit. 2013, 46, 1415–1432. [Google Scholar] [CrossRef]
  52. Freniere, E.R.; Gregory, G.G.; Hassler, R.A. Edge diffraction in Monte Carlo ray tracing. In Proceedings of the Optical Design and Analysis Software, SPIE, Denver, CO, USA, 21–22 July 1999; Volume 3780, pp. 151–157. [Google Scholar]
  53. Mahan, J.; Vinh, N.; Ho, V.; Munir, N. Monte Carlo ray-trace diffraction based on the Huygens–Fresnel principle. Appl. Opt. 2018, 57, D56–D62. [Google Scholar] [CrossRef] [PubMed]
Figure 2. Plenoptic camera modeled by a thick main lens combined with a thin-lens MLA. The pitch of the microlens is described by d ML , the distance between the neighboring microlens image centers (MICs) is denoted as d MLI , X describes the distance between the exit pupil and the camera-side principal plane, and d is the distance between H cam and the MLA. A complete notation overview is provided in Appendix A.
Figure 2. Plenoptic camera modeled by a thick main lens combined with a thin-lens MLA. The pitch of the microlens is described by d ML , the distance between the neighboring microlens image centers (MICs) is denoted as d MLI , X describes the distance between the exit pupil and the camera-side principal plane, and d is the distance between H cam and the MLA. A complete notation overview is provided in Appendix A.
Sensors 24 02522 g002
Figure 3. Integer (red) and metric (black) two-plane parameterization of the light field. Here, s p x describes the size of a sensor pixel and f m the focal length of a microlens, which for an SPC coincides with the distance between the MLA and sensor.
Figure 3. Integer (red) and metric (black) two-plane parameterization of the light field. Here, s p x describes the size of a sensor pixel and f m the focal length of a microlens, which for an SPC coincides with the distance between the MLA and sensor.
Sensors 24 02522 g003
Figure 4. Light field refocusing via shifting of the virtual sensor, i.e., the S T -plane is moved to the image distance i . A ray ( s , u ) can be associated with a ray ( s , u ) by means of the triangle equality, i.e., s = u + F F ( s u ) .
Figure 4. Light field refocusing via shifting of the virtual sensor, i.e., the S T -plane is moved to the image distance i . A ray ( s , u ) can be associated with a ray ( s , u ) by means of the triangle equality, i.e., s = u + F F ( s u ) .
Sensors 24 02522 g004
Figure 5. Left: Relative shift error based on λ = o o f . Middle/Right: The two cases of relative object distance errors for an assumed camera focused at a finite distance which is met for a relative distance of o o f = 1 . Negative error values indicate an underestimation of the ground truth value, while positive errors represent an overestimation.
Figure 5. Left: Relative shift error based on λ = o o f . Middle/Right: The two cases of relative object distance errors for an assumed camera focused at a finite distance which is met for a relative distance of o o f = 1 . Negative error values indicate an underestimation of the ground truth value, while positive errors represent an overestimation.
Sensors 24 02522 g005
Figure 6. Distances X between the exit pupil and principal plane H cam for 866 lenses [47] sorted by focal length. The black line represents the linear model fitted to the data, while the colors of the datapoints indicate the relationship X / f M . Note that the horizontal axis uses logarithmic scaling due to the large number of lenses with a focal length below 100 mm.
Figure 6. Distances X between the exit pupil and principal plane H cam for 866 lenses [47] sorted by focal length. The black line represents the linear model fitted to the data, while the colors of the datapoints indicate the relationship X / f M . Note that the horizontal axis uses logarithmic scaling due to the large number of lenses with a focal length below 100 mm.
Sensors 24 02522 g006
Figure 7. Image formation (light blue) for an object point located at distance o from H scene . The image of this point, located at distance i from H cam , is seen by multiple microlenses, and its projections onto the sensor have a metric disparity of S ^ . In order to determine the object distance o , Hahne et al. [9] proposed using intersecting ray functions (red) from two of the images and transferring the resulting image distance i to the scene via the thin lens equation.
Figure 7. Image formation (light blue) for an object point located at distance o from H scene . The image of this point, located at distance i from H cam , is seen by multiple microlenses, and its projections onto the sensor have a metric disparity of S ^ . In order to determine the object distance o , Hahne et al. [9] proposed using intersecting ray functions (red) from two of the images and transferring the resulting image distance i to the scene via the thin lens equation.
Sensors 24 02522 g007
Figure 8. The central sub-aperture image consists of the MICs, i.e., the image of the aperture center viewed across all microlenses. These MICs originate from the center of the UV plane, i.e., the exit pupil; accordingly, the distance between neighboring MICs is provided by the triangle equality via d MLI = S T · ( f m + F ) / F = S T · ( 1 + f m / ( d X ) ) .
Figure 8. The central sub-aperture image consists of the MICs, i.e., the image of the aperture center viewed across all microlenses. These MICs originate from the center of the UV plane, i.e., the exit pupil; accordingly, the distance between neighboring MICs is provided by the triangle equality via d MLI = S T · ( f m + F ) / F = S T · ( 1 + f m / ( d X ) ) .
Sensors 24 02522 g008
Figure 9. Cross-section and rendering of an example evaluation setup. A fully modeled lens was combined with a two-plane MLA model and a sensor in order simulate a plenoptic camera via ray tracing. Calibration patterns were placed at different distances in front of this setup to verify the analytical models. Note that the housing of the camera and lens were removed for the purpose of visualization. The ISO 12233 pattern [49] is used with the permission of Cornell University.
Figure 9. Cross-section and rendering of an example evaluation setup. A fully modeled lens was combined with a two-plane MLA model and a sensor in order simulate a plenoptic camera via ray tracing. Calibration patterns were placed at different distances in front of this setup to verify the analytical models. Note that the housing of the camera and lens were removed for the purpose of visualization. The ISO 12233 pattern [49] is used with the permission of Cornell University.
Sensors 24 02522 g009
Figure 10. Experiment (I): The two steps of the MIC/exit pupil verification visualized for the Zeiss Batis 85mm F1.8. Rays (light gray) are traced from the main lens aperture center through the main lens and MLA onto the sensor. The resulting means of the sensor hits per microlens represent the MICs. The exit pupil as the approximate source of these points is verified by backwards tracing of the rays (blue/green) from the MICs through the respective microlenses and calculation of the minimum blur spot position along with the mean and variance of intersections along the optical axis.
Figure 10. Experiment (I): The two steps of the MIC/exit pupil verification visualized for the Zeiss Batis 85mm F1.8. Rays (light gray) are traced from the main lens aperture center through the main lens and MLA onto the sensor. The resulting means of the sensor hits per microlens represent the MICs. The exit pupil as the approximate source of these points is verified by backwards tracing of the rays (blue/green) from the MICs through the respective microlenses and calculation of the minimum blur spot position along with the mean and variance of intersections along the optical axis.
Sensors 24 02522 g010
Figure 11. Results of Experiment (I). The orange markers indicate the locations of the exit pupils on the optical axis (horizontal, dotted line) with respect to the respective principal planes ( H cam ). The black markers show the mean and variance of the intersection points between the optical axis and the rays traced back from the MICs through the microlens centers. The colored functions show the blur spot sizes for different subsets of these ray bundles close to the exit pupil (compare Figure 10). These subsets contain the respective portion of rays from the bundle which are closest to the optical axis.
Figure 11. Results of Experiment (I). The orange markers indicate the locations of the exit pupils on the optical axis (horizontal, dotted line) with respect to the respective principal planes ( H cam ). The black markers show the mean and variance of the intersection points between the optical axis and the rays traced back from the MICs through the microlens centers. The colored functions show the blur spot sizes for different subsets of these ray bundles close to the exit pupil (compare Figure 10). These subsets contain the respective portion of rays from the bundle which are closest to the optical axis.
Sensors 24 02522 g011
Figure 12. Results of Experiment (II): Verification of the model S ( o ) according to Equation (6). The datapoints represent measured shift values for the respective object distances, while the underlying lines represent the expected directly calculated values. Left: The error for the setups with finite focus based on the relative object distance o o f . Right: The error for setups focused at infinity.
Figure 12. Results of Experiment (II): Verification of the model S ( o ) according to Equation (6). The datapoints represent measured shift values for the respective object distances, while the underlying lines represent the expected directly calculated values. Left: The error for the setups with finite focus based on the relative object distance o o f . Right: The error for setups focused at infinity.
Sensors 24 02522 g012
Figure 13. Results of Experiment (II): Verification of the model o ( S ) according to Equation (7). The datapoints represent measured focus distance values for the respective shifts and the underlying lines represent the expected directly calculated values.
Figure 13. Results of Experiment (II): Verification of the model o ( S ) according to Equation (7). The datapoints represent measured focus distance values for the respective shifts and the underlying lines represent the expected directly calculated values.
Sensors 24 02522 g013
Figure 14. Results of Experiment (III): Relative object distance error resulting from a correctly estimated shift with incorrect object distance estimation based on the assumption that X = 0 . The thick lines indicate the predicted ground truth error based on Equations (13) and (14), with the points indicating the measurements. Left: The error for the setups with finite focus. To provide a comparative visualization of the different target distance ranges, the results are shown based on the relative object distance λ = o o f . Right: The error for the setups focused at infinity.
Figure 14. Results of Experiment (III): Relative object distance error resulting from a correctly estimated shift with incorrect object distance estimation based on the assumption that X = 0 . The thick lines indicate the predicted ground truth error based on Equations (13) and (14), with the points indicating the measurements. Left: The error for the setups with finite focus. To provide a comparative visualization of the different target distance ranges, the results are shown based on the relative object distance λ = o o f . Right: The error for the setups focused at infinity.
Sensors 24 02522 g014
Figure 15. Results of Experiment (IV): Relative object distance error resulting from an incorrectly calculated shift ignoring the exit pupil in combination with a correct object distance estimation. The thick lines indicate the predicted ground truth error based on Equations (15) and (16), with the points indicating the measurements. Left: The error for the setups with finite focus. To provide a comparative visualization, the results are shown based on the relative object distance o o f , i.e., the finite focus distance for the respective setups is met at o o f = 1 . Right: The error for the setups focused at infinity.
Figure 15. Results of Experiment (IV): Relative object distance error resulting from an incorrectly calculated shift ignoring the exit pupil in combination with a correct object distance estimation. The thick lines indicate the predicted ground truth error based on Equations (15) and (16), with the points indicating the measurements. Left: The error for the setups with finite focus. To provide a comparative visualization, the results are shown based on the relative object distance o o f , i.e., the finite focus distance for the respective setups is met at o o f = 1 . Right: The error for the setups focused at infinity.
Sensors 24 02522 g015
Table 1. Overview of the simulated main lenses and their properties in the finite and infinite focus setup. Note that the focal length of a lens can vary in different setups due to the lens group movements involved in refocus or zoom operations.
Table 1. Overview of the simulated main lenses and their properties in the finite and infinite focus setup. Note that the focal length of a lens can vary in different setups due to the lens group movements involved in refocus or zoom operations.
Finite FocusInfinite Focus
Lens Model f M (mm) X (mm) X f M Focus
Dist. (mm)
f M (mm) X (mm) X f M
Rodenstock
Sironar-N 100 mm F5.6
99.9980.1940.002500.099.9980.1940.002
Zeiss
Batis 85 mm F1.8
82.04740.6520.496500.082.86039.5730.478
Ricoh smc Pentax-A
200 mm F4 Macro ED
167.99499.9080.595500.0173.11591.8540.531
Canon
EF 85 mm F1.8 USM
84.998−28.938−0.341300.084.998−28.938−0.341
Olympus Zuiko
Auto-Zoom 85–250 mm F5
85.120−60.219−0.708300.085.004−60.308−0.709
Table 2. Results of Experiment (V): Model parameters and the resulting RMSE in mm for the ten SPC setups. The parameters for our method and that of Pertuz et al. from [10] were directly calculated based on the known optical properties. The fitted parameter sets were acquired via grid search and subsequent local optimization to fit the model to the data from Experiment (II).
Table 2. Results of Experiment (V): Model parameters and the resulting RMSE in mm for the ten SPC setups. The parameters for our method and that of Pertuz et al. from [10] were directly calculated based on the known optical properties. The fitted parameter sets were acquired via grid search and subsequent local optimization to fit the model to the data from Experiment (II).
Pertuz et al. [10]Ours (Equation (32))Fitted
Setup a 0 a 1 RMSE a 0 a 1 RMSE a 0 a 1 RMSE
Rodens.0.09270.463321.03−0.000140.37050.64−0.000380.369510.49
Zeiss0.18531.129428.05−0.131020.813041.51−0.131090.813611.51
Ricoh0.11400.339189.77−0.074350.150790.52−0.074030.151110.51
Canon0.13410.473345.500.026300.365530.740.0255470.365680.59
Olympus0.06740.237637.490.022670.192861.120.02130.191460.62
Rodens.0.092755.58896.87−0.0001855.4957.190.0019855.6256.83
Zeiss0.1070322.73153.30−0.09773322.539.67−0.08574323.618.35
Ricoh0.0635366.78364.78−0.07175366.656.76−0.06887367.545.03
Canon0.14581715.167.400.03702171510.260.045761724.18.69
Olympus0.137096.68346.520.0568096.6038.230.0601496.6318.07
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Michels, T.; Mäckelmann, D.; Koch, R. Mind the Exit Pupil Gap: Revisiting the Intrinsics of a Standard Plenoptic Camera. Sensors 2024, 24, 2522. https://doi.org/10.3390/s24082522

AMA Style

Michels T, Mäckelmann D, Koch R. Mind the Exit Pupil Gap: Revisiting the Intrinsics of a Standard Plenoptic Camera. Sensors. 2024; 24(8):2522. https://doi.org/10.3390/s24082522

Chicago/Turabian Style

Michels, Tim, Daniel Mäckelmann, and Reinhard Koch. 2024. "Mind the Exit Pupil Gap: Revisiting the Intrinsics of a Standard Plenoptic Camera" Sensors 24, no. 8: 2522. https://doi.org/10.3390/s24082522

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop