Next Article in Journal
Prediction and Elimination of Physiological Tremor During Control of Teleoperated Robot Based on Deep Learning
Previous Article in Journal
Piezoresistive, Piezocapacitive and Memcapacitive Silk Fibroin-Based Cement Mortars
Previous Article in Special Issue
Holo-U2Net for High-Fidelity 3D Hologram Generation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Instant-SFH: Non-Iterative Sparse Fourier Holograms Using Perlin Noise

by
David Li
1,*,
Susmija Jabbireddy
1,
Yang Zhang
2,
Christopher Metzler
1 and
Amitabh Varshney
1
1
Department of Computer Science, University of Maryland, College Park, MD 20742, USA
2
Department of Electrical and Computer Engineering, University of Maryland, College Park, MD 20742, USA
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(22), 7358; https://doi.org/10.3390/s24227358
Submission received: 18 October 2024 / Revised: 15 November 2024 / Accepted: 16 November 2024 / Published: 18 November 2024
(This article belongs to the Special Issue Digital Holography Imaging Techniques and Applications Using Sensors)

Abstract

:
Holographic displays are an upcoming technology for AR and VR applications, with the ability to show 3D content with accurate depth cues, including accommodation and motion parallax. Recent research reveals that only a fraction of holographic pixels are needed to display images with high fidelity, improving energy efficiency in future holographic displays. However, the existing iterative method for computing sparse amplitude and phase layouts does not run in real time; instead, it takes hundreds of milliseconds to render an image into a sparse hologram. In this paper, we present a non-iterative amplitude and phase computation for sparse Fourier holograms that uses Perlin noise in the image–plane phase. We conduct simulated and optical experiments. Compared to the Gaussian-weighted Gerchberg–Saxton method, our method achieves a run time improvement of over 600 times while producing a nearly equal PSNR and SSIM quality. The real-time performance of our method enables the presentation of dynamic content crucial to AR and VR applications, such as video streaming and interactive visualization, on holographic displays.

1. Introduction

Holographic displays are a promising upcoming display technology with the ability to display 3D content by reconstructing the wavefield of a complete 3D scene. For AR and VR devices, holographic displays also have the potential to resolve the vergence–accommodation conflict, which leads to fatigue and virtual reality sickness [1,2,3]. Current holographic displays are prototyped using spatial light modulators (SLMs) based on micro-electro-mechanical systems (DMD or MEMS) or liquid crystal (LCD or LCoS) technology supporting phase-only holograms. Future holographic displays promise control of both amplitude and phase with dual liquid–crystal panels [4], metasurfaces [5,6,7,8], or nanophotonic phased arrays (NPAs) [9,10,11,12].
To address the high power consumption of large-scale NPA-based holographic displays, Jabbireddy et al. [13] propose using sparse holograms, complex holograms with zero amplitude across a large fraction of the hologram–plane pixels. Sparse holograms reduce power consumption by activating a sparse array of active antennas. The hologram layout, phase, and amplitude are computed using a Gaussian-weighted Gerchberg–Saxton (GGS) algorithm for a given target image. By adding a Gaussian weighted amplitude to the Gerchberg–Saxton algorithm, their method redistributes energy in the hologram plane, concentrating it in a few pixels and enabling high sparsity levels while maintaining the image quality of the complete non-sparse image. Due to the iterative nature of Gerchberg–Saxton, GGS takes over 2 s to generate the sparse hologram for a single channel on an RTX 2080 GPU. This limitation prevents sparse holograms from being used for interactive applications which encompass most AR and VR use cases such as data visualization [14], virtual collaboration [15], and live captioning [16].
Non-iterative methods in computer-generated holography, either conventional or learned ones, could help improve the speed of hologram generation. Traditionally, the Gerchberg–Saxton algorithm [17] is used for phase-retrieval applications, such as computing the hologram–plane phase for phase-only holograms. As an alternative to Gerchberg–Saxton, Pang et al. [18] propose a non-iterative method that combines a quadratic image–plane phase with an error diffusion technique. However, this method is unsuitable for generating sparse holograms, which concentrate the amplitude in a few hologram pixels rather than diffusing it throughout all the pixels.
To enable sparse holograms for dynamic and interactive applications, in this paper, we examine how the Gaussian-weighted Gerchberg–Saxton (GGS) algorithm generates sparse holograms. We observe that each iteration of the GGS algorithm smooths out the image–plane phase, converging to a primarily low-frequency phase. Based on this observation, we propose to replace the random phase initialization of GGS with a low-frequency initial phase based on Perlin noise [19,20]. We show that, by selecting a sufficiently low frequency, a Perlin-noise-based phase initialization directly results in a converged solution for GGS, allowing us to instantly generate a sparse hologram.
Figure 1 shows a visual overview of our algorithm results with a random phase, a quadratic phase, and a Perlin-noise-based phase. We examine how the image–plane phase distribution affects the amplitude distribution of Fourier holograms. We discover that using a low-frequency Perlin noise allows us to generate sparse Fourier holograms (visualized at 80 % sparsity) with an over 600 × speedup over the current method.
In summary, the contributions of our paper are as follows:
  • We examine how the Gaussian-weighted Gerchberg–Saxton algorithm optimizes sparse holograms and discover that the frequency of the image–plane phase decreases until convergence.
  • We propose to use low-frequency Perlin noise as the image–plane phase, allowing us to create sparse holograms in about one millisecond.
  • We explore the relationship between the frequency of the image–plane phase and the reconstruction quality of sparse holograms.

2. Background and Related Work

Our work builds upon the existing research in computer-generated holography. First, we present some background on the sparse holograms. Then, we provide a brief overview of recent work in holographic displays. Finally, we discuss related work in Gerchberg–Saxton phase retrieval.

2.1. Sparse Holograms

Computer-generated holograms (CGHs) can be classified into three types: amplitude-only holograms, phase-only holograms, and complex holograms. Most existing work in CGHs focuses on generating phase-only holograms to accommodate the commercially available phase-only spatial light modulators found in existing holographic display setups. Our work focuses on sparse complex holograms designed for future holographic display setups with independent control of amplitude and phase for each hologram pixel.
In a far-field Fourier hologram setup, the hologram plane and the image plane are related by a Fourier transform, as shown in Figure 2. Since a far-field hologram undergoes a Fourier transform, we can use a hologram with sparsely activated pixel to display a complete image. As illustrated in Figure 3b, a sparse hologram is a hologram where only a subset of the hologram–plane pixels are activated with a non-zero amplitude.
Whereas phase-only holograms, illustrated in Figure 3a, are directly applicable to the current generation of phase-only spatial light modulators (SLMs), sparse holograms are intended to address upcoming SLMs that promise control of both amplitude and phase. Jabbireddy et al. [13] explore sparse holograms targeting nanophotonic phased arrays (NPAs) [9,10,11,12], where each antenna, or ‘pixel’, emits amplitude and phase independently. Under such a setup, the usage of each hologram pixel directly contributes to the power consumption and heat generation of the holographic display.
We note that this use of sparsity differs from Shi et al. [21], which uses the sparsity of light fields in the angular domain for light field reconstruction. Most existing methods such as Maimone et al. [22] and Peng et al. [23] are targeted towards generating real-time phase-only CGH rather than sparse CGH.

2.2. Holographic Displays and Algorithms

Holographic displays generate images through the controlled diffraction and interference of coherent light. Unlike traditional displays, holographic displays reconstruct a wavefront of the scene by controlling the phase of the emitted light. Holographic displays are built using spatial light modulators (SLMs). Current SLMs are built upon liquid crystal-based (LCD) or micro-electro-mechanical system (MEMS) technology. Existing SLMs typically only offer phase-only modulation and can also be referred to as phase spatial light modulators (PLMs). For complex holograms which have both amplitude and phase, researchers are in the process of developing two-phase holograms [24], metasurfaces [6,7], and nanophotonic phased arrays [9,10].
Digital holography falls into a class of problems known as inverse problems [25,26,27], since they aim to recover the hologram amplitude and phase which generates the observed target image. Two common types of holograms are Fresnel and Fourier holograms. Fresnel holograms are used in near-field holographic displays. Our paper focuses on Fourier holograms, which are used in far-field holographic displays.

2.3. Gerchberg–Saxton Phase Retrieval

Conventional spatial light modulators (SLMs) can only control either the amplitude or phase of the coherent light. In the case of phase-only SLMs, the Gerchberg–Saxton (GS) algorithm [17] is often used to compute the phase in the hologram plane with strict amplitude constraints in the hologram and image planes. Existing research has focused on improving the convergence performance and robustness of the GS algorithm for generating phase-only holograms.
For phase-only Fourier holograms, Chen et al. [28,29] develop a weighted constraint iterative algorithm (WCIA). Given target amplitude A t , reconstructed amplitude A r , and input amplitude A k , WCIA uses an adaptive over-compensation in their amplitude constraint, | A c o n | = | A k | ( | A t | / | A r | ) β k , in the signal region. Reducing the difference between the reconstructed amplitude and the new amplitude in the image plane allows for better amplitude uniformity in the hologram which enhances the convergence of Gerchberg–Saxton. Pang et al. [18] propose a non-iterative method to generate phase-only Fourier holograms using a quadratic phase ( φ = a m 2 + b n 2 ), which represents illumination with a point light source. Building upon WCIA, Wu et al. [30] develop an adaptive weighted Gerchberg–Saxton (AWGS). AWGS replaces the weighted ratio in the amplitude constraint with e A t A r to achieve more stable convergence. Furthermore, they propose initializing the phase in the target domain with an approximation to a quadratic phase. With both of these techniques, AWGS can generate phase-only Fourier holograms with stable convergence, fewer artifacts, and a PSNR that is 4.8 dB higher.
Thimons and Wittle [31] investigate the Gerchberg–Saxton algorithm under different conditions, including applying a low-frequency filter to the input magnitude, multiplying the input magnitude with Gaussian, and implementing a constant initial phase. In their experiments, they found that, when one set of magnitudes is non-centrosymmetric, using constant phase initialization leads to more stable and accurate convergence in far fewer iterations. Cruz-López and González-Velázquez [32] analyze the effects of several random phase distributions on Fourier and Fresnel hologram reconstruction. While we also examine the phase distribution, our method uses Perlin noise to generate sparse complex holograms rather than amplitude-only or phase-only holograms.
In contrast to these existing works, which develop algorithms for phase-only holograms and existing phase-only SLM hardware, our work is focused on the constraint of sparsity for future complex SLM hardware. To the best of our knowledge, we believe Instant-SFH is the first to use Perlin noise for constructing sparse holograms and the first to generate sparse holograms in real time with equivalent quality to those generated by GGS.

3. Method

3.1. Gaussian-Weighted Gerchberg–Saxton

In a Fourier hologram setup, an SLM produces a complex-valued source hologram E s = S e i ϕ S with amplitude S and phase ϕ S . The source hologram propagates through free space or a Fourier lens to form a target far-field image E t = T e i ϕ T . At the far-field image plane, only the intensity I T 2 is perceptible to the viewer. Jabbireddy et al. [13] propose the Gaussian-weighted Gerchberg–Saxton (GGS) algorithm to generate sparse amplitude and phase layouts for Fourier holograms. GGS leverages the Gerchberg–Saxton algorithm while modulating the amplitude in the hologram and image planes. However, unlike in traditional phase-retrieval applications, GGS does not apply a strict amplitude constraint in the hologram plane. Instead, at each iteration i, the amplitude constraint A c o n used is a element-wise product of a Gaussian weight map W and the current amplitude | A i | in the hologram plane:
A c o n = W | A i |
Upon convergence, or after 1000 iterations, the amplitudes in the hologram plane are truncated to the desired sparsity level, resulting in a sparse layout for a Fourier hologram with amplitude and phase.
In their experiments, Jabbireddy et al. [13] observe that each iteration of GGS takes about 3.5 ms for a single channel. In total, their end-to-end pipeline takes 2.26 s per image channel on average. Without the ability to run in real time, GGS is limited to generating holograms of things that can be pre-computed offline, such as static images and text. This limitation prohibits using sparse holograms in interactive applications and makes it difficult to view even videos, which would require pre-computing and storing a sequence of sparse holograms.
A real-time method would enable sparse holograms to be used in far more scenarios such as video streaming, data visualization, and other interactive applications. With the goal of real-time sparse holograms in mind, we first examine how GGS generates sparse holograms. The standard Gerchberg–Saxton algorithm enforces a strict amplitude constraint in both the image plane and hologram plane. GGS uses a more relaxed constraint by modulating the amplitude in the hologram plane instead of replacing it in each iteration. However, the amplitude is still strictly enforced in the image plane. Since the Fourier transform is an invertible transformation between holograms and images, the amplitude constraint in the hologram plane must be reflected in the image–plane phase. We visualize how the image–plane phase changes in Figure 4 and Figure 5a. In Figure 5a, we see that the phase changes significantly at the first iteration. This is expected of a non-converged phase. As GGS progresses through 1000 iterations in Figure 4, the image–plane phase quickly becomes smoother, consisting primarily of low-frequency components.

3.2. Perlin Noise

Since GGS converges to a low-frequency phase distribution, we hypothesize that initializing with a low-frequency phase will result in faster convergence. To initialize the image–plane phase with a low-frequency distribution, we propose the use of Perlin noise [19,20]. Perlin noise is a band-limited noise function that allows direct control over the frequency content and can be made periodic with a proper selection of gradient vectors. This property is not available in other types of noise such as Gaussian noise. In computer graphics, Perlin noise is commonly used to generate procedural textures and terrain [19,33,34].
Perlin noise [19,20] is generated by enumerating a low-resolution 2D grid with unit-length 2D vectors, known as gradient vectors. Next, we enumerate a high-resolution 2D image grid mapping to the low-resolution grid at a particular frequency. For each image pixel, offset vectors to the four nearest low-resolution grid points are computed. The dot products between gradient and offset vectors are computed at each grid point. Finally, a smooth interpolation is applied across the four dot product values to produce a smooth noise image. We scale the resulting Perlin noise output to an angle between 0 to 2 π . Implementations of Perlin noise are available online (Perlin noise: https://github.com/pvigier/perlin-numpy, accessed on 21 April 2023).
When replacing the random initial phase with a low-frequency Perlin-noise-based phase, we observe that successive GGS iterations do not significantly change the image–plane phase, as shown in Figure 5b, suggesting a converged solution. Hence, with Perlin noise, a single inverse FFT can be used to generate a sparse hologram without the need for GGS, as shown in Figure 6. Directly generating a sparse hologram with a high-frequency random phase [17] or quadratic phase [18] does not result in an acceptable reconstruction, as shown in Figure 7.

4. Experiments

We first examine the relationship between the frequency of the image–plane phase and the sparse Fourier hologram reconstruction quality. Next, we explore whether there is a relationship between the frequency of the image amplitude and the frequency of the image phase when generating sparse holograms. Finally, we evaluate the quality and performance of our instant sparse Fourier hologram method by comparing it to the Gaussian-weighted Gerchberg–Saxton method.

4.1. Implementation and Setup

Our experiments are conducted using an existing image dataset. Following Jabbireddy et al. [13], we use the X2 bicubic split of the DIV2K dataset (Data available at https://data.vision.ee.ethz.ch/cvl/DIV2K/, accessed on 21 April 2023) [35]. This dataset contains a total of 900 images, most of which have a landscape orientation with 1020 × 678 resolution. To ensure consistent quality and performance evaluation results, we rotate portrait images to landscape and resize and center crop all images to 1020 × 678 . We evaluate our method using all 900 images. We implement our algorithm and baseline methods using PyTorch 2.0.1, which enables GPU-accelerated computations. Our experiments are run on a workstation with an Intel Core i7-12700K and an NVIDIA RTX 4090.
We first evaluate the performance and quality of our instant sparse Fourier hologram method (Instant-SFH). We compare our method to the existing offline GGS method [13], which uses a random initial phase and up to 1000 iterations for convergence. Note that the offline GGS method is not suitable for dynamic content such as streaming video or interactive 3D visualization. To quantitatively evaluate the reconstruction quality of our methods, we use two standard reference-based image quality metrics: the peak signal-to-noise ratio (PSNR) and the structural similarity metric (SSIM).

4.2. Frequency Ablation

Perlin noise [19,20] is generated by enumerating a grid with unit-length vectors and interpolating between grid values using dot products and linear interpolation. The sampling rate, or the ratio between image pixels and the underlying grid, represents the frequency of the noise generated. To determine the ideal frequency of Perlin noise to use for the image–plane phase, we evaluate how different frequencies affect the reconstruction quality. Specifically, we conduct experiments across 28 frequencies: 0 , 1 , . . . , 9 , 10 , 20 , . . . , 90 , 100 , 200 , . . . , 900 . This value is the resolution of the underlying grid being sampled when generating Perlin noise. For frequency 0, we use a constant zero phase across the image.
Quantitative results averaged over all 900 images ( 28 × 900 = 25.2 K reconstructions) are shown in Figure 8. Qualitative results at a few frequency levels are shown in Figure 9 with corresponding hologram amplitudes visualized in Figure 10. Across all sparsity levels, we observe that the PSNR and SSIM quality decreases as the frequency of the Perlin noise increases. The PSNR and SSIM values are maximized at the lowest non-zero frequencies, specifically frequencies 1 to 10. Additionally, we observe that the higher the sparsity level, the more sensitive the reconstruction is to the phase frequency. For the majority of our evaluation, namely Section 4.4 and Section 4.5, we use a frequency of 3 for our Instant-SFH method.

4.3. Amplitude vs. Phase Frequency

Next, we examine whether there is a relationship between the frequency of the amplitude and the frequency of the phase in the image plane when generating sparse holograms. To control the frequency of the amplitude, we apply a Gaussian blur to the target image and conduct evaluations using different band-limited targets.
Qualitative results on a single image are shown in Figure 11, and quantitative results averaged over all images are shown in Figure 12. Gaussian-blurred images can be easily reconstructed with low-frequency phases, achieving a higher SSIM than normal images at the same sparsity level due to the lack of high-frequency details. With a high-frequency phase, speckling artifacts emerge in sparse hologram reconstructions. These artifacts are much more noticeable in low-frequency (e.g., blurred) regions of the target image that otherwise lack high-frequency details in the reconstruction. Hence, the SSIM of Gaussian-blurred image reconstructions falls below that of standard images.

4.4. Performance

First, we evaluate the performance of our method to determine whether interactive performance can be achieved by removing GGS. We present cumulative performance results over all three color channels in Table 1. Note that the results shown are rounded so run times for all stages may not add up to the total run time. On our workstation, we observe that offline GGS takes 664.4 ms, making it unsuitable for interactive scenarios. Using GGS in an interactive scenario would result in less than 2 FPS. Meanwhile, our instant sparse Fourier hologram method runs in 1.0 ms on average, a performance improvement of over 600 × . Next, we benchmark our Instant-SFH method at higher resolutions to determine whether it is suitable for large-field-of-view applications. At 720p HD and 1080p full HD resolutions, our method runs in 1.2 ms and 4.0 ms, respectively, as shown in Table 2. The high performance of our method allows sparse holograms to take advantage of the high refresh rate available in holographic displays and enables high-refresh-rate and high-resolution content to be displayed.

4.5. Quality

Finally, we compare the image reconstruction quality of GGS and Instant-SFH. Quantitative results for GGS and Instant-SFH are shown in Table 3. In the offline setup, where up to 1000 iterations of GGS can be computed, initializing with a random phase achieves very high PSNR and SSIM results even at 70 % or 80 % sparsity. With Perlin noise, our Instant-SFH method matches the PSNR and SSIM of GGS without requiring any iterations of GGS. Random phase [17] and quadratic phase [18] generate a sparse hologram with PSNR values around 11 dB, which is not acceptable. A constant zero phase yields a sparse hologram but with ≈1 dB worse PSNR results than low-frequency Perlin noise. Finally, optimized random phase (ORAP) [36], which involves transferring a phase distribution optimized for another image, achieves PSNR and SSIM scores slightly below that of GGS and Perlin noise. In Figure 13, our Instant-SFH method achieves PSNR and SSIM results consistent with GGS across all sparsity levels.

4.6. Optical Verification

In the absence of commercially available amplitude-phase SLMs, we use a phase-only SLM to optically verify that our Instant-SFH method achieves similar results to GGS. Specifically, we generate 80 % sparse complex holograms using GGS and Instant-SFH. Both sets of sparse holograms are then converted to phase-only holograms using the double-phase amplitude coding (DPAC) [22,23,37] method and displayed on a phase-only SLM. While our method is broadly applicable to natural images, phase-only holograms produced by DPAC cannot encode high amplitude information at low frequencies [38]. Consequently, we opt to display sketches from the human sketch image dataset published by Eitz et al. [39] (Sketch Dataset: https://cybertron.cg.tu-berlin.de/eitz/projects/classifysketch/, accessed on 21 April 2023). Our setup consists of a Holoeye Leto SLM with 1080p resolution and 6.4 µm pixel pitch. Coherent light emitted by a 532 nm green laser is passed through a beam expander and is incident on the SLM. The modulated light forms a far-field image, which is captured by a 1024 × 1280 camera through a 200 mm lens. Our optical verification setup and results are shown in Figure 14. From our experiments, we see that Instant-SFH generates similar results to GGS despite only taking a fraction of the time. In this situation, we replace reconstruction with DPAC, which only takes 0.1 ms. We attribute the speckling artifacts to the DPAC conversion and environmental variables since they exist in both the results of GGS and Instant-SFH. As our objective is to show the quality similarity between our real-time method and offline GGS, we have not tuned the Perlin noise to improve the quality for this scenario.

5. Discussion

This paper focuses on Fourier holograms that display flat 2D images, similar to other recent holography papers [13,40]. To achieve the full benefits of holographic displays including accommodation and motion parallax, 3D content can be displayed using near-field holograms with multiple focal planes [41,42]. Since Instant-SFH runs at KHz rates, our method can be extended to display 3D by generating multiple 2D depth slices and time-multiplexing near-field 2D holograms generated with a Fresnel transform. Alternatively, the 3D method by Chakravarthula et al. [43] also applies whereby an eye-tracker is used to determine the depth of the foveal region based on the scene and a 2D Fresnel hologram is shown at the detected depth.
As observed by Chakravarthula et al. [40], a constant phase can lead to a smaller effective eye box as the light energy is concentrated in the small, low-frequency region. In Figure 15, we observe that Instant-SFH and GGS have a similar eyebox as both use smooth phases, prioritizing hologram quality over eyebox size. While outside the scope of this paper, our sub-ms run-times also open up the possibility of adaptively compensating for a moving eyebox. For example, in a pupil-tracked environment [44], one could pair Perlin noise with a few iterations of GGS and a pupil-weighted threshold [40] when determining the amplitudes at the SLM. Such compensation is only viable for algorithms, like ours, that can run at KHz rates.
Time-division multiplexing for multiple focal planes and dynamic eyebox adjustment require fast algorithms to achieve interactive frame rates. The millisecond run times of Instant-SFH open up the possibility of pairing such techniques with sparse holograms. Our evaluation performs FFT using a desktop NVIDIA RTX 4090 GPU. To deploy our algorithm on more portable form factors, we believe a digital signal processor with the ability to perform FFT in real time would be necessary to accommodate the requirements of Instant-SFH.

6. Conclusions

In this paper, we examine how sparse Fourier holograms can be adapted to view dynamic content such as videos and interactive applications. First, we examine how the Gaussian-weighted Gerchberg–Saxton (GGS) algorithm alters the image–plane phase. Next, we discover that using Perlin noise allows for the instantaneous generation of sparse Fourier holograms, eliminating the need for any iterations of GGS. Finally, we quantitatively evaluate how the frequency of the Perlin noise affects the hologram reconstruction quality across an existing dataset of 900 images and on band-limited images. Our Instant-SFH method achieves a performance improvement of over 600 times, opening up the possibility of using sparse holograms for dynamic content. We believe our method will broaden the use of upcoming holographic displays for dynamic applications.

Author Contributions

Conceptualization, D.L., S.J. and A.V.; methodology, D.L., S.J. and A.V.; software, D.L. and S.J.; validation, D.L. and S.J.; investigation, D.L. and S.J.; resources, Y.Z., C.M. and A.V.; writing—original draft preparation, D.L.; writing—review and editing, D.L., S.J., C.M. and A.V.; supervision, A.V.; project administration, A.V.; funding acquisition, A.V. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported in part by the NSF Grants 18-23321, 21-37229, and 22-35050 and the State of Maryland’s MPower initiative. Any opinions, findings, conclusions, or recommendations expressed in this article are those of the authors and do not necessarily reflect the views of the research sponsors.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Our simulated experiments are conducted on the X2 bicubic split of the DIV2K dataset [35] available at https://data.vision.ee.ethz.ch/cvl/DIV2K/ (accessed on 21 April 2023). Our optical verification results are conducted on the sketch dataset by Eitz et al. [39] available at https://cybertron.cg.tu-berlin.de/eitz/projects/classifysketch/ (accessed on 21 April 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hoffman, D.M.; Girshick, A.R.; Akeley, K.; Banks, M.S. Vergence-accommodation conflicts hinder visual performance and cause visual fatigue. J. Vis. 2008, 8, 33. [Google Scholar] [CrossRef] [PubMed]
  2. Kramida, G. Resolving the Vergence-Accommodation Conflict in Head-Mounted Displays. IEEE Trans. Vis. Comput. Graph. 2016, 22, 1912–1931. [Google Scholar] [CrossRef] [PubMed]
  3. Krokos, E.; Varshney, A. Quantifying VR cybersickness using EEG. Virtual Real. 2022, 26, 77–89. [Google Scholar] [CrossRef]
  4. Jang, S.W.; Choi, W.; Kim, S.; Lee, J.; Na, S.; Ham, S.; Park, J.; Kang, H.; Ju, B.K.; Kim, H. Complex spatial light modulation capability of a dual layer in-plane switching liquid crystal panel. Sci. Rep. 2022, 12, 8277. [Google Scholar] [CrossRef]
  5. Wang, Q.; Zhang, X.; Xu, Y.; Gu, J.; Li, Y.; Tian, Z.; Singh, R.; Zhang, S.; Han, J.; Zhang, W. Broadband metasurface holograms: Toward complete phase and amplitude engineering. Sci. Rep. 2016, 6, 32867. [Google Scholar] [CrossRef]
  6. Lee, G.Y.; Yoon, G.; Lee, S.Y.; Yun, H.; Cho, J.; Lee, K.; Kim, H.; Rho, J.; Lee, B. Complete amplitude and phase control of light using broadband holographic metasurfaces. Nanoscale 2018, 10, 4237–4245. [Google Scholar] [CrossRef]
  7. Jiang, Q.; Jin, G.; Cao, L. When metasurface meets hologram: Principle and advances. Adv. Opt. Photon. 2019, 11, 518–576. [Google Scholar] [CrossRef]
  8. Deng, Z.L.; Jin, M.; Ye, X.; Wang, S.; Shi, T.; Deng, J.; Mao, N.; Cao, Y.; Guan, B.O.; Alù, A.; et al. Full-Color Complex-Amplitude Vectorial Holograms Based on Multi-Freedom Metasurfaces. Adv. Funct. Mater. 2020, 30, 1910610. [Google Scholar] [CrossRef]
  9. Sun, J.; Timurdogan, E.; Yaacobi, A.; Hosseini, E.S.; Watts, M.R. Large-scale nanophotonic phased array. Nature 2013, 493, 195–199. [Google Scholar] [CrossRef]
  10. Raval, M.; Yaacobi, A.; Coleman, D.; Fahrenkopf, N.M.; Baiocco, C.; Leake, G.; Adam, T.N.; Coolbaugh, D.; Watts, M.R. Nanophotonic phased array for visible light image projection. In Proceedings of the 2016 IEEE Photonics Conference (IPC), Waikoloa, HI, USA, 2–6 October 2016; pp. 206–207. [Google Scholar] [CrossRef]
  11. Sun, X.; Zhang, Y.; Huang, P.C.; Acharjee, N.; Dagenais, M.; Peckerar, M.; Varshney, A. Correcting the Proximity Effect in Nanophotonic Phased Arrays. IEEE Trans. Vis. Comput. Graph. 2020, 26, 3503–3513. [Google Scholar] [CrossRef]
  12. Sun, X.; Zhang, Y.; Huang, P.; Acharjee, N.; Dagenais, M.; Peckerar, M.; Varshney, A. Proximity Effect Correction for Fresnel Holograms on Nanophotonic Phased Arrays. In Proceedings of the 2021 IEEE Virtual Reality and 3D User Interfaces (VR), Lisbon, Portugal, 27 March–1 April 2021; pp. 353–362. [Google Scholar] [CrossRef]
  13. Jabbireddy, S.; Zhang, Y.; Peckerar, M.; Dagenais, M.; Varshney, A. Sparse Nanophotonic Phased Arrays for Energy-Efficient Holographic Displays. In Proceedings of the 2022 IEEE Conference on Virtual Reality and 3D User Interfaces (VR), Christchurch, New Zealand, 12–16 March 2022; pp. 553–562. [Google Scholar] [CrossRef]
  14. Shirah, G. Inside Hurricane Maria in 360 Degrees. In Proceedings of the ACM SIGGRAPH 2019 Computer Animation Festival, New York, NY, USA, 28 July 2019. [Google Scholar] [CrossRef]
  15. He, Z.; Du, R.; Perlin, K. CollaboVR: A Reconfigurable Framework for Creative Collaboration in Virtual Reality. In Proceedings of the 2020 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Virtual, 9–13 November 2020; pp. 542–554. [Google Scholar] [CrossRef]
  16. Olwal, A.; Balke, K.; Votintcev, D.; Starner, T.; Conn, P.; Chinh, B.; Corda, B. Wearable Subtitles: Augmenting Spoken Communication with Lightweight Eyewear for All-Day Captioning. In Proceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology, UIST ’20, New York, NY, USA, 20–23 October 2020; pp. 1108–1120. [Google Scholar] [CrossRef]
  17. Gerchberg, R.W. A practical algorithm for the determination of plane from image and diffraction pictures. Optik 1972, 35, 237–246. [Google Scholar]
  18. Pang, H.; Wang, J.; Zhang, M.; Cao, A.; Shi, L.; Deng, Q. Non-iterative phase-only Fourier hologram generation with high image quality. Opt. Express 2017, 25, 14323–14333. [Google Scholar] [CrossRef] [PubMed]
  19. Perlin, K. An Image Synthesizer. In Proceedings of the 12th Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH ’85, New York, NY, USA, 22–26 July 1985; pp. 287–296. [Google Scholar] [CrossRef]
  20. Perlin, K. Improving Noise. In Proceedings of the 29th Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH ’02, New York, NY, USA, 21–26 July 2002; pp. 681–682. [Google Scholar] [CrossRef]
  21. Shi, L.; Hassanieh, H.; Davis, A.; Katabi, D.; Durand, F. Light Field Reconstruction Using Sparsity in the Continuous Fourier Domain. ACM Trans. Graph. 2015, 34, 1–13. [Google Scholar] [CrossRef]
  22. Maimone, A.; Georgiou, A.; Kollin, J.S. Holographic Near-Eye Displays for Virtual and Augmented Reality. ACM Trans. Graph. 2017, 36, 1–16. [Google Scholar] [CrossRef]
  23. Peng, Y.; Choi, S.; Padmanaban, N.; Kim, J.; Wetzstein, G. Neural Holography. In Proceedings of the ACM SIGGRAPH 2020 Emerging Technologies, SIGGRAPH ’20, New York, NY, USA, 17 August 2020. [Google Scholar] [CrossRef]
  24. Brown, M.D. Phase and amplitude modulation with acoustic holograms. Appl. Phys. Lett. 2019, 115, 053701. [Google Scholar] [CrossRef]
  25. Rosen, J.; Alford, S.; Allan, B.; Anand, V.; Arnon, S.; Arockiaraj, F.G.; Art, J.; Bai, B.; Balasubramaniam, G.M.; Birnbaum, T.; et al. Roadmap on computational methods in optical imaging and holography [invited]. Appl. Phys. B 2024, 130, 166. [Google Scholar] [CrossRef]
  26. Khare, K.; Butola, M.; Rajora, S. Image Reconstruction from Projections. In Fourier Optics and Computational Imaging; Springer International Publishing: Cham, Switzerland, 2023; pp. 271–279. [Google Scholar] [CrossRef]
  27. Donatelli, M.; Serra-Capizzano, S. Computational Methods for Inverse Problems in Imaging; Springer: Berlin/Heidelberg, Germany, 2019; Volume 1. [Google Scholar]
  28. Chen, L.; Zhang, H.; He, Z.; Wang, X.; Cao, L.; Jin, G. Weighted Constraint Iterative Algorithm for Phase Hologram Generation. Appl. Sci. 2020, 10, 3652. [Google Scholar] [CrossRef]
  29. Chen, L.; Zhang, H.; Cao, L.; Jin, G. Non-iterative phase hologram generation with optimized phase modulation. Opt. Express 2020, 28, 11380–11392. [Google Scholar] [CrossRef]
  30. Wu, Y.; Wang, J.; Chen, C.; Liu, C.J.; Jin, F.M.; Chen, N. Adaptive weighted Gerchberg-Saxton algorithm for generation of phase-only hologram with artifacts suppression. Opt. Express 2021, 29, 1412–1427. [Google Scholar] [CrossRef]
  31. Thimons, T.; Wittle, L. Investigating the Gerchberg-Saxton Phase Retrieval Algorithm. SIAM Undergrad. Res. Online 2018, 11, 1–11. [Google Scholar] [CrossRef]
  32. Cruz-López, M.L.; González-Velázquez, K.G. Analysis of random-phase distributions and Perlin noise in CGH: A study of its effects on Fourier and Fresnel holograms reconstruction. Opt. Eng. 2020, 59, 102419. [Google Scholar] [CrossRef]
  33. Rhoades, J.; Turk, G.; Bell, A.; State, A.; Neumann, U.; Varshney, A. Real-Time Procedural Textures. In Proceedings of the 1992 Symposium on Interactive 3D Graphics, I3D ’92, New York, NY, USA, 29 March–1 April 1992; pp. 95–100. [Google Scholar] [CrossRef]
  34. Génevaux, J.D.; Galin, E.; Guérin, E.; Peytavie, A.; Benes, B. Terrain Generation Using Procedural Models Based on Hydrology. ACM Trans. Graph. 2013, 32, 1–13. [Google Scholar] [CrossRef]
  35. Agustsson, E.; Timofte, R. NTIRE 2017 Challenge on Single Image Super-Resolution: Dataset and Study. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1122–1131. [Google Scholar] [CrossRef]
  36. Zea, A.V.; Ramirez, J.F.B.; Torroba, R. Optimized random phase only holograms. Opt. Lett. 2018, 43, 731–734. [Google Scholar] [CrossRef] [PubMed]
  37. Hsueh, C.K.; Sawchuk, A.A. Computer-generated double-phase holograms. Appl. Opt. 1978, 17, 3874–3883. [Google Scholar] [CrossRef]
  38. Wu, Y.; Wang, J.; Chen, C.; Liu, C.J.; Jin, F.M.; Hu, Y.H. Generation of Phase-Only Fourier Hologram Based on Double Phase Method and Quantization Error Analysis. IEEE Photonics J. 2020, 12, 1–13. [Google Scholar] [CrossRef]
  39. Eitz, M.; Hays, J.; Alexa, M. How Do Humans Sketch Objects? ACM Trans. Graph. (Proc. SIGGRAPH) 2012, 31, 44:1–44:10. [Google Scholar] [CrossRef]
  40. Chakravarthula, P.; Baek, S.H.; Schiffers, F.; Tseng, E.; Kuo, G.; Maimone, A.; Matsuda, N.; Cossairt, O.; Lanman, D.; Heide, F. Pupil-Aware Holography. ACM Trans. Graph. 2022, 41, 1–15. [Google Scholar] [CrossRef]
  41. Choi, S.; Gopakumar, M.; Peng, Y.; Kim, J.; O’Toole, M.; Wetzstein, G. Time-Multiplexed Neural Holography: A Flexible Framework for Holographic Near-Eye Displays with Fast Heavily-Quantized Spatial Light Modulators. In Proceedings of the ACM SIGGRAPH 2022 Conference Proceedings, SIGGRAPH ’22, Vancouver, BC, Canada, 8–11 August 2022. [Google Scholar] [CrossRef]
  42. Kavaklı, K.; Itoh, Y.; Urey, H.; Akşit, K. Realistic Defocus Blur for Multiplane Computer-Generated Holography. In Proceedings of the 2023 IEEE Conference Virtual Reality and 3D User Interfaces (VR), Shanghai, China, 25–29 March 2023; pp. 418–426. [Google Scholar] [CrossRef]
  43. Chakravarthula, P.; Peng, Y.; Kollin, J.; Fuchs, H.; Heide, F. Wirtinger Holography for Near-Eye Displays. ACM Trans. Graph. 2019, 38, 1–13. [Google Scholar] [CrossRef]
  44. Jang, C.; Bang, K.; Moon, S.; Kim, J.; Lee, S.; Lee, B. Retinal 3D: Augmented reality near-eye display via pupil-tracked light field projection on retina. ACM Trans. Graph. 2017, 36, 1–13. [Google Scholar] [CrossRef]
Figure 1. On the (left), we show a comparison between the random phase, the quadratic phase [18], and Perlin noise [19,20]. In the (middle), we show the amplitudes of 80% sparse holograms constructed using each of these image–plane phase distributions. On the (right), we show images reconstructed from the sparse holograms for each of these image–plane phase distributions.
Figure 1. On the (left), we show a comparison between the random phase, the quadratic phase [18], and Perlin noise [19,20]. In the (middle), we show the amplitudes of 80% sparse holograms constructed using each of these image–plane phase distributions. On the (right), we show images reconstructed from the sparse holograms for each of these image–plane phase distributions.
Sensors 24 07358 g001
Figure 2. Visual overview of a Fourier hologram setup. On the hologram plane (left), a complex hologram with amplitude and phase is emitted from the hologram plane. The hologram passes through a Fourier lens (middle). On the image plane (right), a complete image is emitted.
Figure 2. Visual overview of a Fourier hologram setup. On the hologram plane (left), a complex hologram with amplitude and phase is emitted from the hologram plane. The hologram passes through a Fourier lens (middle). On the image plane (right), a complete image is emitted.
Sensors 24 07358 g002
Figure 3. Illustrative comparison between a (a) phase-only hologram and a (b) sparse hologram. In (a), pixels on the hologram plane have equal amplitude (visualized as transparency) and differing phases (visualized as hue). In (b), only a subset of pixels are active with non-zero amplitude and differing phases. In both (a) and (b), a complete image can be produced on the image plane.
Figure 3. Illustrative comparison between a (a) phase-only hologram and a (b) sparse hologram. In (a), pixels on the hologram plane have equal amplitude (visualized as transparency) and differing phases (visualized as hue). In (b), only a subset of pixels are active with non-zero amplitude and differing phases. In both (a) and (b), a complete image can be produced on the image plane.
Sensors 24 07358 g003
Figure 4. Image plane phase of Gaussian-weighted Gerchberg–Saxton (GGS) visualized with angles directly mapped to color hues visualized at initialization (a), after 10 GGS iterations (b), after 100 GGS iterations (c), and after 1000 GGS iterations (d).
Figure 4. Image plane phase of Gaussian-weighted Gerchberg–Saxton (GGS) visualized with angles directly mapped to color hues visualized at initialization (a), after 10 GGS iterations (b), after 100 GGS iterations (c), and after 1000 GGS iterations (d).
Sensors 24 07358 g004
Figure 5. Image–plane phase visualizations at initialization and after a single GGS iteration. In (a), we show a random initial phase and the phase after a single iteration. In (b), we show a Perlin-noise-based initial phase the phase after a single iteration.
Figure 5. Image–plane phase visualizations at initialization and after a single GGS iteration. In (a), we show a random initial phase and the phase after a single iteration. In (b), we show a Perlin-noise-based initial phase the phase after a single iteration.
Sensors 24 07358 g005
Figure 6. Overview of our instant sparse Fourier hologram (Instant-SFH) evaluation pipeline with three components: hologram generation, truncation, and reconstruction. In hologram generation, the target amplitude is augmented with a Perlin noise phase and converted to a complex hologram using inverse FFT. In truncation, the complex hologram is converted to a sparse hologram by zeroing out pixels at the lowest amplitude values. In reconstruction, we apply far-field propagation by taking an FFT to reconstruct the image.
Figure 6. Overview of our instant sparse Fourier hologram (Instant-SFH) evaluation pipeline with three components: hologram generation, truncation, and reconstruction. In hologram generation, the target amplitude is augmented with a Perlin noise phase and converted to a complex hologram using inverse FFT. In truncation, the complex hologram is converted to a sparse hologram by zeroing out pixels at the lowest amplitude values. In reconstruction, we apply far-field propagation by taking an FFT to reconstruct the image.
Sensors 24 07358 g006
Figure 7. Comparison of 80 % sparse hologram reconstructions using several methods. The original image is first shown in (a) as a reference. Then for each method, we show the full image and a zoomed-in crop. (a) Original image; (b) sparse hologram reconstruction with quadratic phase, (≈1 ms); (c) sparse hologram reconstruction with random phase, (≈1 ms); (d) sparse hologram reconstruction with random phase and 1000 GGS iterations (≈600 ms); (e) sparse hologram reconstruction with Perlin-noise-based phase (≈1 ms).
Figure 7. Comparison of 80 % sparse hologram reconstructions using several methods. The original image is first shown in (a) as a reference. Then for each method, we show the full image and a zoomed-in crop. (a) Original image; (b) sparse hologram reconstruction with quadratic phase, (≈1 ms); (c) sparse hologram reconstruction with random phase, (≈1 ms); (d) sparse hologram reconstruction with random phase and 1000 GGS iterations (≈600 ms); (e) sparse hologram reconstruction with Perlin-noise-based phase (≈1 ms).
Sensors 24 07358 g007
Figure 8. Quantitative evaluation of the sparse hologram reconstruction quality across Perlin noise frequency levels from 0 to 100 and sparsity levels from 30 % to 90 % .
Figure 8. Quantitative evaluation of the sparse hologram reconstruction quality across Perlin noise frequency levels from 0 to 100 and sparsity levels from 30 % to 90 % .
Sensors 24 07358 g008
Figure 9. Qualitative ablation results showing the generated Perlin noise (left) and resulting 80 % sparse hologram reconstruction (right) at Perlin noise frequency levels (a) 5, (b) 50, (c) 100, and (d) 500. Please zoom in to see the differences in details.
Figure 9. Qualitative ablation results showing the generated Perlin noise (left) and resulting 80 % sparse hologram reconstruction (right) at Perlin noise frequency levels (a) 5, (b) 50, (c) 100, and (d) 500. Please zoom in to see the differences in details.
Sensors 24 07358 g009
Figure 10. Qualitative ablation results showing the generated Perlin noise (left) and resulting 80% sparse hologram amplitudes (right) at Perlin noise frequency levels (a) 5, (b) 50, (c) 100, and (d) 500. The amplitude corresponds to the red channel of the image in Figure 9, where yellow pixels indicate amplitudes at or above 0.001.
Figure 10. Qualitative ablation results showing the generated Perlin noise (left) and resulting 80% sparse hologram amplitudes (right) at Perlin noise frequency levels (a) 5, (b) 50, (c) 100, and (d) 500. The amplitude corresponds to the red channel of the image in Figure 9, where yellow pixels indicate amplitudes at or above 0.001.
Sensors 24 07358 g010
Figure 11. Qualitative results showing band-limited target images and resulting 80 % sparse hologram reconstruction quality at frequency 200 Perlin noise.
Figure 11. Qualitative results showing band-limited target images and resulting 80 % sparse hologram reconstruction quality at frequency 200 Perlin noise.
Sensors 24 07358 g011
Figure 12. Quantitative evaluation with Gaussian blurred target images at 80 % sparsity.
Figure 12. Quantitative evaluation with Gaussian blurred target images at 80 % sparsity.
Sensors 24 07358 g012
Figure 13. Quantitative comparison between GGS [13] and our Instant-SFH method. Note that GGS and Instant-SFH have overlapping results. Results are averaged over 900 images.
Figure 13. Quantitative comparison between GGS [13] and our Instant-SFH method. Note that GGS and Instant-SFH have overlapping results. Results are averaged over 900 images.
Sensors 24 07358 g013
Figure 14. Optical verification setup and results. Sparse holograms (80% sparsity) from GGS and Instant-SFH are converted to phase-only holograms using double phase-amplitude coding and displayed on a phase-only SLM.The left column shows the target images. The center column shows images generated from GGS. The right column shows images generated from Instant-SFH.
Figure 14. Optical verification setup and results. Sparse holograms (80% sparsity) from GGS and Instant-SFH are converted to phase-only holograms using double phase-amplitude coding and displayed on a phase-only SLM.The left column shows the target images. The center column shows images generated from GGS. The right column shows images generated from Instant-SFH.
Sensors 24 07358 g014
Figure 15. Eyebox comparison between GGS and Instant-SFH with Perlin-noise-based phase at 80% sparsity. The pupil mask (top row) shows the pupil moving progressively to the right. The center row shows images reconstructed from GGS. The bottom row shows images reconstructed from Instant-SFH using Perlin noise.
Figure 15. Eyebox comparison between GGS and Instant-SFH with Perlin-noise-based phase at 80% sparsity. The pupil mask (top row) shows the pupil moving progressively to the right. The center row shows images reconstructed from GGS. The bottom row shows images reconstructed from Instant-SFH using Perlin noise.
Sensors 24 07358 g015
Table 1. Run time (ms) evaluation comparing running GGS with 1000 iterations and our Instant-SFH method at 1020 × 678 resolution. The results are the total run time for all three channels of the RGB images. Numbers in parentheses indicate standard deviation.
Table 1. Run time (ms) evaluation comparing running GGS with 1000 iterations and our Instant-SFH method at 1020 × 678 resolution. The results are the total run time for all three channels of the RGB images. Numbers in parentheses indicate standard deviation.
MethodGenerateTruncateReconstructTotal
GGS 595.6 0.4 0.2 596.2
(std) ( 1.0 ) ( 0.1 ) ( 0.0 ) ( 1.0 )
Instant-SFH 0.4 0.4 0.2 1.0
(std) ( 0.1 ) ( 0.1 ) ( 0.0 ) ( 0.1 )
Table 2. Run time (ms) evaluation for our Instant-SFH method at 720p ( 1280 × 720 ) and 1080p ( 1920 × 1080 ) resolution. Numbers in parentheses indicate standard deviation.
Table 2. Run time (ms) evaluation for our Instant-SFH method at 720p ( 1280 × 720 ) and 1080p ( 1920 × 1080 ) resolution. Numbers in parentheses indicate standard deviation.
ResolutionGenerateTruncateReconstructTotal
1020 × 678 0.4 0.4 0.2 1.0
(std) ( 0.1 ) ( 0.1 ) ( 0.0 ) ( 0.1 )
720p 0.4 0.5 0.3 1.1
(std) ( 0.0 ) ( 0.0 ) ( 0.1 ) ( 0.1 )
1080p 1.8 0.9 1.2 3.8
(std) ( 0.0 ) ( 0.1 ) ( 0.1 ) ( 0.2 )
Table 3. Quantitative comparison of GGS with 1000 iterations and Instant SFH with different image phases. Results are computed at 70 % sparsity and averaged over 900 images. Numbers in parentheses indicate standard deviation. Bold numbers indicate the best results.
Table 3. Quantitative comparison of GGS with 1000 iterations and Instant SFH with different image phases. Results are computed at 70 % sparsity and averaged over 900 images. Numbers in parentheses indicate standard deviation. Bold numbers indicate the best results.
MethodInitial PhasePSNR (dB)SSIM
GGSRandom36.12 (4.16)0.944 (0.034)
Instant-SFHRandom10.88 (2.29)0.118 (0.064)
Instant-SFHQuadratic10.90 (2.18)0.139 (0.070)
Instant-SFHZeros34.98 (4.24)0.918 (0.050)
Instant-SFHORAP36.02 (4.15)0.943 (0.035)
Instant-SFHPerlin36.19 (4.21)0.945 (0.034)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, D.; Jabbireddy, S.; Zhang, Y.; Metzler, C.; Varshney, A. Instant-SFH: Non-Iterative Sparse Fourier Holograms Using Perlin Noise. Sensors 2024, 24, 7358. https://doi.org/10.3390/s24227358

AMA Style

Li D, Jabbireddy S, Zhang Y, Metzler C, Varshney A. Instant-SFH: Non-Iterative Sparse Fourier Holograms Using Perlin Noise. Sensors. 2024; 24(22):7358. https://doi.org/10.3390/s24227358

Chicago/Turabian Style

Li, David, Susmija Jabbireddy, Yang Zhang, Christopher Metzler, and Amitabh Varshney. 2024. "Instant-SFH: Non-Iterative Sparse Fourier Holograms Using Perlin Noise" Sensors 24, no. 22: 7358. https://doi.org/10.3390/s24227358

APA Style

Li, D., Jabbireddy, S., Zhang, Y., Metzler, C., & Varshney, A. (2024). Instant-SFH: Non-Iterative Sparse Fourier Holograms Using Perlin Noise. Sensors, 24(22), 7358. https://doi.org/10.3390/s24227358

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop