Clip space sample culling for motion blur and defocus blur

Yi Jeng Wu, Der Lor Way, Yu Ting Tsai, Zen-Chung Shih

Research output: Contribution to journalArticlepeer-review

2 Scopus citations

Abstract

Motion blur and defocus blur are two common visual effects for rendering realistic camera images. This paper presents a novel clip space culling for stochastic rasterization to render motion and defocus blur effects. Our proposed algorithm reduces the sample coverage using the clip space information in camera lens domain (UV) and time domain (T). First, samples outside the camera lens were culled in stage I using the linear relationship between camera lens and vertex position. Second, samples outside the time bounds were culled in stage II using the triangle similarity in clip space to find the intersection time. Each pixel was computed within two linear bounds only once. Our method achieves good sample test efficiency with low computation cost for real-time stochastic rasterization. Finally, the proposed method is demonstrated by means of various experiments, and a comparison is made with previous works. Our algorithm was able to handle these two blur effects simultaneously and performed better than others did.

Original languageEnglish
Pages (from-to)1071-1084
Number of pages14
JournalJournal of Information Science and Engineering
Volume31
Issue number3
DOIs
StatePublished - 1 May 2015

Keywords

  • Defocus blur
  • Focal depth
  • Motion blur
  • Sample test efficiency (STE)
  • Stochastic rasterization

Fingerprint Dive into the research topics of 'Clip space sample culling for motion blur and defocus blur'. Together they form a unique fingerprint.

Cite this