Autofocusing System using Matching Blurry measure and Working Distance for industrial application
Main Article Content
Abstract
- This paper presents a novel autofocus algorithm in which the type of target object is fixed and the range of working distance is known a priori. The key idea of the algorithm is to compare the blurry value computed from the input image to the ones pre-recorded in the learning step. The blurry value is derived from the Point Spread Function (PSF) estimated from the input image. The information on working distance associated to the reference blurry value is used to adjust the lens position to the focus position. This algorithm requires at most three shots (initial shot and two steps of lens position adjustment) to reach the focus position. From our experiment, the accuracy of our autofocus algorithm is about 93%. The algorithm can be applied for the autofocus module of an industrial visual inspection, especially in the area of high magnification applications.
Article Details
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
I/we certify that I/we have participated sufficiently in the intellectual content, conception and design of this work or the analysis and interpretation of the data (when applicable), as well as the writing of the manuscript, to take public responsibility for it and have agreed to have my/our name listed as a contributor. I/we believe the manuscript represents valid work. Neither this manuscript nor one with substantially similar content under my/our authorship has been published or is being considered for publication elsewhere, except as described in the covering letter. I/we certify that all the data collected during the study is presented in this manuscript and no data from the study has been or will be published separately. I/we attest that, if requested by the editors, I/we will provide the data/information or will cooperate fully in obtaining and providing the data/information on which the manuscript is based, for examination by the editors or their assignees. Financial interests, direct or indirect, that exist or may be perceived to exist for individual contributors in connection with the content of this paper have been disclosed in the cover letter. Sources of outside support of the project are named in the cover letter.
I/We hereby transfer(s), assign(s), or otherwise convey(s) all copyright ownership, including any and all rights incidental thereto, exclusively to the Journal, in the event that such work is published by the Journal. The Journal shall own the work, including 1) copyright; 2) the right to grant permission to republish the article in whole or in part, with or without fee; 3) the right to produce preprints or reprints and translate into languages other than English for sale or free distribution; and 4) the right to republish the work in a collection of articles in any other mechanical or electronic format.
We give the rights to the corresponding author to make necessary changes as per the request of the journal, do the rest of the correspondence on our behalf and he/she will act as the guarantor for the manuscript on our behalf.
All persons who have made substantial contributions to the work reported in the manuscript, but who are not contributors, are named in the Acknowledgment and have given me/us their written permission to be named. If I/we do not include an Acknowledgment that means I/we have not received substantial contributions from non-contributors and no contributor has been omitted.
References
2. C.Y. Chen, R.C. Hwang and Y.J. Chen, “A passive auto-focus camera control system,” Applied Soft Computing, vol. 10, no. 1, pp. 296–303, 2010.
3. S. Jin, J. Cho, K.H. Kwon and J.W. Jeon, "A Dedicated Hardware Architecture for Real-Time Auto-Focusing using an FPGA," Machine Vision and Applications, vol. 21, no. 5, pp. 727-734, 2010.
4. L. C. Chiu and C. S. Fuh, “An efficient auto focus method for digital still camera based on focus value curve prediction model,” Journal of Information Science and Engineering, vol.26, no. 4, pp. 1261-1272, 2010.
5. S. Allegro, C. Chanel and J. Jacot, "Autofocus for Automated Microassembly Under a Microscope", IEEE Int. Conf. on Image Processing, vol. 1, pp. 677-680, 1996.
6. G. Duceux, B. Tamadazte, E. Marchand, N. Le-Fort Piat and G. Fortier, “Autofocusing-Based Visual Servoing : Application to MEMS micromanipulation,” International Symposium on Optomechatronic Technologies, 2010.
7. P. T. Yap and P. Raveendran, "Image focus measure based on chebyshev moments," IEE Proceedings - Vision, Image and Signal Processing, vol. 151, no. 2, pp. 128-136, 2004.
8. J. Kautsky, J. Flusser, B. Zitová and S. Šimberová, " A new wavelet-based measure of image focus," Pattern Recognition Letters, vol. 23, pp. 1785-1794, 2002.
9. Y. Zhang, Y. Zhang and C. Wen, “A new focus measure method using moments,” Image and Vision Computing, vol. 18, pp. 959–965, 2000.
10. R. Fergus, B. Singh, A. Hertzmann, S.T. Roweis and W.T. Freeman, “Removing camera shake from a single photograph,” SIGGRAPH, vol. 25, no. 3, pp. 787-794, 2006.
11. Q. Shan, J. Jia and A. Agarwala, "High-Quality Motion Deblurring From a Single Image," ACM Transactions on Graphics, vol. 27, no. 3, pp. 73:1-73:10, Aug. 2008.
12. S.M. Smith and J.M. Brady, “SUSAN-a new approach to low level image processing,” Journal of Computer Vision, vol. 23, pp. 45-78, 1997