CFP last date
20 May 2024
Call for Paper
June Edition
IJCA solicits high quality original research papers for the upcoming June edition of the journal. The last date of research paper submission is 20 May 2024

Submit your paper
Know more
Reseach Article

An Optimised Approach on Object and Text Detection from Real Time Data using Histogram Equalization Technique

by Monika Kapoor, Saurabh Sharma
International Journal of Computer Applications
Foundation of Computer Science (FCS), NY, USA
Volume 177 - Number 32
Year of Publication: 2020
Authors: Monika Kapoor, Saurabh Sharma
10.5120/ijca2020919779

Monika Kapoor, Saurabh Sharma . An Optimised Approach on Object and Text Detection from Real Time Data using Histogram Equalization Technique. International Journal of Computer Applications. 177, 32 ( Jan 2020), 14-20. DOI=10.5120/ijca2020919779

@article{ 10.5120/ijca2020919779,
author = { Monika Kapoor, Saurabh Sharma },
title = { An Optimised Approach on Object and Text Detection from Real Time Data using Histogram Equalization Technique },
journal = { International Journal of Computer Applications },
issue_date = { Jan 2020 },
volume = { 177 },
number = { 32 },
month = { Jan },
year = { 2020 },
issn = { 0975-8887 },
pages = { 14-20 },
numpages = {9},
url = { https://ijcaonline.org/archives/volume177/number32/31107-2020919779/ },
doi = { 10.5120/ijca2020919779 },
publisher = {Foundation of Computer Science (FCS), NY, USA},
address = {New York, USA}
}
%0 Journal Article
%1 2024-02-07T00:47:30.070646+05:30
%A Monika Kapoor
%A Saurabh Sharma
%T An Optimised Approach on Object and Text Detection from Real Time Data using Histogram Equalization Technique
%J International Journal of Computer Applications
%@ 0975-8887
%V 177
%N 32
%P 14-20
%D 2020
%I Foundation of Computer Science (FCS), NY, USA
Abstract

The initiation of artificial intelligence, gave a curious approach towards the artificial neural networks. The developers are trying to make computers think and perform tasks just like humans. Artificial Neural Networks was developed with the view in mind to make computers to do so. The machines that are to be used in the field of robotics, medicines, industry needs to be so smart that they should be able to perform the day to day tasks easily without needing the human help. Take the case of self-driven cars, it is very crucial for the AI to make out the scenario in the real time and act accordingly take the necessary actions like breaking or evading in the traffics. Since the introduction of the object detection by the AI Researchers at Facebook, it has become easier for us to identify the objects in the images. While the object identifiers can use the image identification in the still images but there is a need to identify the text in the images. The techniques can be used for identification of the objects and texts in the real time scenario also. The present work here shows that image identification can work not only for the object identification but also identifies the text in the images as well. A CLAHE algorithm is developed to identify these objects in the images and text and classify these entities into the categories as desired by the programmer. Here in this paper an attempt is made to show the working of an algorithm based on the EAST text detector and PMTD that can perform object as well as text identification in the real time itself. The algorithm achieved success in achieving the identification of the objects in under 5 seconds and better yet identifying text in under 1.5 seconds. The developed algorithm outperforms its earlier precursors. The Recall, Precision and F-measure values all were found better than both the previous algorithms.

References
  1. G. F. C. Campos, S. M. Mastelini, G. J. Aguiar, R. G. Mantovani, L. F. de Melo, and S. Barbon, “Machine learning hyperparameter selection for Contrast Limited Adaptive Histogram Equalization,” Eurasip J. Image Video Process., vol. 2019, no. 1, 2019.
  2. X. Sun, P. Wu, and S. C. H. Hoi, “Face detection using deep learning: An improved faster RCNN approach,” Neurocomputing, vol. 299, pp. 42–50, 2018.
  3. V. Badrinarayanan, A. Kendall, and R. Cipolla, “SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 39, no. 12, pp. 2481–2495, 2017.
  4. X. Zhou et al., “EAST: An efficient and accurate scene text detector,” Proc. - 30th IEEE Conf. Comput. Vis. Pattern Recognition, CVPR 2017, vol. 2017-Janua, pp. 2642–2651, 2017.
  5. Y. Guo, Y. Liu, A. Oerlemans, S. Lao, S. Wu, and M. S. Lew, “Deep learning for visual understanding: A review,” Neurocomputing, vol. 187, pp. 27–48, 2016.
  6. O. Russakovsky et al., “ImageNet Large Scale Visual Recognition Challenge,” Int. J. Comput. Vis., vol. 115, no. 3, pp. 211–252, 2015.
  7. M. K. Chauhan and G. Kumar, “Automatic Text Detection and Information Retrieval on Mobile,” vol. 5, no. 6, pp. 8285–8292, 2014.
  8. P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. LeCun, “OverFeat: Integrated Recognition, Localization and Detection using Convolutional Networks,” 2013.
  9. R. Socher, “Recursive Deep Models for Semantic Compositionality,” no. October, pp. 1631–1642, 2013.
  10. S. Bell, P. Upchurch, N. Snavely, and K. Bala, “OPENSURFACES: A richly annotated catalog of surface appearance,” ACM Trans. Graph., vol. 32, no. 4, 2013.
  11. F. Mokhtarian, S. Abbasi, and J. Kittler, “Robust and Efficient Shape Indexing through Curvature Scale Space,” pp. 33.1-33.10, 2013.
  12. P. Dollár, C. Wojek, B. Schiele, and P. Perona, “Pedestrian detection: An evaluation of the state of the art,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 34, no. 4, pp. 743–761, 2012.
  13. A. Krizhevsky and G. Hinton, “ImageNet Classification with Deep Convolutional Neural Networks (presentation),” ImageNet Large Scale Vis. Recognit. Chall. 2012, p. 27, 2012.
  14. G. Patterson and J. Hays, “SUN attribute database: Discovering, annotating, and recognizing scene attributes,” Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., pp. 2751–2758, 2012.
  15. D. Hoiem, Y. Chodpathumwan, and Q. Dai, “Diagnosing error in object detectors,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 7574 LNCS, no. PART 3, pp. 340–353, 2012.
  16. J. Xiao, J. Hays, K. A. Ehinger, A. Oliva, and A. Torralba, “SUN database: Large-scale scene recognition from abbey to zoo,” Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., pp. 3485–3492, 2010.
  17. M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman, “The pascal visual object classes (VOC) challenge,” Int. J. Comput. Vis., vol. 88, no. 2, pp. 303–338, 2010.
  18. P. Nagabhushan and S. Nirmala, “Text Extraction in Complex Color Document Images for Enhanced Readability,” Intell. Inf. Manag., vol. 02, no. 02, pp. 120–133, 2010.
  19. A. Farhadi, I. Endres, D. Hoiem, and D. Forsyth, “Describing objects by their attributes_2009 IEEE Conference on Computer Vision and Pattern Recognition_2009_Farhadi et al.pdf.”
  20. G. J. Brostow, J. Fauqueur, and R. Cipolla, “Semantic object classes in video: A high-definition ground truth database,” Pattern Recognit. Lett., vol. 30, no. 2, pp. 88–97, 2009.
  21. B. C. Russell, A. Torralba, K. P. Murphy, and W. T. Freeman, “LabelMe: A database and web-based tool for image annotation,” Int. J. Comput. Vis., vol. 77, no. 1–3, pp. 157–173, 2008.
  22. K. Jansen and H. Zhang, “Scheduling malleable tasks,” Handb. Approx. Algorithms Metaheuristics, pp. 45-1-45–16, 2007.
  23. N. Chen, “A Survey of Indexing and Retrieval of Multimodal Documents : Text and Images,” no. February, p. 40, 2006.
  24. D. Das, D. Chen, and A. G. Hauptmann, “Improving Multimedia Retrieval with a Video OCR,” 2006.
  25. D. Chen, J. M. Odobez, and H. Bourlard, “Text detection and recognition in images and video frames,” Pattern Recognit., vol. 37, no. 3, pp. 595–608, 2004.
  26. R. Lienhart and A. Wernicke, “Localizing and segmenting text in images and videos,” IEEE Trans. Circuits Syst. Video Technol., vol. 12, no. 4, pp. 256–268, 2002.
  27. C. Garcia and X. Apostolidis, “Text detection and segmentation in complex color images,” ICASSP, IEEE Int. Conf. Acoust. Speech Signal Process. - Proc., vol. 4, pp. 2326–2329, 2000.
  28. S. E. Watson and A. F. Kramer, “Object-based visual selective attention and perceptual organization,” Percept. Psychophys., vol. 61, no. 1, pp. 31–49, 1999.
  29. K. Sobottka, H. Bunke, and H. Kronenberg, “Identification of text on colored book and journal covers,” Proc. Int. Conf. Doc. Anal. Recognition, ICDAR, pp. 57–62, 1999.
  30. H. Li and D. Doermann, “Text enhancement in digital video using multiple frame integration,” Proc. ACM Int. Multimed. Conf. Exhib., pp. 19–22, 1999.
  31. B. S. Manjunath, “Texture features for browsing and retrieval of image data,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 18, no. 8, pp. 837–842, 1996.
  32. L. Eikvil, “Line Eikvil, OCR- Optical Character Recognition., December 1993..pdf,” no. December, 1993.
  33. M. H. Brill, “Computer Vision and Pattern Recognition: CVPR 92,” Color Res. Appl., vol. 17, no. 6, pp. 426–427, 1992.
  34. M. J. Swain and D. H. Ballard, “Color indexing,” Int. J. Comput. Vis., vol. 7, no. 1, pp. 11–32, 1991.
  35. I. Bose, A. R. Jana, and S. Chatterjee, “A BCS Theory ot Superconductivity in Heavy Fermion Systems,” Phys. Status Solidi, vol. 136, no. 1, pp. 387–392, 1986.
  36. L. S. Vishnu and S. Rao, “Object Recognition and Object Counting using CNNs,” no. 12376, 1237.
Index Terms

Computer Science
Information Sciences

Keywords

ANN AI PMTD EAST CLAHE Text Classification Image Identification Object Classification.