Current Search: Image processing (x)
View All Items
- Title
- Digital Image Processing Using NTEC Facilities.
- Creator
-
Roesch, James F., Simons, Jr., Fred O., Engineering
- Abstract / Description
-
University of Central Florida College of Engineering Thesis; Digital image enhancement refers to the improvement of a given image for human interpretation. Digital image processing facilities are those in which hardware and software computing elements are combined in such a way as to enable the processing of digital images. This report describes the use of the Naval Training Equipment Center (NTEC) Computer Systems Laboratory computing facilities to enhance digital images. Described are two...
Show moreUniversity of Central Florida College of Engineering Thesis; Digital image enhancement refers to the improvement of a given image for human interpretation. Digital image processing facilities are those in which hardware and software computing elements are combined in such a way as to enable the processing of digital images. This report describes the use of the Naval Training Equipment Center (NTEC) Computer Systems Laboratory computing facilities to enhance digital images. Described are two major hardware systems, the IKONAS RDS-3000 raster display graphics system and the VAX-11/780, and the digital image processing program (DIMPRP) written by the author. Digital image enhancement theory and practice are addressed through a discussion of the DIMPRP software. Finally, enhancements to the NTEC digital image processing facility such as improvements in hardware reliability, documentation, and increased speed of program esecution are discussed.
Show less - Date Issued
- 1984
- Identifier
- CFR0008160, ucf:53072
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFR0008160
- Title
- ANALYSIS OF KOLMOGOROV'S SUPERPOSITION THEOREM AND ITS IMPLEMENTATION IN APPLICATIONS WITH LOW AND HIGH DIMENSIONAL DATA.
- Creator
-
Bryant, Donald, Li, Xin, University of Central Florida
- Abstract / Description
-
In this dissertation, we analyze Kolmogorov's superposition theorem for high dimensions. Our main goal is to explore and demonstrate the feasibility of an accurate implementation of Kolmogorov's theorem. First, based on Lorentz's ideas, we provide a thorough discussion on the proof and its numerical implementation of the theorem in dimension two. We present computational experiments which prove the feasibility of the theorem in applications of low dimensions (namely, dimensions...
Show moreIn this dissertation, we analyze Kolmogorov's superposition theorem for high dimensions. Our main goal is to explore and demonstrate the feasibility of an accurate implementation of Kolmogorov's theorem. First, based on Lorentz's ideas, we provide a thorough discussion on the proof and its numerical implementation of the theorem in dimension two. We present computational experiments which prove the feasibility of the theorem in applications of low dimensions (namely, dimensions two and three). Next, we present high dimensional extensions with complete and detailed proofs and provide the implementation that aims at applications with high dimensionality. The amalgamation of these ideas is evidenced by applications in image (two dimensional) and video (three dimensional) representations, the content based image retrieval, video retrieval, de-noising and in-painting, and Bayesian prior estimation of high dimensional data from the fields of computer vision and image processing.
Show less - Date Issued
- 2008
- Identifier
- CFE0002236, ucf:47909
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0002236
- Title
- A METHOD OF CONTENT-BASED IMAGE RETRIEVAL FOR THE GENERATION OF IMAGE MOSAICS.
- Creator
-
Snead, Michael, Richie, Samuel, University of Central Florida
- Abstract / Description
-
An image mosaic is an artistic work that uses a number of smaller images creatively combined together to form another larger image. Each building block image, or tessera, has its own distinctive and meaningful content, but when viewed from a distance the tesserae come together to form an aesthetically pleasing montage. This work presents the design and implementation of MosaiX, a computer software system that generates these image mosaics automatically. To control the image mosaic creation...
Show moreAn image mosaic is an artistic work that uses a number of smaller images creatively combined together to form another larger image. Each building block image, or tessera, has its own distinctive and meaningful content, but when viewed from a distance the tesserae come together to form an aesthetically pleasing montage. This work presents the design and implementation of MosaiX, a computer software system that generates these image mosaics automatically. To control the image mosaic creation process, several parameters are used within the system. Each parameter affects the overall mosaic quality, as well as required processing time, in its own unique way. A detailed analysis is performed to evaluate each parameter individually. Additionally, this work proposes two novel ways by which to evaluate the quality of an image mosaic in a quantitative way. One method focuses on the perceptual color accuracy of the mosaic reproduction, while the other concentrates on edge replication. Both measures include preprocessing to take into account the unique visual features present in an image mosaic. Doing so minimizes quality penalization due the inherent properties of an image mosaic that make them visually appealing.
Show less - Date Issued
- 2007
- Identifier
- CFE0001585, ucf:47115
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0001585
- Title
- OPTIMIZING THE HIGH DYNAMIC RANGE IMAGING PIPELINE.
- Creator
-
Akyuz, Ahmet, Reinhard, Erik, University of Central Florida
- Abstract / Description
-
High dynamic range (HDR) imaging is a rapidly growing field in computer graphics and image processing. It allows capture, storage, processing, and display of photographic information within a scene-referred framework. The HDR imaging pipeline consists of the major steps an HDR image is expected to go through from capture to display. It involves various techniques to create HDR images, pixel encodings and file formats for storage, tone mapping for display on conventional display devices and...
Show moreHigh dynamic range (HDR) imaging is a rapidly growing field in computer graphics and image processing. It allows capture, storage, processing, and display of photographic information within a scene-referred framework. The HDR imaging pipeline consists of the major steps an HDR image is expected to go through from capture to display. It involves various techniques to create HDR images, pixel encodings and file formats for storage, tone mapping for display on conventional display devices and direct display on HDR capable screens. Each of these stages have important open problems, which need to be addressed for a smoother transition to an HDR imaging pipeline. We addressed some of these important problems such as noise reduction in HDR imagery, preservation of color appearance, validation of tone mapping operators, and image display on HDR monitors. The aim of this thesis is thus, to present our findings and describe the research we have conducted within the framework of optimizing the HDR imaging pipeline.
Show less - Date Issued
- 2007
- Identifier
- CFE0001875, ucf:47404
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0001875
- Title
- Tessellation for computer image generation.
- Creator
-
Panzitta, Michael James, Bauer, Christian S., Engineering
- Abstract / Description
-
University of Central Florida College of Engineering Thesis; Of the vast number of algorithms used in modern computer image generation, most rely upon data bases comprised of polygons. This becomes a severe impediment when curved objects must be modeled and displayed with an acceptable level of speed and accuracy. A technique is needed to provide a means of modeling curved surfaces, storing them in a data base, and displaying them using existing algorithms. Tessellation is one method of...
Show moreUniversity of Central Florida College of Engineering Thesis; Of the vast number of algorithms used in modern computer image generation, most rely upon data bases comprised of polygons. This becomes a severe impediment when curved objects must be modeled and displayed with an acceptable level of speed and accuracy. A technique is needed to provide a means of modeling curved surfaces, storing them in a data base, and displaying them using existing algorithms. Tessellation is one method of achieving such goals.
Show less - Date Issued
- 1987
- Identifier
- CFR0001375, ucf:52922
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFR0001375
- Title
- An investigation into a least squares method for image registration.
- Creator
-
Cordon, Ernest William, Patz, B.W., Engineering
- Abstract / Description
-
Florida Technological University College of Engineering Thesis; One of the problems associated with the automatic image processing of satellite photographs such as weather maps is the need for image registration; that is, the fitting of a map that has some translational and rotational bias to a known data base. This paper investigates a least square method of image registration using an image that has been converted into a boundary map with a pixel representation 1 for land, -1 for water and...
Show moreFlorida Technological University College of Engineering Thesis; One of the problems associated with the automatic image processing of satellite photographs such as weather maps is the need for image registration; that is, the fitting of a map that has some translational and rotational bias to a known data base. This paper investigates a least square method of image registration using an image that has been converted into a boundary map with a pixel representation 1 for land, -1 for water and zero for cloud pixels. A sampled correlation array is constructed by shifting the weather map to locations on a given grid, centered around a sampled correlation peak, and performing an accumulation of the pixel-by-pixel comparisons between the weather map and its data base over the whole map or a smaller search window. A least square approximation 0 f the translational and rotational bias is performed using the data from this sampled correlation array, fitted against the shape of an elliptical cone.
Show less - Date Issued
- 1978
- Identifier
- CFR0003516, ucf:53005
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFR0003516
- Title
- Realization of a fast automatic correlation algorithm for registration of satellite images.
- Creator
-
Kassak, John E., Patz, B.W., Engineering
- Abstract / Description
-
Florida Technological University College of Engineering Thesis; The requirement for a fast automated correlation algorithm for registration of satellite images is discussed. An overview of current registration techniques is presented indicating a correlator, matching binary maps compressed from the original imagery, may provide the required throughput when implemented with a dedicated hardware/processor. An actual registration problem utilizing GOES digitally processed imagery is chosen and...
Show moreFlorida Technological University College of Engineering Thesis; The requirement for a fast automated correlation algorithm for registration of satellite images is discussed. An overview of current registration techniques is presented indicating a correlator, matching binary maps compressed from the original imagery, may provide the required throughput when implemented with a dedicated hardware/processor. An actual registration problem utilizing GOES digitally processed imagery is chosen and defined. The realization of a fast correlator, matching image input data with sampled data base reference image data in real time is considered.
Show less - Date Issued
- 1978
- Identifier
- CFR0003495, ucf:53010
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFR0003495
- Title
- AUTONOMOUS ROBOTIC AUTOMATION SYSTEMWITH VISION FEEDBACK.
- Creator
-
Rosino, Jeffery, Qu, Zhihua, University of Central Florida
- Abstract / Description
-
In this thesis, a full design, development and application of an autonomous robotic automation system using vision feedback is performed. To realize this system, a cylindrical manipulator configuration is implemented, using a personal computer (PC) based PID controller from National Instruments. Full autonomous control will be achieved via a programmable human machine interface (HMI) developed on a PC using Borland C++ Builder. The vision feedback position control is accomplished using an...
Show moreIn this thesis, a full design, development and application of an autonomous robotic automation system using vision feedback is performed. To realize this system, a cylindrical manipulator configuration is implemented, using a personal computer (PC) based PID controller from National Instruments. Full autonomous control will be achieved via a programmable human machine interface (HMI) developed on a PC using Borland C++ Builder. The vision feedback position control is accomplished using an ordinary "off-the-shelf" web camera. The manuscript is organized as follows; After Chapter 1, an introduction to automation history and its role in the manufacturing industry, Chapter 2 discusses and outlines the development of the robotic kinematics and dynamics of the system. A control strategy is also developed and simulated in this chapter. Chapter 3 discusses color image processing and shows the development of the algorithm used for the vision feedback position control. Chapter 4 outlines the system development, which includes the hardware and software. Chapter 5 concludes with a summary, and improvement section. The process used as a basis for the design and development of this thesis of this thesis topic was constructed from a manual capacitor orientation check test station. A more detailed definition and objective is presented in the introduction.
Show less - Date Issued
- 2004
- Identifier
- CFE0000277, ucf:46220
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0000277
- Title
- Computerized Evaluatution of Microsurgery Skills Training.
- Creator
-
Jotwani, Payal, Foroosh, Hassan, Hughes, Charles, Hua, Kien, University of Central Florida
- Abstract / Description
-
The style of imparting medical training has evolved, over the years. The traditional methods of teaching and practicing basic surgical skills under apprenticeship model, no longer occupy the first place in modern technically demanding advanced surgical disciplines like neurosurgery. Furthermore, the legal and ethical concerns for patient safety as well as cost-effectiveness have forced neurosurgeons to master the necessary microsurgical techniques to accomplish desired results. This has lead...
Show moreThe style of imparting medical training has evolved, over the years. The traditional methods of teaching and practicing basic surgical skills under apprenticeship model, no longer occupy the first place in modern technically demanding advanced surgical disciplines like neurosurgery. Furthermore, the legal and ethical concerns for patient safety as well as cost-effectiveness have forced neurosurgeons to master the necessary microsurgical techniques to accomplish desired results. This has lead to increased emphasis on assessment of clinical and surgical techniques of the neurosurgeons. However, the subjective assessment of microsurgical techniques like micro-suturing under the apprenticeship model cannot be completely unbiased. A few initiatives using computer-based techniques, have been made to introduce objective evaluation of surgical skills.This thesis presents a novel approach involving computerized evaluation of different components of micro-suturing techniques, to eliminate the bias of subjective assessment. The work involved acquisition of cine clips of micro-suturing activity on synthetic material. Image processing and computer vision based techniques were then applied to these videos to assess different characteristics of micro-suturing viz. speed, dexterity and effectualness. In parallel subjective grading on these was done by a senior neurosurgeon. Further correlation and comparative study of both the assessments was done to analyze the efficacy of objective and subjective evaluation.
Show less - Date Issued
- 2015
- Identifier
- CFE0006221, ucf:51056
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0006221
- Title
- Applied Advanced Error Control Coding for General Purpose Representation and Association Machine Systems.
- Creator
-
Dai, Bowen, Wei, Lei, Lin, Mingjie, Rahnavard, Nazanin, Turgut, Damla, Sun, Qiyu, University of Central Florida
- Abstract / Description
-
General-Purpose Representation and Association Machine (GPRAM) is proposed to be focusing on computations in terms of variation and flexibility, rather than precision and speed. GPRAM system has a vague representation and has no predefined tasks. With several important lessons learned from error control coding, neuroscience and human visual system, we investigate several types of error control codes, including Hamming code and Low-Density Parity Check (LDPC) codes, and extend them to...
Show moreGeneral-Purpose Representation and Association Machine (GPRAM) is proposed to be focusing on computations in terms of variation and flexibility, rather than precision and speed. GPRAM system has a vague representation and has no predefined tasks. With several important lessons learned from error control coding, neuroscience and human visual system, we investigate several types of error control codes, including Hamming code and Low-Density Parity Check (LDPC) codes, and extend them to different directions.While in error control codes, solely XOR logic gate is used to connect different nodes. Inspired by bio-systems and Turbo codes, we suggest and study non-linear codes with expanded operations, such as codes including AND and OR gates which raises the problem of prior-probabilities mismatching. Prior discussions about critical challenges in designing codes and iterative decoding for non-equiprobable symbols may pave the way for a more comprehensive understanding of bio-signal processing. The limitation of XOR operation in iterative decoding with non-equiprobable symbols is described and can be potentially resolved by applying quasi-XOR operation and intermediate transformation layer. Constructing codes for non-equiprobable symbols with the former approach cannot satisfyingly perform with regarding to error correction capability. Probabilistic messages for sum-product algorithm using XOR, AND, and OR operations with non-equiprobable symbols are further computed. The primary motivation for the constructing codes is to establish the GPRAM system rather than to conduct error control coding per se. The GPRAM system is fundamentally developed by applying various operations with substantial over-complete basis. This system is capable of continuously achieving better and simpler approximations for complex tasks.The approaches of decoding LDPC codes with non-equiprobable binary symbols are discussed due to the aforementioned prior-probabilities mismatching problem. The traditional Tanner graph should be modified because of the distinction of message passing to information bits and to parity check bits from check nodes. In other words, the message passing along two directions are identical in conventional Tanner graph, while the message along the forward direction and backward direction are different in our case. A method of optimizing signal constellation is described, which is able to maximize the channel mutual information.A simple Image Processing Unit (IPU) structure is proposed for GPRAM system, to which images are inputted. The IPU consists of a randomly constructed LDPC code, an iterative decoder, a switch, and scaling and decision device. The quality of input images has been severely deteriorated for the purpose of mimicking visual information variability (VIV) experienced in human visual systems. The IPU is capable of (a) reliably recognizing digits from images of which quality is extremely inadequate; (b) achieving similar hyper-acuity performance comparing to human visual system; and (c) significantly improving the recognition rate with applying randomly constructed LDPC code, which is not specifically optimized for the tasks.
Show less - Date Issued
- 2016
- Identifier
- CFE0006449, ucf:51413
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0006449
- Title
- Analytical study of computer vision-based pavement crack quantification using machine learning techniques.
- Creator
-
Mokhtari, Soroush, Yun, Hae-Bum, Nam, Boo Hyun, Catbas, Necati, Shah, Mubarak, Xanthopoulos, Petros, University of Central Florida
- Abstract / Description
-
Image-based techniques are a promising non-destructive approach for road pavement condition evaluation. The main objective of this study is to extract, quantify and evaluate important surface defects, such as cracks, using an automated computer vision-based system to provide a better understanding of the pavement deterioration process. To achieve this objective, an automated crack-recognition software was developed, employing a series of image processing algorithms of crack extraction, crack...
Show moreImage-based techniques are a promising non-destructive approach for road pavement condition evaluation. The main objective of this study is to extract, quantify and evaluate important surface defects, such as cracks, using an automated computer vision-based system to provide a better understanding of the pavement deterioration process. To achieve this objective, an automated crack-recognition software was developed, employing a series of image processing algorithms of crack extraction, crack grouping, and crack detection. Bottom-hat morphological technique was used to remove the random background of pavement images and extract cracks, selectively based on their shapes, sizes, and intensities using a relatively small number of user-defined parameters. A technical challenge with crack extraction algorithms, including the Bottom-hat transform, is that extracted crack pixels are usually fragmented along crack paths. For de-fragmenting those crack pixels, a novel crack-grouping algorithm is proposed as an image segmentation method, so called MorphLink-C. Statistical validation of this method using flexible pavement images indicated that MorphLink-C not only improves crack-detection accuracy but also reduces crack detection time.Crack characterization was performed by analysing imagerial features of the extracted crack image components. A comprehensive statistical analysis was conducted using filter feature subset selection (FSS) methods, including Fischer score, Gini index, information gain, ReliefF, mRmR, and FCBF to understand the statistical characteristics of cracks in different deterioration stages. Statistical significance of crack features was ranked based on their relevancy and redundancy. The statistical method used in this study can be employed to avoid subjective crack rating based on human visual inspection. Moreover, the statistical information can be used as fundamental data to justify rehabilitation policies in pavement maintenance.Finally, the application of four classification algorithms, including Artificial Neural Network (ANN), Decision Tree (DT), k-Nearest Neighbours (kNN) and Adaptive Neuro-Fuzzy Inference System (ANFIS) is investigated for the crack detection framework. The classifiers were evaluated in the following five criteria: 1) prediction performance, 2) computation time, 3) stability of results for highly imbalanced datasets in which, the number of crack objects are significantly smaller than the number of non-crack objects, 4) stability of the classifiers performance for pavements in different deterioration stages, and 5) interpretability of results and clarity of the procedure. Comparison results indicate the advantages of white-box classification methods for computer vision based pavement evaluation. Although black-box methods, such as ANN provide superior classification performance, white-box methods, such as ANFIS, provide useful information about the logic of classification and the effect of feature values on detection results. Such information can provide further insight for the image-based pavement crack detection application.
Show less - Date Issued
- 2015
- Identifier
- CFE0005671, ucf:50186
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0005671
- Title
- Vision-Based Sensing and Optimal Control for Low-Cost and Small Satellite Platforms.
- Creator
-
Sease, Bradley, Xu, Yunjun, Lin, Kuo-Chi, Bradley, Eric, University of Central Florida
- Abstract / Description
-
Current trends in spacecraft are leading to smaller, more inexpensive options whenever possible. This shift has been primarily pursued for the opportunity to open a new frontier for technologies with a small financial obligation. Limited power, processing, pointing, and communication capabilities are all common issues which must be considered when miniaturizing systems and implementing low-cost components. This thesis addresses some of these concerns by applying two methods, in attitude...
Show moreCurrent trends in spacecraft are leading to smaller, more inexpensive options whenever possible. This shift has been primarily pursued for the opportunity to open a new frontier for technologies with a small financial obligation. Limited power, processing, pointing, and communication capabilities are all common issues which must be considered when miniaturizing systems and implementing low-cost components. This thesis addresses some of these concerns by applying two methods, in attitude estimation and control. Additionally, these methods are not restricted to only small, inexpensive satellites, but offer a benefit to large-scale spacecraft as well.First, star cameras are examined for the tendency to generate streaked star images during maneuvers. This issue also comes into play when pointing capabilities and camera hardware quality are low, as is often the case in small, budget-constrained spacecraft. When pointing capabilities are low, small residual velocities can cause movement of the stars in the focal plane during an exposure, causing them to streak across the image. Additionally, if the camera quality is low, longer exposures may be required to gather sufficient light from a star, further contributing to streaking. Rather than improving the pointing or hardware directly, an algorithm is presented to retrieve and utilize the endpoints of streaked stars to provide feedback where traditional methods do not. This allows precise attitude and angular rate estimates to be derived from an image which, with traditional methods, would return large attitude and rate error. Simulation results are presented which demonstrate endpoint error of approximately half a pixel and rate estimates within 2% of the true angular velocity. Three methods are also considered to remove overlapping star streaks and resident space objects from images to improve performance of both attitude and rate estimates. Results from a large-scale Monte Carlo simulation are presented in order to characterize the performance of the method.Additionally, a rapid optimal attitude guidance method is experimentally validated in a ground-based, pico-scale satellite test bed. Fast slewing performance is demonstrated for an incremental step maneuver with low average power consumption. Though the focus of this thesis is primarily on increasing the capabilities of small, inexpensive spacecraft, the methods discussed have the potential to increase the capabilities of current and future large-scale missions as well.
Show less - Date Issued
- 2013
- Identifier
- CFE0005249, ucf:50603
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0005249
- Title
- Harnessing Spatial Intensity Fluctuations for Optical Imaging and Sensing.
- Creator
-
Akhlaghi Bouzan, Milad, Dogariu, Aristide, Saleh, Bahaa, Pang, Sean, Atia, George, University of Central Florida
- Abstract / Description
-
Properties of light such as amplitude and phase, temporal and spatial coherence, polarization, etc. are abundantly used for sensing and imaging. Regardless of the passive or active nature of the sensing method, optical intensity fluctuations are always present! While these fluctuations are usually regarded as noise, there are situations where one can harness the intensity fluctuations to enhance certain attributes of the sensing procedure. In this thesis, we developed different sensing...
Show moreProperties of light such as amplitude and phase, temporal and spatial coherence, polarization, etc. are abundantly used for sensing and imaging. Regardless of the passive or active nature of the sensing method, optical intensity fluctuations are always present! While these fluctuations are usually regarded as noise, there are situations where one can harness the intensity fluctuations to enhance certain attributes of the sensing procedure. In this thesis, we developed different sensing methodologies that use statistical properties of optical fluctuations for gauging specific information. We examine this concept in the context of three different aspects of computational optical imaging and sensing. First, we study imposing specific statistical properties to the probing field to image or characterize certain properties of an object through a statistical analysis of the spatially integrated scattered intensity. This offers unique capabilities for imaging and sensing techniques operating in highly perturbed environments and low-light conditions. Next, we examine optical sensing in the presence of strong perturbations that preclude any controllable field modification. We demonstrate that inherent properties of diffused coherent fields and fluctuations of integrated intensity can be used to track objects hidden behind obscurants. Finally, we address situations where, due to coherent noise, image accuracy is severely degraded by intensity fluctuations. By taking advantage of the spatial coherence properties of optical fields, we show that this limitation can be effectively mitigated and that a significant improvement in the signal-to-noise ratio can be achieved even in one single-shot measurement. The findings included in this dissertation illustrate different circumstances where optical fluctuations can affect the efficacy of computational optical imaging and sensing. A broad range of applications, including biomedical imaging and remote sensing, could benefit from the new approaches to suppress, enhance, and exploit optical fluctuations, which are described in this dissertation.
Show less - Date Issued
- 2017
- Identifier
- CFE0007274, ucf:52200
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0007274
- Title
- Video categorization using semantics and semiotics.
- Creator
-
Rasheed, Zeeshan, Shah, Mubarak, Engineering and Computer Science
- Abstract / Description
-
University of Central Florida College of Engineering Thesis; There is a great need to automatically segment, categorize, and annotate video data, and to develop efficient tools for browsing and searching. We believe that the categorization of videos can be achieved by exploring the concepts and meanings of the videos. This task requires bridging the gap between low-level content and high-level concepts (or semantics). Once a relationship is established between the low-level computable...
Show moreUniversity of Central Florida College of Engineering Thesis; There is a great need to automatically segment, categorize, and annotate video data, and to develop efficient tools for browsing and searching. We believe that the categorization of videos can be achieved by exploring the concepts and meanings of the videos. This task requires bridging the gap between low-level content and high-level concepts (or semantics). Once a relationship is established between the low-level computable features of the video and its semantics, .the user would be able to navigate through videos through the use of concepts and ideas (for example, a user could extract only those scenes in an action film that actually contain fights) rat her than sequentially browsing the whole video. However, this relationship must follow the norms of human perception and abide by the rules that are most often followed by the creators (directors) of these videos. These rules are called film grammar in video production literature. Like any natural language, this grammar has several dialects, but it has been acknowledged to be universal. Therefore, the knowledge of film grammar can be exploited effectively for the understanding of films. To interpret an idea using the grammar, we need to first understand the symbols, as in natural languages, and second, understand the rules of combination of these symbols to represent concepts. In order to develop algorithms that exploit this film grammar, it is necessary to relate the symbols of the grammar to computable video features.
Show less - Date Issued
- 2003
- Identifier
- CFR0001717, ucf:52920
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFR0001717
- Title
- Sampling and Subspace Methods for Learning Sparse Group Structures in Computer Vision.
- Creator
-
Jaberi, Maryam, Foroosh, Hassan, Pensky, Marianna, Gong, Boqing, Qi, GuoJun, Pensky, Marianna, University of Central Florida
- Abstract / Description
-
The unprecedented growth of data in volume and dimension has led to an increased number of computationally-demanding and data-driven decision-making methods in many disciplines, such as computer vision, genomics, finance, etc. Research on big data aims to understand and describe trends in massive volumes of high-dimensional data. High volume and dimension are the determining factors in both computational and time complexity of algorithms. The challenge grows when the data are formed of the...
Show moreThe unprecedented growth of data in volume and dimension has led to an increased number of computationally-demanding and data-driven decision-making methods in many disciplines, such as computer vision, genomics, finance, etc. Research on big data aims to understand and describe trends in massive volumes of high-dimensional data. High volume and dimension are the determining factors in both computational and time complexity of algorithms. The challenge grows when the data are formed of the union of group-structures of different dimensions embedded in a high-dimensional ambient space.To address the problem of high volume, we propose a sampling method referred to as the Sparse Withdrawal of Inliers in a First Trial (SWIFT), which determines the smallest sample size in one grab so that all group-structures are adequately represented and discovered with high probability. The key features of SWIFT are: (i) sparsity, which is independent of the population size; (ii) no prior knowledge of the distribution of data, or the number of underlying group-structures; and (iii) robustness in the presence of an overwhelming number of outliers. We report a comprehensive study of the proposed sampling method in terms of accuracy, functionality, and effectiveness in reducing the computational cost in various applications of computer vision. In the second part of this dissertation, we study dimensionality reduction for multi-structural data. We propose a probabilistic subspace clustering method that unifies soft- and hard-clustering in a single framework. This is achieved by introducing a delayed association of uncertain points to subspaces of lower dimensions based on a confidence measure. Delayed association yields higher accuracy in clustering subspaces that have ambiguities, i.e. due to intersections and high-level of outliers/noise, and hence leads to more accurate self-representation of underlying subspaces. Altogether, this dissertation addresses the key theoretical and practically issues of size and dimension in big data analysis.
Show less - Date Issued
- 2018
- Identifier
- CFE0007017, ucf:52039
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0007017