Current Search: Hua, Kien A. (x)
View All Items
- Title
- INTERACTIVITY AND USER-HETEROGENEITY IN ON DEMAND BROADCAST VIDEO.
- Creator
-
Tantaoui El Araki, Mounir, Hua, Kien A., University of Central Florida
- Abstract / Description
-
Video-On-Demand (VOD) has appeared as an important technology for many multimedia applications such as news on demand, digital libraries, home entertainment, and distance learning. In its simplest form, delivery of a video stream requires a dedicated channel for each video session. This scheme is very expensive and non-scalable. To preserve server bandwidth, many users can share a channel using multicast. Two types of multicast have been considered. In a non-periodic multicast setting, users...
Show moreVideo-On-Demand (VOD) has appeared as an important technology for many multimedia applications such as news on demand, digital libraries, home entertainment, and distance learning. In its simplest form, delivery of a video stream requires a dedicated channel for each video session. This scheme is very expensive and non-scalable. To preserve server bandwidth, many users can share a channel using multicast. Two types of multicast have been considered. In a non-periodic multicast setting, users make video requests to the server; and it serves them according to some scheduling policy. In a periodic broadcast environment, the server does not wait for service requests. It broadcasts a video cyclically, e.g., a new stream of the same video is started every t seconds. Although, this type of approach does not guarantee true VOD, the worst service latency experienced by any client is less than t seconds. A distinct advantage of this approach is that it can serve a very large community of users using minimal server bandwidth. In VOD System it is desirable to provide the user with the video-cassette-recorder-like (VCR) capabilities such as fast-forwarding a video or jumping to a specific frame. This issue in the broadcast framework is addressed, where each video and its interactive version are broadcast repeatedly on the network. Existing techniques rely on data prefetching as the mechanism to provide this functionality. This approach provides limited usability since the prefetching rate cannot keep up with typical fast-forward speeds. In the same environment, end users might have access to different bandwidth capabilities at different times. Current periodic broadcast schemes, do not take advantage of high-bandwidth capabilities, nor do they adapt to the low-bandwidth limitation of the receivers. A heterogeneous technique is presented that can adapt to a range of receiving bandwidth capability. Given a server bandwidth and a range of different client bandwidths, users employing the proposed technique will choose either to use their full reception bandwidth capability and therefore accessing the video at a very short time, or using part or enough reception bandwidth at the expense of a longer access latency.
Show less - Date Issued
- 2004
- Identifier
- CFE0000085, ucf:46129
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0000085
- Title
- CAMERA SYSTEM SUPPORT FOR HIGHWAY TRANSPORTATION USING MOBILE DEVICES.
- Creator
-
Minh, Le, Hua, Kien A., University of Central Florida
- Abstract / Description
-
With the very fast growing technology in wireless, advancement in hardware and the dramatically falling cost of mobile computing devices such as PDA, handheld device, People nowadays can have a personal device that fits in their hand but has computing power as a desktop did few years ago. The same device now is able to communicate over a wireless network and view office document at the same time. The combination of size, power and flexibility makes the personal devices increasingly appear in...
Show moreWith the very fast growing technology in wireless, advancement in hardware and the dramatically falling cost of mobile computing devices such as PDA, handheld device, People nowadays can have a personal device that fits in their hand but has computing power as a desktop did few years ago. The same device now is able to communicate over a wireless network and view office document at the same time. The combination of size, power and flexibility makes the personal devices increasingly appear in many aspects of life.In this proposal, we focus on a simple yet useful application of mobile devices and wireless capabilities. The application can help commuters in traffic system to find an optimal route based on video camera surveillance information. This surveillance information is made available to the user through his/her handheld devices. As an example, suppose we have installed several cameras along the expressway. If commuters can access to these cameras, they can observe the situation currently happening along the way, and decide which path would be the most effective to avoid the traffic congestion. This application will eventually improve the effectiveness of current traffic system since it will help to reduce traffic congestions.
Show less - Date Issued
- 2004
- Identifier
- CFE0000094, ucf:46090
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0000094
- Title
- AUTOMATIC ANNOTATION OF DATABASE IMAGES FOR QUERY-BY-CONCEPT.
- Creator
-
Hiransakolwong, Nualsawat, Hua, kien A., University of Central Florida
- Abstract / Description
-
As digital images become ubiquitous in many applications, the need for efficient and effective retrieval techniques is more demanding than ever. Query by Example (QBE) and Query by Concept (QBC) are among the most popular query models. The former model accepts example images as queries and searches for similar ones based on low-level features such as colors and textures. The latter model allows queries to be expressed in the form of high-level semantics or concept words, such as "boat" or ...
Show moreAs digital images become ubiquitous in many applications, the need for efficient and effective retrieval techniques is more demanding than ever. Query by Example (QBE) and Query by Concept (QBC) are among the most popular query models. The former model accepts example images as queries and searches for similar ones based on low-level features such as colors and textures. The latter model allows queries to be expressed in the form of high-level semantics or concept words, such as "boat" or "car," and finds images that match the specified concepts. Recent research has focused on the connections between these two models and attempts to close the semantic-gap between them. This research involves finding the best method that maps a set of low-level features into high-level concepts. Automatic annotation techniques are investigated in this dissertation to facilitate QBC. In this approach, sets of training images are used to discover the relationship between low-level features and predetermined high-level concepts. The best mapping with respect to the training sets is proposed and used to analyze images, annotating them with the matched concept words. One principal difference between QBE and QBC is that, while similarity matching in QBE must be done at the query time, QBC performs concept exploration off-line. This difference allows QBC techniques to shift the time-consuming task of determining similarity away from the query time, thus facilitating the additional processing time required for increasingly accurate matching. Consequently, QBC's primary design objective is to achieve accurate annotation within a reasonable processing time. This objective is the guiding principle in the design of the following proposed methods which facilitate image annotation: 1.A novel dynamic similarity function. This technique allows users to query with multiple examples: relevant, irrelevant or neutral. It uses the range distance in each group to automatically determine weights in the distance function. Among the advantages of this technique are higher precision and recall rates with fast matching time. 2.Object recognition based on skeletal graphs. The topologies of objects' skeletal graphs are captured and compared at the node level. Such graph representation allows preservation of the skeletal graph's coherence without sacrificing the flexibility of matching similar portions of graphs across different levels. The technique is robust to translation, scaling, and rotation invariants at object level. This technique achieves high precision and recall rates with reasonable matching time and storage space. 3.ASIA (Automatic Sampling-based Image Annotation) is a technique based on a new sampling-based matching framework allowing users to identify their area of interest. ASIA eliminates noise, or irrelevant areas of the image. ASIA is robust to translation, scaling, and rotation invariants at the object level. This technique also achieves high precision and recall rates. While the above techniques may not be the fastest when contrasted with some other recent QBE techniques, they very effectively perform image annotation. The results of applying these processes are accurately annotated database images to which QBC may then be applied. The results of extensive experiments are presented to substantiate the performance advantages of the proposed techniques and allow them to be compared with other recent high-performance techniques. Additionally, a discussion on merging the proposed techniques into a highly effective annotation system is also detailed.
Show less - Date Issued
- 2004
- Identifier
- CFE0000262, ucf:46239
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0000262
- Title
- LEARNING TECHNIQUES FOR INFORMATION RETRIEVAL AND MINING IN HIGH-DIMENSIONAL DATABASES.
- Creator
-
Cheng, Hao, Hua, Kien A., University of Central Florida
- Abstract / Description
-
The main focus of my research is to design effective learning techniques for information retrieval and mining in high-dimensional databases. There are two main aspects in the retrieval and mining research: accuracy and efficiency. The accuracy problem is how to return results which can better match the ground truth, and the efficiency problem is how to evaluate users' requests and execute learning algorithms as fast as possible. However, these problems are non-trivial because of the...
Show moreThe main focus of my research is to design effective learning techniques for information retrieval and mining in high-dimensional databases. There are two main aspects in the retrieval and mining research: accuracy and efficiency. The accuracy problem is how to return results which can better match the ground truth, and the efficiency problem is how to evaluate users' requests and execute learning algorithms as fast as possible. However, these problems are non-trivial because of the complexity of the high-level semantic concepts, the heterogeneous natures of the feature space, the high dimensionality of data representations and the size of the databases. My dissertation is dedicated to addressing these issues. Specifically, my work has five main contributions as follows. The first contribution is a novel manifold learning algorithm, Local and Global Structures Preserving Projection (LGSPP), which defines salient low-dimensional representations for the high-dimensional data. A small number of projection directions are sought in order to properly preserve the local and global structures for the original data. Specifically, two groups of points are extracted for each individual point in the dataset: the first group contains the nearest neighbors of the point, and the other set are a few sampled points far away from the point. These two point sets respectively characterize the local and global structures with regard to the data point. The objective of the embedding is to minimize the distances of the points in each local neighborhood and also to disperse the points far away from their respective remote points in the original space. In this way, the relationships between the data in the original space are well preserved with little distortions. The second contribution is a new constrained clustering algorithm. Conventionally, clustering is an unsupervised learning problem, which systematically partitions a dataset into a small set of clusters such that data in each cluster appear similar to each other compared with those in other clusters. In the proposal, the partial human knowledge is exploited to find better clustering results. Two kinds of constraints are integrated into the clustering algorithm. One is the must-link constraint, indicating that the involved two points belong to the same cluster. On the other hand, the cannot-link constraint denotes that two points are not within the same cluster. Given the input constraints, data points are arranged into small groups and a graph is constructed to preserve the semantic relations between these groups. The assignment procedure makes a best effort to assign each group to a feasible cluster without violating the constraints. The theoretical analysis reveals that the probability of data points being assigned to the true clusters is much higher by the new proposal, compared to conventional methods. In general, the new scheme can produce clusters which can better match the ground truth and respect the semantic relations between points inferred from the constraints. The third contribution is a unified framework for partition-based dimension reduction techniques, which allows efficient similarity retrieval in the high-dimensional data space. Recent similarity search techniques, such as Piecewise Aggregate Approximation (PAA), Segmented Means (SMEAN) and Mean-Standard deviation (MS), prove to be very effective in reducing data dimensionality by partitioning dimensions into subsets and extracting aggregate values from each dimension subset. These partition-based techniques have many advantages including very efficient multi-phased pruning while being simple to implement. They, however, are not adaptive to different characteristics of data in diverse applications. In this study, a unified framework for these partition-based techniques is proposed and the issue of dimension partitions is examined in this framework. An investigation of the relationships of query selectivity and the dimension partition schemes discovers indicators which can predict the performance of a partitioning setting. Accordingly, a greedy algorithm is designed to effectively determine a good partitioning of data dimensions so that the performance of the reduction technique is robust with regard to different datasets. The fourth contribution is an effective similarity search technique in the database of point sets. In the conventional model, an object corresponds to a single vector. In the proposed study, an object is represented by a set of points. In general, this new representation can be used in many real-world applications and carries much more local information, but the retrieval and learning problems become very challenging. The Hausdorff distance is the common distance function to measure the similarity between two point sets, however, this metric is sensitive to outliers in the data. To address this issue, a novel similarity function is defined to better capture the proximity of two objects, in which a one-to-one mapping is established between vectors of the two objects. The optimal mapping minimizes the sum of distances between each paired points. The overall distance of the optimal matching is robust and has high retrieval accuracy. The computation of the new distance function is formulated into the classical assignment problem. The lower-bounding techniques and early-stop mechanism are also proposed to significantly accelerate the expensive similarity search process. The classification problem over the point-set data is called Multiple Instance Learning (MIL) in the machine learning community in which a vector is an instance and an object is a bag of instances. The fifth contribution is to convert the MIL problem into a standard supervised learning in the conventional vector space. Specially, feature vectors of bags are grouped into clusters. Each object is then denoted as a bag of cluster labels, and common patterns of each category are discovered, each of which is further reconstructed into a bag of features. Accordingly, a bag is effectively mapped into a feature space defined by the distances from this bag to all the derived patterns. The standard supervised learning algorithms can be applied to classify objects into pre-defined categories. The results demonstrate that the proposal has better classification accuracy compared to other state-of-the-art techniques. In the future, I will continue to explore my research in large-scale data analysis algorithms, applications and system developments. Especially, I am interested in applications to analyze the massive volume of online data.
Show less - Date Issued
- 2009
- Identifier
- CFE0002882, ucf:48022
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFE0002882
- Title
- Efficient techniques for management and delivery of video data.
- Creator
-
Oh, Junghwan, Hua, Kien A., Engineering and Computer Science
- Abstract / Description
-
University of Central Florida College of Engineering Thesis; The rapid advances in electronic imaging, storage, data compression telecommunications, and networking technology have resulted in a vast creation and use of digital videos in many important applications such as digital libraries, distance learning, public information systems, electronic commerce, movie on demand, etc. This brings about the need for management as well as delivery of video data. Organizing and managing video data,...
Show moreUniversity of Central Florida College of Engineering Thesis; The rapid advances in electronic imaging, storage, data compression telecommunications, and networking technology have resulted in a vast creation and use of digital videos in many important applications such as digital libraries, distance learning, public information systems, electronic commerce, movie on demand, etc. This brings about the need for management as well as delivery of video data. Organizing and managing video data, however, is much more complex than managing conventional text data due to their semantically rich and unstructured contents. Also, the enormous size of video files requires high communication bandwidth for data delivery. In this dissertation, I present the following techniques for video data management and delivery. Decomposing video into meaningful pieces (i.e., shots) is a very fundamental step to handling the complicated contents of video data. Content-based video parsing techniques are presented and analyzed. In order to reduce the computation cost substantially, a non-sequential approach to shot boundary detection is investigated. Efficient browsing and indexing of video data are essential for video data management. Non-linear browsing and cost-effective indexing schemes for video data based on their contents are described and evaluated. In order to satisfy various user requests, delivering long videos through the limited capacity of bandwidth is challenging work. To reduce the demand on this bandwidth, a hybrid of two effective approaches, periodic broadcast and scheduled multicast, is discussed and simulated. The current techniques related to the above works are discussed thoroughly to explain their advantages and disadvantages, and to make the new improved schemes. The substantial amount of experiments and simulations as well as the concepts are provided to compare the introduced techniques with the other existing ones. The results indicate that they outperform recent techniques by a significant margin. I conclude the dissertation with a discussing of future research directions.
Show less - Date Issued
- 2000
- Identifier
- CFR0001719, ucf:52918
- Format
- Document (PDF)
- PURL
- http://purl.flvc.org/ucf/fd/CFR0001719