National Repository of Grey Literature 33 records found  beginprevious31 - 33  jump to record: Search took 0.01 seconds. 
Image Database Query by Example
Dobrotka, Matúš ; Hradiš, Michal (referee) ; Veľas, Martin (advisor)
This thesis deals with content-based image retrieval. The objective of the thesis is to develop an application, which will compare different approaches of image retrieval. First basic approach consists of keypoints detection, local features extraction and creating a visual vocabulary by clustering algorithm - k-means. Using this visual vocabulary is computed histogram of occurrence count of visual words - Bag of Words (BoW), which globally represents an image. After applying an appropriate metrics, it follows finding similar images. Second approach uses deep convolutional neural networks (DCNN) to extract feature vectors. These vectors are used to create a visual vocabulary, which is used to calculate BoW. Next procedure is then similar to the first approach. Third approach uses extracted vectors from DCNN as BoW vectors. It is followed by applying an appropriate metrics and finding similar images. The conclusion describes mentioned approaches, experiments and the final evaluation.
Automatic Content-Based Image Categorization
Němec, Ladislav ; Španěl, Michal (referee) ; Veľas, Martin (advisor)
This thesis deals with automatic content-based image classification. The main goal of this work is implementation of application which is able to perform this task automatically. The solution consists of variable system using local image features extraction and visual vocabulary built by k-means method. Bag Of Words representation is used as a global feature describing each image. Support Vector Machines - the final component of this system - perform the classification based on this representation. In the last chapter, the results of this experimental system are presented.
Klasifikace entit pomocí Wikipedie a WordNetu
Kliegr, Tomáš ; Rauch, Jan (advisor) ; Berka, Petr (referee) ; Smrž, Pavel (referee) ; Žabokrtský, Zdeněk (referee)
This dissertation addresses the problem of classification of entities in text represented by noun phrases. The goal of this thesis is to develop a method for automated classification of entities appearing in datasets consisting of short textual fragments. The emphasis is on unsupervised and semi-supervised methods that will allow for fine-grained character of the assigned classes and require no labeled instances for training. The set of target classes is either user-defined or determined automatically. Our initial attempt to address the entity classification problem is called Semantic Concept Mapping (SCM) algorithm. SCM maps the noun phrases representing the entities as well as the target classes to WordNet. Graph-based WordNet similarity measures are used to assign the closest class to the noun phrase. If a noun phrase does not match any WordNet concept, a Targeted Hypernym Discovery (THD) algorithm is executed. The THD algorithm extracts a hypernym from a Wikipedia article defining the noun phrase using lexico-syntactic patterns. This hypernym is then used to map the noun phrase to a WordNet synset, but it can also be perceived as the classification result by itself, resulting in an unsupervised classification system. SCM and THD algorithms were designed for English. While adaptation of these algorithms for other languages is conceivable, we decided to develop the Bag of Articles (BOA) algorithm, which is language agnostic as it is based on the statistical Rocchio classifier. Since this algorithm utilizes Wikipedia as a source of data for classification, it does not require any labeled training instances. WordNet is used in a novel way to compute term weights. It is also used as a positive term list and for lemmatization. A disambiguation algorithm utilizing global context is also proposed. We consider the BOA algorithm to be the main contribution of this dissertation. Experimental evaluation of the proposed algorithms is performed on the WordSim353 dataset, which is used for evaluation in the Word Similarity Computation (WSC) task, and on the Czech Traveler dataset, the latter being specifically designed for the purpose of our research. BOA performance on WordSim353 achieves Spearman correlation of 0.72 with human judgment, which is close to the 0.75 correlation for the ESA algorithm, to the author's knowledge the best performing algorithm for this gold-standard dataset, which does not require training data. The advantage of BOA over ESA is that it has smaller requirements on preprocessing of the Wikipedia data. While SCM underperforms on the WordSim353 dataset, it overtakes BOA on the Czech Traveler dataset, which was designed specifically for our entity classification problem. This discrepancy requires further investigation. In a standalone evaluation of THD on Czech Traveler dataset the algorithm returned a correct hypernym for 62% of entities.

National Repository of Grey Literature : 33 records found   beginprevious31 - 33  jump to record:
Interested in being notified about new results for this query?
Subscribe to the RSS feed.