UNIVERSITY of CALIFORNIA, SAN DIEGO Bag-Of-Concepts As a Movie

Total Page:16

File Type:pdf, Size:1020Kb

UNIVERSITY of CALIFORNIA, SAN DIEGO Bag-Of-Concepts As a Movie UNIVERSITY OF CALIFORNIA, SAN DIEGO Bag-of-Concepts as a Movie Genome and Representation A Thesis submitted in partial satisfaction of the requirements for the degree Master of Science in Computer Science by Colin Zhou Committee in charge: Professor Shlomo Dubnov, Chair Professor Sanjoy Dasgupta Professor Lawrence Saul 2016 The Thesis of Colin Zhou is approved, and it is acceptable in quality and form for publication on microfilm and electronically: Chair University of California, San Diego 2016 iii DEDICATION For Mom and Dad iv TABLE OF CONTENTS Signature Page........................................................................................................... iii Dedication.................................................................................................................. iv Table of Contents....................................................................................................... v List of Figures and Tables.......................................................................................... vi Abstract of the Thesis................................................................................................ vii Introduction................................................................................................................ 1 Chapter 1 Background............................................................................................... 3 1.1 Vector Space Models............................................................................... 3 1.1.1. Bag-of-Words......................................................................... 4 1.1.2. TF-IDF................................................................................... 4 1.1.3. Latent Semantic Indexing...................................................... 5 1.2 Neural Net Models................................................................................... 5 1.2.1. Neural Net Architectures....................................................... 7 1.3 Bag-of-Concepts Model........................................................................... 9 1.4 Recommender Systems............................................................................ 10 1.4.1. Collaborative Recommenders................................................ 10 1.4.2. Content-based Recommenders............................................... 11 1.4.3. Recommendation Through “Genes”...................................... 11 Chapter 2 Concept Navigation and Tuning............................................................... 14 2.1 Methodology............................................................................................ 14 2.2 Concepts................................................................................................... 15 2.3 Recommendation Tuning......................................................................... 17 2.4 Discussion................................................................................................ 19 Chapter 3 Genre Classification.................................................................................. 22 3.1 Methodology............................................................................................ 22 3.2 Results...................................................................................................... 23 3.3 Discussion................................................................................................ 24 Chapter 4 Future Work.............................................................................................. 26 Appendix.................................................................................................................... 28 References.................................................................................................................. 30 v LIST OF FIGURES AND TABLES Figure 1: Ten BOC Vectors for the Nearest Neighbors of Cold Thoughts (2013)................................................................................................. 16 Table 1: Some words from selected word clusters and a human “best guess” for the cluster topic................................................................. 15 Table 2: Precision, recall, and F-score averages over all genres...................... 23 Table 3: Precision, Recall, and F-1 Scores broken down by genre.................. 28 vi ABSTRACT OF THE THESIS Bag-of-Concepts as a Movie Genome and Representation by Colin Zhou Master of Science in Computer Science University of California, San Diego, 2016 Professor Shlomo Dubnov, Chair As online retailers, media providers, and others have learned, the ability to provide quality recommendations for users is a strong profit multiplier: it drives sales, maintains retention, and provides value for users. Video streaming, retail, and video databases particularly have had extensive effort and research devoted to finding effective ways to make representations. One of these is MovieLens.org, a video recommender website that uses a representation called the tag genome to understand movies and make tuned recommendations. The tag genome representation is built from user input on the website, where users can apply tags to movies and rate them in order to provide the vii information needed to make recommendations. In this work, we draw from research in information retrieval to implement a bag- of-concepts representation for movies and make tuned recommendations in the same manner as MovieLens. Our implementation is fully unsupervised and does not require the user data needed in the implementation of MovieLens while still having similar properties to the tag genome that enable interesting tuned recommendations. viii Introduction Online movie/video information databases and video streaming services often come with a recommendation feature that finds new videos that users may enjoy. These recommendations can keep users engaged and improve their experience with the site or service, which makes these recommendation systems an important feature for designers to focus their efforts on. The Netflix Prize challenge [20] shows the interest in and importance placed on recommender systems by researchers and companies. The competition's goal was to find a recommender system that could beat the accuracy of Netflix's own system by ten percent. Netflix offered a prize of one million dollars and gathered significant media attention, and more than forty thousand teams submitted their own recommender systems. Many recommender systems for movies that are discussed in literature require both detailed information about the films themselves as well as a body of user data and feedback in order to make their recommendations. This work is an attempt at implementing and evaluating a recommender system that has much lower data requirements: we use only text descriptions of each movie's plot and no user data in order to build representations for our movies and make recommendations. The data set used in this paper comes from IMDB.com, accessed through its public FTP interface [3]. Plot descriptions are contributed by its userbase, and contain a large amount of variation. 1 2 Users differ in their writing style, tone, descriptiveness, quality, and more when writing plot summaries, which makes the dataset a challenging one to work with. In Chapter 1 we review relevant literature from the areas of document representation and recommender systems. Models of representation discussed are the vector space model and neural net derived distributed representations. The two approaches of collaborative and content-based recommenders are explained, and relevant previous work is reviewed. In Chapter 2 we describe the methods of manipulating bag- of-concept vectors with meaningful semantics in their composition as an analogue of tag genome navigation pioneered by the GroupLens research lab [27]. Chapter 3 evaluates the performance of bag-of-concepts in comparison to other representation methods in the task of genre classification. The paper concludes with a discussion of possible future work in chapter 4. Chapter 1 Background Document representation has always been a important area of work for fields such as information retrieval and data mining. Once computer systems were put to use in storing and retrieving large bodies of information, researchers began looking for ways to automatically search for documents in a large collection. One influential idea that emerged in this early era, published by H.P. Luhn in 1957 [4], was that the encoding of a document could be made automatic, and that it could capture the “concepts” in a document without needing to understand it in the same way that a human cataloger would. Luhn proposed that the conceptual makeup of documents could be indicated by the words in the document, and that documents could be retrieved based on the similarity of their word distributions. 1.1 Vector Space Models One key development of the 1960's was the SMART information retrieval system, developed by a team led by Gerard Salton[5,6]. The SMART system introduced the vector space model [7, 8] for representing documents, which has become a very widespread model for representing documents and objects in general. In the vector space model, documents are represented as vectors with each dimension corresponding to a 3 4 different term. In most cases a term would be a single word, but in different applications a term can be a phrase or other unit. If a term appears in the document, then its corresponding dimension in the representation has a non-zero value [41].
Recommended publications
  • The Netflix Prize Contest
    1) New Paths to New Machine Learning Science 2) How an Unruly Mob Almost Stole the Grand Prize at the Last Moment Jeff Howbert University of Washington February 4, 2014 Netflix Viewing Recommendations Recommender Systems DOMAIN: some field of activity where users buy, view, consume, or otherwise experience items PROCESS: 1. users provide ratings on items they have experienced 2. Take all < user, item, rating > data and build a predictive model 3. For a user who hasn’t experienced a particular item, use model to predict how well they will like it (i.e. predict rating) Roles of Recommender Systems y Help users deal with paradox of choice y Allow online sites to: y Increase likelihood of sales y Retain customers by providing positive search experience y Considered essential in operation of: y Online retailing, e.g. Amazon, Netflix, etc. y Social networking sites Amazon.com Product Recommendations Social Network Recommendations y Recommendations on essentially every category of interest known to mankind y Friends y Groups y Activities y Media (TV shows, movies, music, books) y News stories y Ad placements y All based on connections in underlying social network graph, and the expressed ‘likes’ and ‘dislikes’ of yourself and your connections Types of Recommender Systems Base predictions on either: y content‐based approach y explicit characteristics of users and items y collaborative filtering approach y implicit characteristics based on similarity of users’ preferences to those of other users The Netflix Prize Contest y GOAL: use training data to build a recommender system, which, when applied to qualifying data, improves error rate by 10% relative to Netflix’s existing system y PRIZE: first team to 10% wins $1,000,000 y Annual Progress Prizes of $50,000 also possible The Netflix Prize Contest y CONDITIONS: y Open to public y Compete as individual or group y Submit predictions no more than once a day y Prize winners must publish results and license code to Netflix (non‐exclusive) y SCHEDULE: y Started Oct.
    [Show full text]
  • Lessons Learned from the Netflix Contest Arthur Dunbar
    Lessons Learned from the Netflix Contest Arthur Dunbar Background ● From Wikipedia: – “The Netflix Prize was an open competition for the best collaborative filtering algorithm to predict user ratings for films, based on previous ratings without any other information about the users or films, i.e. without the users or the films being identified except by numbers assigned for the contest.” ● Two teams dead tied for first place on the test set: – BellKor's Pragmatic Chaos – The Ensemble ● Both are covered here, but technically BellKor won because they submitted their algorithm 20 minutes earlier. This even though The Ensemble did better on the training set (barely). ● Lesson one learned! Be 21 minutes faster! ;-D – Don't feel too bad, that's the reason they won the competition, but the actual test set data was slightly better for them: ● BellKor RMSE: 0.856704 ● The Ensemble RMSE: 0.856714 Background ● Netflix dataset was 100,480,507 date-stamped movie rating performed by anonymous Netflix customers between December 31st 1999 and December 31st 2005. (Netflix has been around since 1997) ● This was 480,189 users and 17,770 movies Background ● There was a hold-out set of 4.2 million ratings for a test set consisting of the last 9 movies rated by each user that had rated at lest 18 movies. – This was split in to 3 sets, Probe, Quiz and Test ● Probe was sent out with the training set and labeled ● Quiz set was used for feedback. The results on this were published along the way and they are what showed upon the leaderboard ● Test is what they actually had to do best on to win the competition.
    [Show full text]
  • Movie Genome: Alleviating New Item Cold Start in Movie Recommendation
    User Modeling and User-Adapted Interaction https://doi.org/10.1007/s11257-019-09221-y Movie genome: alleviating new item cold start in movie recommendation Yashar Deldjoo, et al. [full author details at the end of the article] Received: 23 December 2017 / Accepted in revised form: 19 January 2019 © The Author(s) 2019 Abstract As of today, most movie recommendation services base their recommendations on col- laborative filtering (CF) and/or content-based filtering (CBF) models that use metadata (e.g., genre or cast). In most video-on-demand and streaming services, however, new movies and TV series are continuously added. CF models are unable to make predic- tions in such a scenario, since the newly added videos lack interactions—a problem technically known as new item cold start (CS). Currently, the most common approach to this problem is to switch to a purely CBF method, usually by exploiting textual meta- data. This approach is known to have lower accuracy than CF because it ignores useful collaborative information and relies on human-generated textual metadata, which are expensive to collect and often prone to errors. User-generated content, such as tags, can also be rare or absent in CS situations. In this paper, we introduce a new movie recommender system that addresses the new item problem in the movie domain by (i) integrating state-of-the-art audio and visual descriptors, which can be automati- cally extracted from video content and constitute what we call the movie genome; (ii) exploiting an effective data fusion method named canonical correlation analysis, which was successfully tested in our previous works Deldjoo et al.
    [Show full text]
  • Netflix and the Development of the Internet Television Network
    Syracuse University SURFACE Dissertations - ALL SURFACE May 2016 Netflix and the Development of the Internet Television Network Laura Osur Syracuse University Follow this and additional works at: https://surface.syr.edu/etd Part of the Social and Behavioral Sciences Commons Recommended Citation Osur, Laura, "Netflix and the Development of the Internet Television Network" (2016). Dissertations - ALL. 448. https://surface.syr.edu/etd/448 This Dissertation is brought to you for free and open access by the SURFACE at SURFACE. It has been accepted for inclusion in Dissertations - ALL by an authorized administrator of SURFACE. For more information, please contact [email protected]. Abstract When Netflix launched in April 1998, Internet video was in its infancy. Eighteen years later, Netflix has developed into the first truly global Internet TV network. Many books have been written about the five broadcast networks – NBC, CBS, ABC, Fox, and the CW – and many about the major cable networks – HBO, CNN, MTV, Nickelodeon, just to name a few – and this is the fitting time to undertake a detailed analysis of how Netflix, as the preeminent Internet TV networks, has come to be. This book, then, combines historical, industrial, and textual analysis to investigate, contextualize, and historicize Netflix's development as an Internet TV network. The book is split into four chapters. The first explores the ways in which Netflix's development during its early years a DVD-by-mail company – 1998-2007, a period I am calling "Netflix as Rental Company" – lay the foundations for the company's future iterations and successes. During this period, Netflix adapted DVD distribution to the Internet, revolutionizing the way viewers receive, watch, and choose content, and built a brand reputation on consumer-centric innovation.
    [Show full text]
  • Document and Topic Models: Plsa
    10/4/2018 Document and Topic Models: pLSA and LDA Andrew Levandoski and Jonathan Lobo CS 3750 Advanced Topics in Machine Learning 2 October 2018 Outline • Topic Models • pLSA • LSA • Model • Fitting via EM • pHITS: link analysis • LDA • Dirichlet distribution • Generative process • Model • Geometric Interpretation • Inference 2 1 10/4/2018 Topic Models: Visual Representation Topic proportions and Topics Documents assignments 3 Topic Models: Importance • For a given corpus, we learn two things: 1. Topic: from full vocabulary set, we learn important subsets 2. Topic proportion: we learn what each document is about • This can be viewed as a form of dimensionality reduction • From large vocabulary set, extract basis vectors (topics) • Represent document in topic space (topic proportions) 푁 퐾 • Dimensionality is reduced from 푤푖 ∈ ℤ푉 to 휃 ∈ ℝ • Topic proportion is useful for several applications including document classification, discovery of semantic structures, sentiment analysis, object localization in images, etc. 4 2 10/4/2018 Topic Models: Terminology • Document Model • Word: element in a vocabulary set • Document: collection of words • Corpus: collection of documents • Topic Model • Topic: collection of words (subset of vocabulary) • Document is represented by (latent) mixture of topics • 푝 푤 푑 = 푝 푤 푧 푝(푧|푑) (푧 : topic) • Note: document is a collection of words (not a sequence) • ‘Bag of words’ assumption • In probability, we call this the exchangeability assumption • 푝 푤1, … , 푤푁 = 푝(푤휎 1 , … , 푤휎 푁 ) (휎: permutation) 5 Topic Models: Terminology (cont’d) • Represent each document as a vector space • A word is an item from a vocabulary indexed by {1, … , 푉}. We represent words using unit‐basis vectors.
    [Show full text]
  • Introduction
    Introduction James Bennett Charles Elkan Bing Liu Netflix Department of Computer Science and Department of Computer Science 100 Winchester Circle Engineering University of Illinois at Chicago Los Gatos, CA 95032 University of California, San Diego 851 S. Morgan Street La Jolla, CA 92093-0404 Chicago, IL 60607-7053 [email protected] [email protected] [email protected] Padhraic Smyth Domonkos Tikk Department of Computer Science Department of Telecom. & Media Informatics University of California, Irvine Budapest University of Technology and Economics CA 92697-3425 H-1117 Budapest, Magyar Tudósok krt. 2, Hungary [email protected] [email protected] INTRODUCTION Both tasks employed the Netflix Prize training data set [2], which The KDD Cup is the oldest of the many data mining competitions consists of more than 100 million ratings from over 480 thousand that are now popular [1]. It is an integral part of the annual ACM randomly-chosen, anonymous customers on nearly 18 thousand SIGKDD International Conference on Knowledge Discovery and movie titles. The data were collected between October, 1998 and Data Mining (KDD). In 2007, the traditional KDD Cup December, 2005 and reflect the distribution of all ratings received competition was augmented with a workshop with a focus on the by Netflix during this period. The ratings are on a scale from 1 to concurrently active Netflix Prize competition [2]. The KDD Cup 5 (integral) stars. itself in 2007 consisted of a prediction competition using Netflix Task 1 (Who Rated What in 2006): The task was to predict which movie rating data, with tasks that were different and separate from users rated which movies in 2006.
    [Show full text]
  • Redalyc.Latent Dirichlet Allocation Complement in the Vector Space Model for Multi-Label Text Classification
    International Journal of Combinatorial Optimization Problems and Informatics E-ISSN: 2007-1558 [email protected] International Journal of Combinatorial Optimization Problems and Informatics México Carrera-Trejo, Víctor; Sidorov, Grigori; Miranda-Jiménez, Sabino; Moreno Ibarra, Marco; Cadena Martínez, Rodrigo Latent Dirichlet Allocation complement in the vector space model for Multi-Label Text Classification International Journal of Combinatorial Optimization Problems and Informatics, vol. 6, núm. 1, enero-abril, 2015, pp. 7-19 International Journal of Combinatorial Optimization Problems and Informatics Morelos, México Available in: http://www.redalyc.org/articulo.oa?id=265239212002 How to cite Complete issue Scientific Information System More information about this article Network of Scientific Journals from Latin America, the Caribbean, Spain and Portugal Journal's homepage in redalyc.org Non-profit academic project, developed under the open access initiative © International Journal of Combinatorial Optimization Problems and Informatics, Vol. 6, No. 1, Jan-April 2015, pp. 7-19. ISSN: 2007-1558. Latent Dirichlet Allocation complement in the vector space model for Multi-Label Text Classification Víctor Carrera-Trejo1, Grigori Sidorov1, Sabino Miranda-Jiménez2, Marco Moreno Ibarra1 and Rodrigo Cadena Martínez3 Centro de Investigación en Computación1, Instituto Politécnico Nacional, México DF, México Centro de Investigación e Innovación en Tecnologías de la Información y Comunicación (INFOTEC)2, Ags., México Universidad Tecnológica de México3 – UNITEC MÉXICO [email protected], {sidorov,marcomoreno}@cic.ipn.mx, [email protected] [email protected] Abstract. In text classification task one of the main problems is to choose which features give the best results. Various features can be used like words, n-grams, syntactic n-grams of various types (POS tags, dependency relations, mixed, etc.), or a combinations of these features can be considered.
    [Show full text]
  • Evaluating Vector-Space Models of Word Representation, Or, the Unreasonable Effectiveness of Counting Words Near Other Words
    Evaluating Vector-Space Models of Word Representation, or, The Unreasonable Effectiveness of Counting Words Near Other Words Aida Nematzadeh, Stephan C. Meylan, and Thomas L. Griffiths University of California, Berkeley fnematzadeh, smeylan, tom griffi[email protected] Abstract angle between word vectors (e.g., Mikolov et al., 2013b; Pen- nington et al., 2014). Vector-space models of semantics represent words as continuously-valued vectors and measure similarity based on In this paper, we examine whether these constraints im- the distance or angle between those vectors. Such representa- ply that Word2Vec and GloVe representations suffer from the tions have become increasingly popular due to the recent de- same difficulty as previous vector-space models in capturing velopment of methods that allow them to be efficiently esti- mated from very large amounts of data. However, the idea human similarity judgments. To this end, we evaluate these of relating similarity to distance in a spatial representation representations on a set of tasks adopted from Griffiths et al. has been criticized by cognitive scientists, as human similar- (2007) in which the authors showed that the representations ity judgments have many properties that are inconsistent with the geometric constraints that a distance metric must obey. We learned by another well-known vector-space model, Latent show that two popular vector-space models, Word2Vec and Semantic Analysis (Landauer and Dumais, 1997), were in- GloVe, are unable to capture certain critical aspects of human consistent with patterns of semantic similarity demonstrated word association data as a consequence of these constraints. However, a probabilistic topic model estimated from a rela- in human word association data.
    [Show full text]
  • Gensim Is Robust in Nature and Has Been in Use in Various Systems by Various People As Well As Organisations for Over 4 Years
    Gensim i Gensim About the Tutorial Gensim = “Generate Similar” is a popular open source natural language processing library used for unsupervised topic modeling. It uses top academic models and modern statistical machine learning to perform various complex tasks such as Building document or word vectors, Corpora, performing topic identification, performing document comparison (retrieving semantically similar documents), analysing plain-text documents for semantic structure. Audience This tutorial will be useful for graduates, post-graduates, and research students who either have an interest in Natural Language Processing (NLP), Topic Modeling or have these subjects as a part of their curriculum. The reader can be a beginner or an advanced learner. Prerequisites The reader must have basic knowledge about NLP and should also be aware of Python programming concepts. Copyright & Disclaimer Copyright 2020 by Tutorials Point (I) Pvt. Ltd. All the content and graphics published in this e-book are the property of Tutorials Point (I) Pvt. Ltd. The user of this e-book is prohibited to reuse, retain, copy, distribute or republish any contents or a part of contents of this e-book in any manner without written consent of the publisher. We strive to update the contents of our website and tutorials as timely and as precisely as possible, however, the contents may contain inaccuracies or errors. Tutorials Point (I) Pvt. Ltd. provides no guarantee regarding the accuracy, timeliness or completeness of our website or its contents including this tutorial. If you discover any errors on our website or in this tutorial, please notify us at [email protected] ii Gensim Table of Contents About the Tutorial ..........................................................................................................................................
    [Show full text]
  • Deconstructing Word Embedding Models
    Deconstructing Word Embedding Models Koushik K. Varma Ashoka University, [email protected] Abstract – Analysis of Word Embedding Models through Analysis, one must be certain that all linguistic aspects of a a deconstructive approach reveals their several word are captured within the vector representations because shortcomings and inconsistencies. These include any discrepancies could be further amplified in practical instability of the vector representations, a distorted applications. While these vector representations have analogical reasoning, geometric incompatibility with widespread use in modern natural language processing, it is linguistic features, and the inconsistencies in the corpus unclear as to what degree they accurately encode the data. A new theoretical embedding model, ‘Derridian essence of language in its structural, contextual, cultural and Embedding,’ is proposed in this paper. Contemporary ethical assimilation. Surface evaluations on training-test embedding models are evaluated qualitatively in terms data provide efficiency measures for a specific of how adequate they are in relation to the capabilities of implementation but do not explicitly give insights on the a Derridian Embedding. factors causing the existing efficiency (or the lack there of). We henceforth present a deconstructive review of 1. INTRODUCTION contemporary word embeddings using available literature to clarify certain misconceptions and inconsistencies Saussure, in the Course in General Linguistics, concerning them. argued that there is no substance in language and that all language consists of differences. In language there are only forms, not substances, and by that he meant that all 1.1 DECONSTRUCTIVE APPROACH apparently substantive units of language are generated by other things that lie outside them, but these external Derrida, a post-structualist French philosopher, characteristics are actually internal to their make-up.
    [Show full text]
  • Vector Space Models for Words and Documents Statistical Natural Language Processing ELEC-E5550 Jan 30, 2019 Tiina Lindh-Knuutila D.Sc
    Vector space models for words and documents Statistical Natural Language Processing ELEC-E5550 Jan 30, 2019 Tiina Lindh-Knuutila D.Sc. (Tech) Lingsoft Language Services & Aalto University tiina.lindh-knuutila -at- aalto dot fi Material also by Mari-Sanna Paukkeri (mari-sanna.paukkeri -at- utopiaanalytics dot com) Today’s Agenda • Vector space models • word-document matrices • word vectors • stemming, weighting, dimensionality reduction • similarity measures • Count models vs. predictive models • Word2vec • Information retrieval (Briefly) • Course project details You shall know the word by the company it keeps • Language is symbolic in nature • Surface form is in an arbitrary relation with the meaning of the word • Hat vs. cat • One substitution: Levenshtein distance of 1 • Does not measure the semantic similarity of the words • Distributional semantics • Linguistic items which appear in similar contexts in large samples of language data tend to have similar meanings Firth, John R. 1957. A synopsis of linguistic theory 1930–1955. In Studies in linguistic analysis, 1–32. Oxford: Blackwell. George Miller and Walter Charles. 1991. Contextual correlates of semantic similarity. Language and Cognitive Processes, 6(1):1–28. Vector space models (VSM) • The use of a high-dimensional space of documents (or words) • Closeness in the vector space resembles closeness in the semantics or structure of the documents (depending on the features extracted). • Makes the use of data mining possible • Applications: – Document clustering/classification/… • Finding similar documents • Finding similar words – Word disambiguation – Information retrieval • Term discrimination: ranking keywords in the order of usefulness Vector space models (VSM) • Steps to build a vector space model 1. Preprocessing 2.
    [Show full text]
  • ŠKODA AUTO VYSOKÁ ŠKOLA O.P.S. Valuation of Netflix, Inc. Diploma
    ŠKODA AUTO VYSOKÁ ŠKOLA o.p.s. Course: Economics and Management Field of study/specialisation: Corporate Finance in International Business Valuation of Netflix, Inc. Diploma Thesis Bc. Alena Vershinina Thesis Supervisor: doc. Ing. Tomáš Krabec, Ph.D., MBA 2 I declare that I have prepared this thesis on my own and listed all the sources used in the bibliography. I declare that, while preparing the thesis, I followed the internal regulation of ŠKODA AUTO VYSOKÁ ŠKOLA o.p.s. (hereinafter referred to as ŠAVŠ), directive OS.17.10 Thesis guidelines. I am aware that this thesis is covered by Act No. 121/2000 Coll., the Copyright Act, that it is schoolwork within the meaning of Section 60 and that under Section 35 (3) ŠAVŠ is entitled to use my thesis for educational purposes or internal requirements. I agree with my thesis being published in accordance with Section 47b of Act No. 111/1998 Coll., on Higher Education Institutions. I understand that ŠAVŠ has the right to enter into a licence agreement for this work under standard conditions. If I use this thesis or grant a licence for its use, I agree to inform ŠAVŠ about it. In this case, ŠAVŠ is entitled to demand a contribution to cover the costs incurred in the creation of this work up to their actual amount. Mladá Boleslav, Date ......... 3 I would like to thank doc. Ing. Tomáš Krabec, Ph.D., MBA for his professional supervision of my thesis, advice, and information. I also want to thank the ŠAVS management for their support of the students’ initiative.
    [Show full text]