Matches in SemOpenAlex for { <https://semopenalex.org/work/W3000805596> ?p ?o ?g. }
Showing items 1 to 67 of
67
with 100 items per page.
- W3000805596 abstract "Machine Learning (ML) has enjoyed huge successes in recent years and an ever- growing number of real-world applications rely on it. However, designing promising algorithms for a specific problem still requires huge human effort. Automated Machine Learning (AutoML) aims at taking the human out of the loop and develop machines that generate / recommend good algorithms for a given ML tasks. AutoML is usually treated as an algorithm / hyper-parameter selection problems, existing approaches include Bayesian optimization, evolutionary algorithms as well as reinforcement learning. Among them, auto-sklearn which incorporates meta-learning techniques in their search initialization, ranks consistently well in AutoML challenges. This observation oriented my research to the Meta-Learning domain. This direction led me to develop a novel framework based on Markov Decision Processes (MDP) and reinforcement learning (RL).After a general introduction (Chapter 1), my thesis work starts with an in-depth analysis of the results of the AutoML challenge (Chapter 2). This analysis oriented my work towards meta-learning, leading me first to propose a formulation of AutoML as a recommendation problem, and ultimately to formulate a novel conceptualisation of the problem as a MDP (Chapter 3). In the MDP setting, the problem is brought back to filling up, as quickly and efficiently as possible, a meta-learning matrix S, in which lines correspond to ML tasks and columns to ML algorithms. A matrix element S(i, j) is the performance of algorithm j applied to task i. Searching efficiently for the best values in S allows us to identify quickly algorithms best suited to given tasks. In Chapter 4 the classical hyper-parameter optimization framework (HyperOpt) is first reviewed. In Chapter 5 a first meta-learning approach is introduced along the lines of our paper ActivMetaL that combines active learning and collaborative filtering techniques to predict the missing values in S. Our latest research applies RL to the MDP problem we defined to learn an efficient policy to explore S. We call this approach REVEAL and propose an analogy with a series of toy games to help visualize agents’ strategies to reveal information progressively, e.g. masked areas of images to be classified, or ship positions in a battleship game. This line of research is developed in Chapter 6. The main results of my PhD project are: 1) HP / model selection: I have explored the Freeze-Thaw method and optimized the algorithm to enter the first AutoML challenge, achieving 3rd place in the final round (Chapter 3). 2) ActivMetaL: I have designed a new algorithm for active meta-learning (ActivMetaL) and compared it with other baseline methods on real-world and artificial data. This study demonstrated that ActiveMetaL is generally able to discover the best algorithm faster than baseline methods. 3) REVEAL: I developed a new conceptualization of meta-learning as a Markov Decision Process and put it into the more general framework of REVEAL games. With a master student intern, I developed agents that learns (with reinforcement learning) to predict the next best algorithm to be tried. To develop this agent, we used surrogate toy tasks of REVEAL games. We then applied our methods to AutoML problems. The work presented in my thesis is empirical in nature. Several real world meta-datasets were used in this research. Artificial and semi-artificial meta-datasets are also used in my work. The results indicate that RL is a viable approach to this problem, although much work remains to be done to optimize algorithms to make them scale to larger meta-learning problems." @default.
- W3000805596 created "2020-01-30" @default.
- W3000805596 creator A5051369578 @default.
- W3000805596 date "2019-12-19" @default.
- W3000805596 modified "2023-09-26" @default.
- W3000805596 title "Meta-Learning as a Markov Decision Process" @default.
- W3000805596 hasPublicationYear "2019" @default.
- W3000805596 type Work @default.
- W3000805596 sameAs 3000805596 @default.
- W3000805596 citedByCount "0" @default.
- W3000805596 crossrefType "dissertation" @default.
- W3000805596 hasAuthorship W3000805596A5051369578 @default.
- W3000805596 hasConcept C105795698 @default.
- W3000805596 hasConcept C106189395 @default.
- W3000805596 hasConcept C114466953 @default.
- W3000805596 hasConcept C119857082 @default.
- W3000805596 hasConcept C127413603 @default.
- W3000805596 hasConcept C154945302 @default.
- W3000805596 hasConcept C159886148 @default.
- W3000805596 hasConcept C199360897 @default.
- W3000805596 hasConcept C201995342 @default.
- W3000805596 hasConcept C2780451532 @default.
- W3000805596 hasConcept C2781002164 @default.
- W3000805596 hasConcept C33923547 @default.
- W3000805596 hasConcept C41008148 @default.
- W3000805596 hasConcept C97541855 @default.
- W3000805596 hasConceptScore W3000805596C105795698 @default.
- W3000805596 hasConceptScore W3000805596C106189395 @default.
- W3000805596 hasConceptScore W3000805596C114466953 @default.
- W3000805596 hasConceptScore W3000805596C119857082 @default.
- W3000805596 hasConceptScore W3000805596C127413603 @default.
- W3000805596 hasConceptScore W3000805596C154945302 @default.
- W3000805596 hasConceptScore W3000805596C159886148 @default.
- W3000805596 hasConceptScore W3000805596C199360897 @default.
- W3000805596 hasConceptScore W3000805596C201995342 @default.
- W3000805596 hasConceptScore W3000805596C2780451532 @default.
- W3000805596 hasConceptScore W3000805596C2781002164 @default.
- W3000805596 hasConceptScore W3000805596C33923547 @default.
- W3000805596 hasConceptScore W3000805596C41008148 @default.
- W3000805596 hasConceptScore W3000805596C97541855 @default.
- W3000805596 hasLocation W30008055961 @default.
- W3000805596 hasOpenAccess W3000805596 @default.
- W3000805596 hasPrimaryLocation W30008055961 @default.
- W3000805596 hasRelatedWork W2726717203 @default.
- W3000805596 hasRelatedWork W2808682055 @default.
- W3000805596 hasRelatedWork W2912722426 @default.
- W3000805596 hasRelatedWork W2923504512 @default.
- W3000805596 hasRelatedWork W2949555518 @default.
- W3000805596 hasRelatedWork W2962928691 @default.
- W3000805596 hasRelatedWork W2969422541 @default.
- W3000805596 hasRelatedWork W2979869797 @default.
- W3000805596 hasRelatedWork W2984492910 @default.
- W3000805596 hasRelatedWork W2995957611 @default.
- W3000805596 hasRelatedWork W3015553638 @default.
- W3000805596 hasRelatedWork W3036413187 @default.
- W3000805596 hasRelatedWork W3041510421 @default.
- W3000805596 hasRelatedWork W3089433505 @default.
- W3000805596 hasRelatedWork W3092688116 @default.
- W3000805596 hasRelatedWork W3109198942 @default.
- W3000805596 hasRelatedWork W3114060346 @default.
- W3000805596 hasRelatedWork W3130794906 @default.
- W3000805596 hasRelatedWork W3159045549 @default.
- W3000805596 hasRelatedWork W3205801758 @default.
- W3000805596 isParatext "false" @default.
- W3000805596 isRetracted "false" @default.
- W3000805596 magId "3000805596" @default.
- W3000805596 workType "dissertation" @default.