Abstract
The researchers explored building a generative model with the combination of Variational Autoencoders (VAEs) and Generative Adversarial Network (GAN) to offer better results when the agent interacts with the environment. The premise is that the agent model can train the unsupervised environment and increase the imaging quality. An experiment was performed on car racing based on the designed agent model with features that could effectively be extracted to help produce reliable information. The combination of VAE and GAN provides better feature extraction to gather relevant data and solve the complexity. Another novel approach was performed in which the world model has a modified Recurrent Neural Network (RNN) model and was carried out by a bi-directional gated recurrent unit (BGRU) as opposed to a traditional long short-term memory (LSTM) model. BGRU tends to use less memory while executing and training faster than an LSTM, as it uses fewer training parameters. However, the LSTM model provided greater accuracy with datasets using longer sequences. Based on practical implementation, the BGRU model produced better performance results.
Notes
If this is your thesis or dissertation, and want to learn how to access it or for more information about readership statistics, contact us at STARS@ucf.edu
Graduation Date
2021
Semester
Summer
Advisor
Wocjan, Pawel
Degree
Doctor of Philosophy (Ph.D.)
College
College of Engineering and Computer Science
Department
Computer Science
Degree Program
Computer Science
Format
application/pdf
Identifier
CFE0009096; DP0026429
URL
https://purls.library.ucf.edu/go/DP0026429
Language
English
Release Date
February 2027
Length of Campus-only Access
5 years
Access Status
Doctoral Dissertation (Campus-only Access)
STARS Citation
Almalki, Ali, "Innovative Modification in Reinforcement Learning Models Using Recurrent Neural Network and Autoencoder" (2021). Electronic Theses and Dissertations, 2020-2023. 1125.
https://stars.library.ucf.edu/etd2020/1125
Restricted to the UCF community until February 2027; it will then be open access.