Stock volatility forecasting with transformer network

dc.contributor.authorSababipour Asl, Golnaz
dc.contributor.examiningcommitteeRouhani, Sara (Computer Science)
dc.contributor.examiningcommitteeThavaneswaran, Aerambamoorthy (Statistics)
dc.contributor.supervisorThulasiram, Ruppa
dc.date.accessioned2024-02-26T20:15:56Z
dc.date.available2024-02-26T20:15:56Z
dc.date.issued2023-12-20
dc.date.submitted2024-02-25T17:29:52Zen_US
dc.degree.disciplineComputer Scienceen_US
dc.degree.levelMaster of Science (M.Sc.)
dc.description.abstractFinancial world faces many uncertainties due to decisions by governments, business entities, technology trends, natural calamities etc. Stocks and stock market form one of the major activities in financial world. Volatility is one of the main measures of uncertainty in financial stock markets. Hence, forecasting stock volatility is a critical component in many financial problems such as financial risk management, optimizing portfolios etc. In addition to traditional statistical techniques, there have been few artificial intelligence (AI) and machine learning (ML) techniques used in the literature for this volatility forecasting problem. Transformer Network (TN) architecture is one of newest ML techniques. For this thesis work, we utilized TN architecture with multi-head attention (MHA) mechanism for stock volatility forecasting. To enhance the performance of the TN, we incorporated different variations of the feed forward layer. The performance of three distinct TN models was evaluated by implementing three different deep learning (DL) layers (CNN, LSTM, and a hybrid layer CNN LSTM) in the encoder block of TN as the feed forward layer. The results clearly demonstrate that the TN model with the hybrid layer (CNN-LSTM) outperformed the other models, including a recently proposed data-driven approach. Furthermore, we assessed the performance of another latest model that is on built on TN known as Informer model on a minute-scale Bitcoin dataset across various forecast lengths. Our findings underscore the advantages of the Informer model, specifically its ProbSparse attention mechanism and distilling operation, which substantially enhances its efficiency in handling long sequence time-series forecasting (LSTF) tasks.
dc.description.noteMay 2024
dc.identifier.urihttp://hdl.handle.net/1993/38029
dc.language.isoeng
dc.rightsopen accessen_US
dc.subjectForecasting
dc.subjectVolatility
dc.subjectDeep Learning
dc.subjectTransformer Network
dc.titleStock volatility forecasting with transformer network
dc.typemaster thesisen_US
local.subject.manitobano
oaire.awardURIhttps://umanitoba.ca/faculties/graduate_studies/media/GETS_Program_Guidelines_-_2020.pdf
project.funder.nameGRADUATE ENHANCEMENT OF TRI-AGENCY STIPENDS (GETS)
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
GOLNAZ_SABABIPOUR_ASL_Thesis_document.pdf
Size:
1.74 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
770 B
Format:
Item-specific license agreed to upon submission
Description: