Το έργο με τίτλο Stochastic optimization on tensor factorization and completion από τον/τους δημιουργό/ούς Siaminou Ioanna διατίθεται με την άδεια Creative Commons Αναφορά Δημιουργού 4.0 Διεθνές
Βιβλιογραφική Αναφορά
Ιωάννα Σιάμινου, "Stochastic optimization on tensor factorization and completion", Μεταπτυχιακή Διατριβή, Σχολή Ηλεκτρολόγων Μηχανικών και Μηχανικών Υπολογιστών, Πολυτεχνείο Κρήτης, Χανιά, Ελλάς, 2021
https://doi.org/10.26233/heallink.tuc.88551
We consider the problem of structured canonical polyadic decomposition (CPD). If the size of the problem is very big, then stochastic optimization approaches are viable alternatives to classical methods, such as Alternating Optimization (AO) and All-At-Once (AAO) optimization. We extend a recent stochastic gradient approach by employing an acceleration step (Nesterov momentum) in each iteration. We compare our approach with state-of-the-art alternatives, using both synthetic and real-world data, and find it to be very competitive. Furthermore, we examine the drawbacks of a parallel implementation of our accelerated stochastic algorithm and describe an alternative method that deals with these limitations. Finally, we propose an accelerated stochastic algorithm for the Nonnegative Tensor Completion problem and its parallel implementation via the shared–memory API OpenMP. Through numerical experiments, we test its efficiency in very large problems.