Please use this identifier to cite or link to this item: https://doi.org/10.48441/4427.3295.2
DC FieldValueLanguage
dc.contributor.advisorHendrix, Eligius María Theodorus-
dc.contributor.advisorNowak, Ivo-
dc.contributor.authorAziz, Vanya-
dc.date.accessioned2026-04-08T11:44:41Z-
dc.date.available2026-03-12T13:23:29Z-
dc.date.available2026-04-08T11:44:41Z-
dc.date.issued2025-
dc.identifier.urihttps://hdl.handle.net/20.500.12738/19053.2-
dc.description.abstractThe term “Industry 4.0” describes the fourth industrial revolution and is characterized by the integration of digital technology into manufacturing processes. The transformative concepts in Industry 4.0 enable economic production at radically small lot sizes, requiring unprecedented levels of automation and adaptability. These requirements on production facilities necessitate autonomously acting and self-optimizing systems. The field of machine learning offers promising solutions to achieve the objectives of Industry 4.0, particularly by enabling data-driven decision-making and adaptive control mechanisms. This dissertation focuses on Deep Reinforcement Learning (DRL), a neural network-based approach for solving Markov Decision Processes in high-dimensional spaces with unknown transition dynamics. The main contribution of this thesis is the development of a novel state-of-the-art distributional reinforcement learning algorithm within the maximum-entropy Actor-Critic framework. This algorithm, termed ”Cramér-based Soft Distributional Soft Actor-critic” (CDSAC), demonstrates superior performance to other RL algorithms, especially in environments with high-dimensional spaces and complex dynamics. Its performance is shown to be partly rooted in a phenomenon arising in Cramér-metric-based Distributional Reinforcement Learning, referred to as confidence-driven model updates. This mechanism ensures that the value function approximator is updated more conservatively when confidence in its estimates is low. Theoretical justifications for the algorithm are provided, demonstrating its convergence in the policy evaluation setting and, under widely accepted mild assumptions, in the control setting as well. Beyond foundational algorithmic research, this thesis contributes to the practical application of RL in robotics. Given the crucial role of multi-joint robotic systems in modern production technology, a RL meta-algorithm called ”Reinforcement Learning - Inverse Kinematics” (RL-IK) is devised. This approach enhances the applicability of reinforcement learning to robotic control tasks by significantly accelerating convergence to near-optimal policiesen
dc.language.isoenen_US
dc.publisherUMA Editorialen_US
dc.relation.replaceshttp://dx.doi.org/10.48441/4427.3295-
dc.subjectRobótica - Tesis doctoralesen_US
dc.subjectProgramación linealen_US
dc.subjectAprendizaje automático (Inteligencia artificial)en_US
dc.subjectRedes neuronales (Informática)en_US
dc.subjectDistributional Reinforcement Learningen_US
dc.subjectSoft Actor-Criticen_US
dc.subjectRoboticsen_US
dc.subjectLinear Programmingen_US
dc.subjectEnsembleen_US
dc.subject.ddc620: Ingenieurwissenschaftenen_US
dc.titleNovel distributional reinforcement and ensemble learning algorithmsen
dc.title.alternativeNuevos algoritmos de aprendizaje por refuerzo distributivo y aprendizaje ensambladores
dc.typeThesisen_US
dc.identifier.doi10.48441/4427.3295.2-
dcterms.dateAccepted2025-
dc.description.versionAlternativeRevieweden_US
openaire.rightsinfo:eu-repo/semantics/openAccessen_US
thesis.grantor.departmentUniversidad de Málaga. Departamento de Ingeniería mecánica, térmica y de fluidosen_US
thesis.grantor.placeMálagaen_US
thesis.grantor.universityOrInstitutionUniversidad de Málagaen_US
tuhh.identifier.urnurn:nbn:de:gbv:18302-reposit-237178-
tuhh.oai.showtrueen_US
tuhh.publication.instituteUniversidad de Málagaen_US
tuhh.publication.instituteUniversidad de Málaga. Departamento de Ingeniería mecánica, térmica y de fluidosen_US
tuhh.publication.instituteDepartment Maschinenbau und Produktion (ehemalig, aufgelöst 10.2025)en_US
tuhh.publication.instituteFakultät Technik und Informatik (ehemalig, aufgelöst 10.2025)en_US
tuhh.publisher.urlhttps://hdl.handle.net/10630/39287-
tuhh.type.opusDissertation-
tuhh.type.rdmtrue-
dc.rights.cchttps://creativecommons.org/licenses/by-nc-nd/4.0/en_US
dc.type.casraiDissertation-
dc.type.dinidoctoralThesis-
dc.type.driverdoctoralThesis-
dc.type.statusinfo:eu-repo/semantics/updatedVersionen_US
dc.type.thesisdoctoralThesisen_US
dcterms.DCMITypeText-
local.comment.externalAziz, Vanya. (2025). Novel distributional reinforcement and ensemble learning algorithms, I-VII, 1-153. dissertation. UMA Editorial. https://hdl.handle.net/10630/39287en_US
tuhh.apc.statusfalseen_US
item.fulltextWith Fulltext-
item.grantfulltextopen-
item.creatorOrcidAziz, Vanya-
item.creatorGNDAziz, Vanya-
item.languageiso639-1en-
item.openairecristypehttp://purl.org/coar/resource_type/c_46ec-
item.cerifentitytypePublications-
item.advisorGNDHendrix, Eligius María Theodorus-
item.advisorGNDNowak, Ivo-
item.openairetypeThesis-
crisitem.author.deptDepartment Maschinenbau und Produktion (ehemalig, aufgelöst 10.2025)-
crisitem.author.parentorgFakultät Technik und Informatik (ehemalig, aufgelöst 10.2025)-
Appears in Collections:Publications with full text
Files in This Item:
File Description SizeFormat
Thesis_30_03_26.pdf3.75 MBAdobe PDFView/Open
Show simple item record

Google ScholarTM

Check

HAW Katalog

Check

Note about this record

Version History
Version Item Date Summary
2 doi:10.48441/4427.3295.2 2026-04-02 07:58:52.153 updated version (only minor changes): 2026-03-30
1 doi:10.48441/4427.3295 2026-03-12 14:23:29.0 old version: 2025

This item is licensed under a Creative Commons License Creative Commons