Narrow your search

Library

KU Leuven (3)

Odisee (3)

Thomas More Kempen (3)

Thomas More Mechelen (3)

VIVES (3)

ULB (2)

ULiège (2)

AP (1)

KDG (1)

UCLL (1)

More...

Resource type

book (3)

digital (1)


Language

English (4)


Year
From To Submit

2024 (2)

2018 (2)

Listing 1 - 4 of 4
Sort by

Book
New Perspectives and Applications of Modern Control Theory : In Honor of Alexander S. Poznyak
Authors: ---
ISBN: 3319624644 3319624636 Year: 2018 Publisher: Cham : Springer International Publishing : Imprint: Springer,

Loading...
Export citation

Choose an application

Bookmark

Abstract

This edited monograph contains research contributions on a wide range of topics such as stochastic control systems, adaptive control, sliding mode control and parameter identification methods. The book also covers applications of robust and adaptice control to chemical and biotechnological systems. This collection of papers commemorates the 70th birthday of Dr. Alexander S. Poznyak. .


Digital
New Perspectives and Applications of Modern Control Theory : In Honor of Alexander S. Poznyak
Authors: ---
ISBN: 9783319624648 Year: 2018 Publisher: Cham Springer International Publishing

Loading...
Export citation

Choose an application

Bookmark

Abstract

This edited monograph contains research contributions on a wide range of topics such as stochastic control systems, adaptive control, sliding mode control and parameter identification methods. The book also covers applications of robust and adaptice control to chemical and biotechnological systems. This collection of papers commemorates the 70th birthday of Dr. Alexander S. Poznyak. .


Book
Optimization and Games for Controllable Markov Chains : Numerical Methods with Application to Finance and Engineering
Authors: ---
ISBN: 3031435753 Year: 2024 Publisher: Cham : Springer Nature Switzerland : Imprint: Springer,

Loading...
Export citation

Choose an application

Bookmark

Abstract

This book considers a class of ergodic finite controllable Markov's chains. The main idea behind the method, described in this book, is to develop the original discrete optimization problems (or game models) in the space of randomized formulations, where the variables stand in for the distributions (mixed strategies or preferences) of the original discrete (pure) strategies in the use. The following suppositions are made: a finite state space, a limited action space, continuity of the probabilities and rewards associated with the actions, and a necessity for accessibility. These hypotheses lead to the existence of an optimal policy. The best course of action is always stationary. It is either simple (i.e., nonrandomized stationary) or composed of two nonrandomized policies, which is equivalent to randomly selecting one of two simple policies throughout each epoch by tossing a biased coin. As a bonus, the optimization procedure just has to repeatedly solve the time-average dynamic programming equation, making it theoretically feasible to choose the optimum course of action under the global restriction. In the ergodic cases the state distributions, generated by the corresponding transition equations, exponentially quickly converge to their stationary (final) values. This makes it possible to employ all widely used optimization methods (such as Gradient-like procedures, Extra-proximal method, Lagrange's multipliers, Tikhonov's regularization), including the related numerical techniques. In the book we tackle different problems and theoretical Markov models like controllable and ergodic Markov chains, multi-objective Pareto front solutions, partially observable Markov chains, continuous-time Markov chains, Nash equilibrium and Stackelberg equilibrium, Lyapunov-like function in Markov chains, Best-reply strategy, Bayesian incentive-compatible mechanisms, Bayesian Partially Observable Markov Games, bargaining solutions for Nash and Kalai-Smorodinsky formulations, multi-traffic signal-control synchronization problem, Rubinstein's non-cooperative bargaining solutions, the transfer pricing problem as bargaining.


Book
Optimization and Games for Controllable Markov Chains
Authors: --- ---
ISBN: 9783031435751 Year: 2024 Publisher: Cham Springer Nature Switzerland :Imprint: Springer

Loading...
Export citation

Choose an application

Bookmark

Abstract

This book considers a class of ergodic finite controllable Markov's chains. The main idea behind the method, described in this book, is to develop the original discrete optimization problems (or game models) in the space of randomized formulations, where the variables stand in for the distributions (mixed strategies or preferences) of the original discrete (pure) strategies in the use. The following suppositions are made: a finite state space, a limited action space, continuity of the probabilities and rewards associated with the actions, and a necessity for accessibility. These hypotheses lead to the existence of an optimal policy. The best course of action is always stationary. It is either simple (i.e., nonrandomized stationary) or composed of two nonrandomized policies, which is equivalent to randomly selecting one of two simple policies throughout each epoch by tossing a biased coin. As a bonus, the optimization procedure just has to repeatedly solve the time-average dynamic programming equation, making it theoretically feasible to choose the optimum course of action under the global restriction. In the ergodic cases the state distributions, generated by the corresponding transition equations, exponentially quickly converge to their stationary (final) values. This makes it possible to employ all widely used optimization methods (such as Gradient-like procedures, Extra-proximal method, Lagrange's multipliers, Tikhonov's regularization), including the related numerical techniques. In the book we tackle different problems and theoretical Markov models like controllable and ergodic Markov chains, multi-objective Pareto front solutions, partially observable Markov chains, continuous-time Markov chains, Nash equilibrium and Stackelberg equilibrium, Lyapunov-like function in Markov chains, Best-reply strategy, Bayesian incentive-compatible mechanisms, Bayesian Partially Observable Markov Games, bargaining solutions for Nash and Kalai-Smorodinsky formulations, multi-traffic signal-control synchronization problem, Rubinstein's non-cooperative bargaining solutions, the transfer pricing problem as bargaining.

Listing 1 - 4 of 4
Sort by