• No results found

Strongly convergent dynamic programming : some results

N/A
N/A
Protected

Academic year: 2021

Share "Strongly convergent dynamic programming : some results"

Copied!
10
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

Strongly convergent dynamic programming : some results

Citation for published version (APA):

Hee, van, K. M., & Wal, van der, J. (1976). Strongly convergent dynamic programming : some results. (Memorandum COSOR; Vol. 7626). Technische Hogeschool Eindhoven.

Document status and date: Published: 01/01/1976

Document Version:

Publisher’s PDF, also known as Version of Record (includes final page, issue and volume numbers)

Please check the document version of this publication:

• A submitted manuscript is the version of the article upon submission and before peer-review. There can be important differences between the submitted version and the official published version of record. People interested in the research are advised to contact the author for the final version of the publication, or visit the DOI to the publisher's website.

• The final author version and the galley proof are versions of the publication after peer review.

• The final published version features the final layout of the paper including the volume, issue and page numbers.

Link to publication

General rights

Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain

• You may freely distribute the URL identifying the publication in the public portal.

If the publication is distributed under the terms of Article 25fa of the Dutch Copyright Act, indicated by the “Taverne” license above, please follow below link for the End User Agreement:

www.tue.nl/taverne

Take down policy

If you believe that this document breaches copyright please contact us at:

openaccess@tue.nl

providing details and we will investigate your claim.

(2)

RRC

131

COS

EINDHOVEN UNIVERSITY OF TECHNOLOGY

Department of Mathematics.

. PROBABILITY THEORY, STATISTICS AND OPERATIONS RESEARCH GROUP

Memorandum COSOR 76-26

Strongly convergent dynamic programming: some results

by

K.M. van Hee and J. van der Wal

Eindhoven, December 1976 (Revised February 1977)

(3)

Strbngly convergent dynamic programming: some results

by

K.M. van Hee and J. van der Wal

1. Introduction

In this paper we consider Markov decision processes with respect to the total expected reward criterion. We work under a convergence condition which guarantees that the total expected reward from time n onwards, tends

to zero uniformly in the strategy. This condition is weaker than the con-traction conditons considered by Wessels (1974) and Van Nunen (1976) which are extensions of the discounted model studied by Blackwell (1965). A nice feature of our condition is the fact that convergence of the method of successive approximations can be shown by elementary calculus. see Van Hee, Hordijk and Van der Wal (1977). Here we concentrate the attention on

Howard's policy iteration method and the existence of nearly optimal stationary strategies. Although our results are partially known they seem to be unpublished. Before we formulate our condition in detail. we first sketch the framework of dynamic programming. using notations of Hordijk

(1974). Consider a countable set S. the state space and an arbitrary set A, endowed with a a-field containing all one-point sets, the action space.

There is a transition probability Q from S x A to S, and a reward function

r from 5 x A tom such that r(i,') is measurable for all i E S and if Q(·li,a1)

=

Q(. li,a2) then r(i,a)

=

r(i.a

2). aJ,a2 E A, ~ E S. With Q one

can compose the set

P

of all transition probabilities P from S to S such that. for any i E S P(·\i) = Q(" i ,a) for some a E A. A (Markov) strategy

R may now be defined as a sequence PO.P

I,P2... with Pn EQ)P.n = 0,J,2, . . . .

Each i E: Sand R determine a probabilitylP

i

,

R on (8 x A) and a stochastic proces

{ex

n.A ) • nn = 0.1.2 •..• } where X is the state and An n is the action at time n. (The expectation with respect to P. R is denoted byE. R and if

~, ~.

we omit the subscript i inE. R we mean the function on 5.)

~.

Throughout this paper we assume

Q)

sup]E. R[ ~ r+eX .A ) ] < <Xl for all i E S

R ~'n=O n n

+

(4)

- 2

-As shown in Van Ree (1975) this assumption guarantees that the restriction to pure Markov strategies gives no loss of generality.

On S we define the following functions:

i)

00

v := supE

R[

z:

r(X ,A )] ,the criterion function.

R . n=O n n

ii)

+

for a function s : S -+-lR with suplE

R [s (~)] R N-I v N S := sup lE R [

z:

reX ,A ) + s(~)] R n=O n n < co iii) iv) for a sequence a

:=

(a O,a1,a2,···) of functions an co

wa(i) := sup La (i)

IlE.

Rr(X ,A )

I

,

i E S

R n=O n 1, n n c»

za(i) := sup

l

a (i)lE. R

I

r(X ,A )

I,

i E: S

R n=

°

n 1, n n

s

-+- [1,co),n ... 0,1,2, ••• :

v) w := z := z

a if an - 1 for n'" 0,1,2, •••

"

The conditions we are working with in this paper,state the existence of a sequence a

=

(a

O,a1,a2, ••• ) of functions an : S -+- [1,(0) with an -+- co (pointwise) while still w < 00 .or even z < 00 holds.

a a

We suggest to use the term 'strongly convergent' for models satisfying the weaker

(w

< co) condition.

a

We conclude this section with some notational conventions. It is easy to see that for P E: P there is a function f : S -+- A such that P(j\i) = Q(j!i,f (i» ,

p p

i, j E S and we sometimes write rp(i) := r(i,fp(i». For R

=

(P

O

,Pl,P2' ••• )

we easily obtain:IE. rr(X ,A ) ]

=

PO ••• P )rp (i) • An empty product of elements

1,R" n n

n-of P is defined as the identity operator. n

For two (extended) real functions a a~d b on S we write

t

for the (extended) real function c defined by c(i) :=

:~~~

if b(i) , 0 • With convergence of a sequence of functions on S we mean pointwise convergence; the supremum of a sequence of functions is the pointwise supremum.

2. Standard successive approximations

In this section we present some inequalities which imply, for strongly convergent models, the convergence of the method of successive

approxi-mations. Further we give a sufficient oondition for a Markov decision process to be strongly convergent. For proofs, not given here, we refer to Van Hee,

(5)

- 3 -

)

\

1

1

Hordijk and Van der Wal (1977). In this sec.tion we assume that a

=

(a

O,a1,a2, ••• ) is a nondecreasing sequence of functions an

Theorem 1.

The following holds:

sup R w a a n co and sup

l

R k=n z ~....!. a n Proof: 00 w (i) s a a

(i)

n 00

~

a

\i)

sup

I

ak(i>!lE i

Rr(~,,\)1

s

n R k=n '

The proof of the second inequality is identical.

o

Corollary 1. sup! lERv (X

)1

s R n wa

a'

since n co

Another direct consequence of theorem ] is the following.

Theorem 2.

+

Let s : S -+

m.

be such that lERs (X

n) < 00 for all R then:

w

lvs - vI ~

2:.

+ sup IlERs(x

)! .

n an R n

Hence if a -+ 00 and w < 00 the method of successive approximations converges

n a

to the value function v for any scrapfunction s satisfying suplERs(X

n) -+ 0 • The bound given in theorem 2 is rather rough, which becomes cfear if we set s equal to v and note that VV

=

v for n - 0,1,2, •.•

n

In corollary 2 we give sufficient conditions for scrapfunctions to guarantee convergence:

(6)

."

4

-Corollary 2.

Let an -+ ... and za < .... If the real valued function s satisfies lsI ~ k z

for some k e: JR. we have

w

+

kz

a a

a n

It follows from theorem 1 that the existence of a sequence a with an -+ ... and wa < ... implies that

lim sup

n-+e» R

o .

The following theorem states that this limit property almost implies the existence of such a sequence a a (a

O,a1,a2••• ) •

Theorem 3.

Let w< ... and lim sup

a+'"' R

sequence of functions

1

ImRr(Xk,A

k

)!

=

0, then there is a nondecreasing k=n

a : S -+ [1,"') such that: a -+ ... and w < (Xl

n n a

Finally we remark that our restriction to a countable state space is not essential; it seems that these results carryover to the general case without any difficulty.

3. The policy iteration method

In this section we assume the existence of a nondecreasing sequence of functions a : S -+ [1 ,(Xl) such that w <... In section 2 we have seen that

n a

in this situation the method of successive approximations converges and now we show that the same holds for the policy iteration method given by

Howard (1960). In fact the convergence of both methods is wellknown for the contracting dynamic programming model. The proofs given here are quite simple and usc the same ideas as i.n the contracting case.

We first introduce Howard's iteration method. Let for P e: p R := (P , P, P, ••• ) •

P

3.1 • i) choose Po e: P and define V

o :a 1::1 '£2"" such that I::n oj. O.

m~ [

L

r (X ,A )J, choose a sequence O n n

(7)

n-ii)

5

-Determine P E: P such that

n r p + P v ~ max{sup[rp + Pv 1 - £ e] , v I} n n n-l p n- n n-and define 00 v

n :=lER..- [-1' n=O

L

reX ,A )]n n n

(e is the unit function on S).

In the remainder of this section we show that v converges monotonically

n

to the criterion function v. First we prove two lemma's.

Lemma I.

Proof:

v ~ v

n n-l n = 1,2,3, •••

yields

From 3.1. ii) we have rp

n

+ Pnvn

-I ~ vn-1 • Iterating this equation k times

Since k

t

PR. rp + pkfolv I ~ v I · R.=O n n n n- n-k w \' pi d

I

pk+l

I

a L rp converges to v an . v I S - we get v ~ v I. n n n n- a. n n-R.=O n k+1

o

Obviously v s v . Defining

v

:=

lim v we get

v

S v •

n n n~ Lemma 2. Proof: { + Pv.... } sup r p p

s

V • + P V

n n

=

vn and so, by lemma 1, we have vn ~ r p

n

+ P v I . Hence

n

n-v ~ r p + Pv 1 - E efor all P ( P . Using the monotone convergence theorem

n n- n

(8)

-,

6

-Now we are ready to prove

v

= v • Theorem S.

v

=

v •

Proof:

Since

v

~ v it suffices to show

v

~ v . Let R = (P

O,P1,P2, .•• ) be an

arbitrary strategy. Then, by lemma 2, we get

we have w :=;....! a n (by theorem 1) 00 V ~lE R[n=O

I

reX ,A )] •n n

Since this holds for all

R

the theorem ~s proved.

4. Nearly optimal stationary strategies

In this section we again assume the existence of a nondecreasing sequence

[1

of functions a = (a

O,al ,a2, •••),an : S ~ [1,00) such that an ~ 00 and wa It follows from theorem 5 that there is for each finite subset So c S

and for all £ > 0 a stationary strategy R - (P,P,P, ••• ) such that

00

vR(i) :=lE. R

[L

r(X ,A)J ~ v(i) - £ for i E So .

~, n=O n n

< 00

We show in this section under some additional assumptions the existence of everywhere nearly optimal stationary strategies.

Theorem 6.

w

If -! ~ 0 uniformly on S, then there exists for any £ > 0 a stationary

a n

(9)

7

-Proof:

Choose € > 0, N such

Then iterating this shows

v

~ v R + Ee • w a € that ~ ~

"3

~ inequality N

e and P such that v ~ r +

Pv

+ ~ e •

p 3N

times and using theorem lone easily

o

Under a weaker additional assumption we have a weaker sense of €-optimality.

Theorem 7.

Let a ~ ~ uniformly and

z

< 00 then there exists for any € > 0 a stationary

n a

strategy R such that v

R ~ v - € za

Proof:

3

Choose € > 0 , N such that aN ~

£

and P such that r + Pv ~ v

-P

N-I -I -1

€ {

I

an } Z •

n=O

Iterating this inequality N times yields:

z Slnce P. nZ S a

a

n

00

, I

pnrp

Is

1

Za and pnv

s

j

za we get for R

=

(P,P,P, ••• )

n=N

References

S EZ

a

o

Blackwell, D., (1965) Discounted Dynamic Programming, Ann. Math. Statist. 36, 226-235.

Van Ree, K.M., (1975) Markov strategies in dynamic programming, Eindhoven, University of Technology (Dept. of Math.) Memorandum COSOR 75-20.

Submitted for publication.

Van Hee, K.M., A. Hordijk, J. van der Walt (1977) To appear in the proce~dinRs of the advanc~d seminar on Markov decision theory in Amsterdan, in

(10)

8

-Bordijk, A. (1974) Dynamic prograuming and Markov potential theory.

Amsterdam. Mathematical Centre Tracts, no. 51.

Howard. R.A. (1960) Dynamic programming and Markov processes. Caabridge

(Mass.) M.I.T. Press.

Van Nunen, J.A.E.E. (1976) Contracting Markov decision processes,

Amsterdam. Mathematical Centre Tracts. no. 71.

Wessels, J. (1974) Markov programming by successive approximations with

respect to weighted suppremum norms to appear in: Journ. of

Referenties

GERELATEERDE DOCUMENTEN

Obwohl seine Familie auch iüdische Rituale feierte, folgte daraus also keineswegs, dass sie einer anderen als der deutschen ldentität añgehörte, weder in ethnischer,

Although in the emerging historicity of Western societies the feasible stories cannot facilitate action due to the lack of an equally feasible political vision, and although

If a plant R can be arbitrarily pole assigned by real memoryless output feedback in the sense of Definition 2.3.1, then in particular does there exist a regular feedback law (2.29)

After summarizing some results concerning nonnegative ma- trices, we will give a short overview of the different steps, leading to a more or less complete analysis of the first

1) Synthetic networks: The agreement between the true memberships and the partitions predicted by the kernel spectral clustering model is good for all the cases. Moreover, the

To deal with this issue, and obtain a closed-form expression that yields a more accurate estimate of the convergence of the MEM in the case of closely spaced cylinders, we develop

1) Synthetic networks: The agreement between the true memberships and the partitions predicted by the kernel spectral clustering model is good for all the cases. Moreover, the

Abbreviations: BMI, body mass index; CVID, common variable immunodeficiency disorders; ENT, ear nose throat; ESID, European Society for Immunodeficiencies; HRCT, high