• English
    • français
    • Deutsch
    • español
    • português (Brasil)
    • Bahasa Indonesia
    • русский
    • العربية
    • 中文
  • English 
    • English
    • français
    • Deutsch
    • español
    • português (Brasil)
    • Bahasa Indonesia
    • русский
    • العربية
    • 中文
  • Login
View Item 
  •   Home
  • OAI Data Pool
  • OAI Harvested Content
  • View Item
  •   Home
  • OAI Data Pool
  • OAI Harvested Content
  • View Item
JavaScript is disabled for your browser. Some features of this site may not work without it.

Browse

All of the LibraryCommunitiesPublication DateTitlesSubjectsAuthorsThis CollectionPublication DateTitlesSubjectsAuthorsProfilesView

My Account

Login

The Library

AboutNew SubmissionSubmission GuideSearch GuideRepository PolicyContact

Statistics

Most Popular ItemsStatistics by CountryMost Popular Authors

PERTURBATION THEORY AND MARKOVIAN DECISION PROCESSES.

  • CSV
  • RefMan
  • EndNote
  • BibTex
  • RefWorks
Author(s)
Schweitzer,Paul J.
Contributor(s)
MASSACHUSETTS INST OF TECH CAMBRIDGE OPERATIONS RESEARCH CENTER
Keywords
(*PERTURBATION THEORY
STATISTICAL PROCESSES)
(*STATISTICAL PROCESSES
PERTURBATION THEORY)
(*DYNAMIC PROGRAMMING
STATISTICAL PROCESSES)
OPERATIONS RESEARCH
PROBABILITY
MATRICES(MATHEMATICS)
ITERATIONS
MARKOV RENEWAL PROGRAMMING
Show allShow less

Full record
Show full item record
URI
http://hdl.handle.net/20.500.12424/1110909
Online Access
http://oai.dtic.mil/oai/oai?&verb=getRecord&metadataPrefix=html&identifier=AD0618406
Abstract
The Howard-Jewell algorithm for programming over a Markov-renewal process is analyzed in terms of a perturbation theory formalism which describes how the stationary distribution changes when the transition probabilities change. The policy improvement technique is derived from this new viewpoint. The relative values may be interpreted as partial derivatives of the gain rate with respect to policy. The value equations are shown to be solvable, with the relative values unique up to one additive constant, if and only if the underlying Markov chain is irreducible. The policy iteration algorithm is shown not to cycle, this guaranteeing convergence. A discussion of the existence, uniqueness, and characterization of the solution to the functional equation of dynamic programming is given. Emphasis is placed upon the value maximization of transient states. The fundamental matrix is developed as a useful tool for doing perturbation theory, describing firstpassage properties of semi-Markov processes, and for dealing with semi-Markov processes with rewards. (Author)
Rept. on Research in the Control of Complex Systems.
Date
1965-06
Type
Text
Identifier
oai:AD0618406
http://oai.dtic.mil/oai/oai?&verb=getRecord&metadataPrefix=html&identifier=AD0618406
Copyright/License
APPROVED FOR PUBLIC RELEASE
Collections
OAI Harvested Content

entitlement

 
DSpace software (copyright © 2002 - 2022)  DuraSpace
Quick Guide | Contact Us
Open Repository is a service operated by 
Atmire NV
 

Export search results

The export option will allow you to export the current search results of the entered query to a file. Different formats are available for download. To export the items, click on the button corresponding with the preferred download format.

By default, clicking on the export buttons will result in a download of the allowed maximum amount of items.

To select a subset of the search results, click "Selective Export" button and make a selection of the items you want to export. The amount of items that can be exported at once is similarly restricted as the full export.

After making a selection, click one of the export format buttons. The amount of items that will be exported is indicated in the bubble next to export format.