Citation Relationships

Legends: Link to a Model Reference cited by multiple papers


Morimoto J, Doya K (2007) Reinforcement learning state estimator. Neural Comput 19:730-56 [PubMed]

References and models cited by this paper

References and models that cite this paper

Arulampalam MS, Maskell S, Gordon N, Clapp T (2002) A tutorial on particle filters for online nonlinear-non-gaussian bayesian tracking IEEE Trans Signal Process 50:174-188
Baird LC, Moore AW (1999) Gradient descent for general reinforcement learning Advances in neural information processing systems, Kearns MS:Solla S:Dohn D, ed.
Barto AG, Sutton RS, Anderson CW (1983) Neuronlike elements that can solve difficult learning control problems IEEE Trans Systems Man Cybern 13:835-846
Crassidis JL, Junkins JL (2004) Optimal estimation of dynamic systems
Doucet A, Godsill S, Andrieu C (2000) On sequential Monte Carlo sampling methods for Bayesian filtering Stat Comput 10:197-208
Doya K (2000) Reinforcement learning in continuous time and space. Neural Comput 12:219-45 [PubMed]
Erdogmus D, Genc AU, Principe JC (2002) A neural network perspectiveto extended Luenberger observers Institute Of Measurement And Control 35:10-16
Goswami A, Thuilot B, Espiau B (1996) Compass-like biped robot part I: Stability and bifurcation of passive gaits Tech Rep No RR-2996 INRI
Jaakkola T, Singh SP, Jordan MI (1995) Reinforcement learning algorithm for partially observable Markov decision problems Advances in neural information processing systems, Tesauro G:Touretzky D:Leen T, ed. pp.345
Kalman RE, Bucy R (1961) New results in linear filtering and prediction Trans ASME J Basic Eng 83:95-108
Kimura H, Kobayashi S (1998) An analysis of actor-critic algorithms using eligibility traces: Reinforcement learning with imperfect value functions Proc 15th Intl Conf Mach Learn :284-292
Littman ML, Cassandra AR, Kaelbling LP (1995) Learning policies for partially observable environments: Scaling up Proc 12th Intl Conf Mach Learn :363-370
Luenberger DG (1971) An introduction to observers IEEE Trans 16:596-602
McCallum RA (1995) Reinforcement learning with selective perception and hidden state Unpublished doctoral dissertation, University of Rochester
Meuleau N, Kim KE, Kaelbling LP (2001) Exploration in gradient-based reinforcement learning Tech Rep MIT
Meuleau N, Peshkin L, Kim KE, Kaelbling LP (2000) Learning finite state controllers for partially observable environments Proc 15th Ann Conf Uncertainty in Artificial Intelligence :427-436
Morimoto J, Doya K (2002) Development of an observer by using reinforcement learning Proc 12th Annual Conf Japn Neural Network Soc :275-278
Porter LL, Passino KM (1995) Genetic adaptive observers Engineering Applications of Artificial Intelligence 8:261-269
Raghavan IR, Hedrick JK (1994) Observer design for a class of nonlinear systems International Journal Of Control 59:515-528
Thau FE (1973) Observing the state of nonlinear dynamic systems Intl J Control 17:471-479
Thrun S (2000) Monte Carlo POMDPs Advances in neural information processing systems, Solla SA:Leen TK:Muller KR, ed. pp.1064
Wan E, van_der_Merwe R (2000) The unscented Kalman filter for nonlinear estimation Proc IEEE Symposium
Wan EA, Nelson AT (1997) Dual Kalman filtering methods for nonlinear prediction, smoothing, and estimation Advances in neural information processing systems, Mozer M:Jordan M:Petsche T, ed. pp.793
Xu JW, Erdogmus D, Principe JC (2005) Minimum error entropy Luenberger observer Proc Am Control Conf
(25 refs)