search for: markov

Displaying 20 results from an estimated 332 matches for "markov".

Did you mean: marko
2016 Jun 30
2
Shared mailboxes not showing up in shared namespace
Hi, I think I have configured everything correctly but for some reason I can?t get a list of the shared mailboxes to show up. When I run: doveadm acl debug -u m.markov Shared/d.marteva/INBOX doveadm(root): Debug: Skipping module doveadm_fts_plugin, because dlopen() failed: /usr/lib/dovecot/modules/doveadm/lib20_doveadm_fts_plugin.so: undefined symbol: fts_backend_rescan (this is usually intentional, so just ignore this message) doveadm(m.markov): Debug: Added u...
2008 Feb 12
1
Markov and Hidden Markov models
Hi, Is there a package that will estimate simple Markov models and hidden Markov models for discrete time processes in R? Thanks in advance, David -- =============================================================== David Kaplan, Ph.D. Professor Department of Educational Psychology University of Wisconsin - Madison Educational Sciences, Room, 1061 102...
2012 Mar 02
2
回复: Bayesian Hidden Markov Models
Dear Oscar,   Thanks for your help.It's so nice of you to explain this package to me.   Best Regards,   James LAN 发件人: Oscar Rueda [via R] <ml-node+s789695n4431468h14@n4.nabble.com> 收件人: monkeylan <lanjinchi@yahoo.com.cn> 发送日期: 2012年2月29日, 星期三, 下午 9:21 主题: Re: Bayesian Hidden Markov Models Dear James, The distances are normalized between zero and 1, so in your case all of them will be zero. You can check that with > res$Dist.for.model And do > Q.NH(summary(res)[[1]]$beta, x=0) To obtain the common transition matrix. Cheers, Oscar On 29/2/12 03:59, &quo...
2012 Jul 27
1
fitting Markov Switching Model
Dear Users, i have this time series, the tree lines means different level, i would use a Markov switching model with two states to modelling this time series. i would obtain the relative transition matrix (2X2) the first state is above the value of 23.65 (the higher line) the second state is below the value of 23.65 You can ignore the other two lines http://r.789695.n4.nabble.com/file/n4638...
2016 Jul 02
2
Shared mailboxes not showing up in shared namespace
...oChildren) "/" INBOX 7 OK List completed. By connecting using `openssl` from a remote machine. > On 1 Jul 2016, at 09:02, Steffen Kaiser <skdovecot at smail.inf.fh-brs.de> wrote: > > -----BEGIN PGP SIGNED MESSAGE----- > Hash: SHA1 > > On Thu, 30 Jun 2016, Marti Markov wrote: > >> I think I have configured everything correctly but for some reason I can?t get a list of the shared mailboxes to show up. >> >> When I run: >> >> doveadm acl debug -u m.markov Shared/d.marteva/INBOX >> >> >> doveadm(m.markov): Info...
2007 Oct 30
2
markov regime switching models
Hi, I am looking for a package to estimate regime switching models (states following a markov chain). I found packages for Hidden Markov Models but I am looking for something a little different: In the HMM the conditional distribution of the observations (give the state) is a known distribution (normal or others), while the package I need should allow to set a conditional distribution (give...
2010 Oct 26
1
Markov Switching with TVTP - problems with convergence
Greetings fellow R entusiasts! We have some problems converting a computer routine written initially for Gauss to estimate a Markov Regime Switching analysis with Time Varying Transition Probability. The source code in Gauss is here: http://www.econ.washington.edu/user/cnelson/markov/programs/hmt_tvp.opt We have converted the code to R, and it's running without errors, but we have some convergence problems. According to th...
2003 Jun 25
2
Markov chain simulation
Hi, Does anybody know a function to simulate a Markov chain given a probability transition matrix and an initial state ? Thanks. Philippe -- -------------------------------------------------- Philippe Hup? Institut Curie - Equipe Bioinformatique 26, rue d'Ulm - 75005 PARIS France +33 (0)1 42 34 65 29 Philippe.Hupe at curie.fr <mailto:Phil...
2001 Nov 07
3
Examples for Markov Chain in Economics
Could anyone tell me where can I find some examples of the applications to economics of a Markov chain? Many thanks in advance. Luis Rivera. -.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.- r-help mailing list -- Read http://www.ci.tuwien.ac.at/~hornik/R/R-FAQ.html Send "info", "help", or "[un]subscribe" (in the "body",...
2008 Nov 11
1
R: R: Hidden Markov Models
...s a time series length constraint that I don't quite understand. Thank you in advance for your attention. Kind regards, Maura Edelweiss -----Messaggio originale----- Da: Walter Zucchini [mailto:wzucchi@uni-goettingen.de] Inviato: mar 11/11/2008 11.32 A: mauede@alice.it Oggetto: Re: R: Hidden Markov Models Dear Ms Monville, Hidden Markov models (HMMs), and that includes the msm implementation, are not based on the assumption that the observations are independent. Indeed HMMs are specifically designed to model serially dependent observations. Of course that doesn't mean that they can...
2004 Apr 29
3
Probability(Markov chain transition matrix)
Hello, My name is Maria, MBA student in Sanfransisco, USA. In my credit scoring class, I have hard time making "transition matrix", which explains probability especially in relation to "Markov chain model". It is regarding people's monthly credit payment behavior. Does R have function to caculate it? I am actually a novice in using 'R'. Please help me!!! Maria Gu
2012 Mar 25
2
Updating a Markov Chain
Hello, my question is if anyone has any good ideas how to create a Markov Chain from ordered data. So, I have some sort of time series, and if value1 happens as time1 and value2 happens at time2 I record this as an update to the probability transition matrix. The problem is that I cannot predefine the size of the matrix (as I don't know how many states(values) I will...
2009 Mar 03
1
spatial markov chain methods
Hello, can any one point me to R-packages (if available) which include spatial Markov Chain methods? My second question is more general but hopefully not OT: Currently we are using the software TPROGS, which let people simulate property distributions in space by some Markov Chain approaches. We face some problems due to the lack of information between distances of samples along...
2006 Jan 13
1
multivariate markov switching
Dear helpers, Does anyone know about a package or a function that allows to estimate Multivariate Markov-Switching Models, like MS-VAR as introduced by Krolzig(1997) with R ? Thanks a lot!! Carlo
2009 May 09
1
R package for estimating markov transition matrix from observations + confidence?
Dear R gurus, I have data for which I want to estimate the markov transition matrix that generated the sequence, and preferably obtain some measure of confidence for that estimation. e.g., for a series such as 1 3 4 1 2 3 1 2 1 3 4 3 2 4 2 1 4 1 2 4 1 2 4 1 2 1 2 1 3 1 I would want to get an estimate of the matrix that generated it [[originally: [,1] [,2]...
2009 Nov 15
1
how to permute, simulate Markov chain
Hi all, I am new to R. Can someone please give me some hints in how to do the following things: 1- Get ONE permutation of a set. I have looked at the gregmisc package's permutations() method, but I just want to get one permutation at a time. 2- Simulate a Markov chain in R. For instance, I want to simulate the simple random walk problem, in which a person can walk randomly around 4 places. I know how to set up the transition matrix in R. I'm stuck at what to do next. I'm grateful if someone can give me hint or a pointer. Thanks. Martin -- View...
2011 Feb 25
1
Markov chain transition model, data replication project
Hello all, I am currently attempting to replicate data from a political science article that utilized a Markov chain transition model to predict voter turnout intention at time *t*; the data was separated into two different models based on whether prior intent was to vote or not to vote. The details don't really matter. Mostly I am curious how to run a Markov chain transition model in R, estimating di...
2010 Sep 17
1
Markov Model problem
...'t really find a solution for my problem. And maybe I formulate it incorrectly, so bear with me. How would I calculate a 'constant transition matrix' if I know a given value at a given time? Let's say I know that my value is 54,0 at t=12. How do I get the initial chain value? t markov.growth 00 00.0 01 06.3 <- My value of interest... 02 12.1 03 17.7 04 22.8 05 27.7 06 32.2 07 36.4 08 40.4 09 44.2 10 47.7 11 50.9 12 54.0 <- These two is what I know... 13 56.9 14 59.6 15 62.1 I can't find out how to do this kind of 'reversed' calulation. Maybe there is some libr...
2006 Jan 22
6
Making a markov transition matrix
Folks, I am holding a dataset where firms are observed for a fixed (and small) set of years. The data is in "long" format - one record for one firm for one point in time. A state variable is observed (a factor). I wish to make a markov transition matrix about the time-series evolution of that state variable. The code below does this. But it's hardcoded to the specific years that I observe. How might one generalise this and make a general function which does this? :-) -ans. set.seed(1001) # Raw data in long for...
2003 Oct 01
3
fitting Markov chains
I need to find a computationally simple process for the movement of interest rates. In this simplified model, an interest rate can have 3--5 possible values, and its movement is characterized by a matrix of transition probabilities (ie, it is a Markov process). I would like to estimate this process from a given set of data. For example, let the interest rate time series be: 7 3 8 2 5 9 6 Assume that the discretized variable can take the following values: (3, 5, 8), then we find the nearest discrete point and give its index: 3 1 3 1 2 3 2...