Yes,
I already have solid code to estimate the probabilities and gather the
public estimates.
What I'm stuck on is how to train the "race-wise" logit and then
somehow
combine them to come up with a final set of coefficients.
I could just train a glm on the whole data set, but would be losing the
"race-wise" relationships.
If I follow step 1 of the paper, I wind up with 1000+ logit models
(large training set.)
Now how do I combine them??
Thanks,
-N
On 8/6/09 6:14 PM, Eduardo Leoni wrote:> If I follow it correctly (though I am quite sure I don't) what the
> authors do in the paper is:
>
> 1) Estimate logit models separately per race (using, I assume, horse
> specific covariates.) This step is not described in the attachment you
> sent.
>
> 2) Get (from external data source?) public implied estimates.
>
> 3) Combine the probabilities from model with those from the public.
> These estimates are considered as "data" (that is, the errors in
the
> coefficients are ignored.) The final coefficients \alpha and \beta are
> estimated using a run of the mill multinomial logit model. It is a
> weighted average betwee the two (log) probabilities.
>
> hth,
>
> -eduardo
>
[[alternative HTML version deleted]]