by Shiraz » Thu 04 Aug 2005, 00:56:11
$this->bbcode_second_pass_quote('', 'D')id you estimate URR with each local fit, or you kept that parameter fixed?
I kept all parameters (inc. URR) for the comparison curve (the verhulst) fixed. I did not use an estimation of URR besides the implicit one derived from the comparison curve.
$this->bbcode_second_pass_quote('', 'I') can see each period of continuity being a fragment of its own verhulst (or other) curve with its own trajectory.
Yes, this is the idea. I was thinking about using a new comparison curve based upon the same parameter for U, with constant cumulative production, and passing through the post shock data point. This should generate a unique new verhulst comparison curve for each exponential regime. My problem with this is, well, 1) it is too arbitrary. too many random actions. It could fit anything. What's that famous quote... something like, "give me 4 parameters, I'll draw you an elephant, give me five, she'll be doing somersaults" 2) way too much emphasis on the post shock data-point. Why should this data-point exert such a high degree of control over the future threshold for shock?
Alternatively, the new verhulst could be created so that the log of the new verhulst was tangential to the log straight line that is to be the baseline for the next exponential regime. I hope that's kind of clear.
Also, it may be possible, somehow, as khebab suggests, to use the GBM to be the fluid verhulst, so to speak.
With regard to the GBM...
Although it's probably not clear anymore, I actually began working on this model as a reaction to the GBM. When playing around with the GBM, I noticed that exponential shocks were not only the best fitting mathematically, but also the best fits logically as well. The effect of any shock has an exponential effect into the future, because it knocks out the base upon which new production builds. Unfortunately, however, the GBM model proceeds by biasing the underlying model according to tension between a set of exponential curves. So long as the exponential behaviour of the innovators and immitiators functions overwhelms the exponential behaviour of the shocks, everything is fine. But is practice, especially with three or more shocks, what you see is the behaviour of the shocks totally overwhelming the underlying model. I think this is a property of fitting a polynomial type curve with a tension between exponentials. You know the idea about higher degree polynomials being arbitrarily good about fitting as order increases. But of course, the higher the order, then in general, the worse the fit will be outside the domain of the test data (ie. overfitting). Well, I've never heard of an equivalent idea with exponentials, but I imagine some theorem could exist whereby any polynomial of order n can be approximated by an exponential series of n terms over a finite domain. There would also be a corollary, that outside the domain, the difference between the polynomial and the exponential series is itself some exponential function.
To try to bring this back to reality, you can see what I'm getting at in Khebabs GBM chart (above). Look at the 'control model' (purple). It is an approximation of 'Z score of the GBM without shocks' (green). Notice the fit is quite nice inside the fitting domain (ie yr < 2005). But what happens outside this domain? Well we KNOW the trend in the Z score is zero. After all, they are z scores. By comparison the control model appears to rise exponentially. In my experience, this is almost always the case with GBM models. That is also the conclusion I was getting at in the previous paragraph - that ouside the fitting domain, the tension between exponential curves would resolve to an exponential difference between the original trend and the continuation of the fitted curve. The problem with the GBM for me is that the control function always exhibits this 'aberant' behaviour outside the fitting domain.
Now it may be easier to see how the model i presented earlier was in reaction to the GBM. I was trying to capture the exponential differences between shock periods without the tension between exponentials messing up the forecast. When I plotted the log of production, it was apparent that this was going to be much easier than assumed, because the underlying behaviour was exponential anyway. (therefore I could proceed by evolving two parameters, m and b, controlling 'straight lines on the log chart')
This introduced a whole bunch of new negatives, such as the relative arbitrariness of the choice of shock timing, magnitude, and new exponential regime. Recall that I resolved this by introducing with a global verhulst for 'comparison'. It looks like a dog chasing his tail, I know. I'm not sure whether in the end, I gained more or lost more, by comparison with the GBM. If i'm to be honest with myself, I have to imagine I've lost more, because there are so many arbitrary assumptions. It's just that my baby looks beautiful to me... However, if I could just find ways to cut out some of the arbitrariness...
$this->bbcode_second_pass_quote('', 'E')xcel sucks ... there was an article a couple of years on statistical packages and add ons.
Yes, but it's so easy. I like easy. I've started to learn other packages now and then, but I get impatient and go to excel to check my latest ideas. In general, errors on the order of the 8th decimal place don't bother me, and i'll accept these happily for ease of use. The only exception perhaps is non-linear fitting. This is where I prefer to use third party software because this stuff is difficult, and I don't trust excel with it.
At the end of the day, I'm doing this oil modelling stuff for fun. I don't expect my 2nd decimal place to save the world, let alone the 8th.