The Imperial College Algorithmic Trading Conference was Saturday.
Talks
Massoud Mussavian
Massoud gave a great talk on “Algo Evolution”. It started with a historical review of how trading used to be done “by hand”. It culminated in a phylogenetic tree of trading algorithms. There was an herbivore branch and a carnivore branch.
Robert Macrae
Robert talked about risk. He put it into a 2-by-2 table:
- Are you measuring or controlling?
- Are you thinking of normal times or extreme times?
Measuring for normal times is easy. Easy in quotation marks.
Controlling risk in normal times is possible, but there are some tricky bits.
Measuring extremes is too hard. We can get estimates but the variability of those estimates is extremely high. To get practical value we have to impose prior beliefs.
Controlling extreme risk is: “Err …” It is basically impossible. On the other hand it is essential for civilization to continue. In Robert’s words, “If the ATMs stop working, things are going to get rough.” Robert also pointed out that it is very easy to do things in the name of controlling extreme risk that will make it worse.
Panos Parpas
Panos talked on “Multiscale Stochastic Volatility”. Stochastic volatility is the continuous-time cousin of garch (see below). However, the trail to stochastic volatility was hard to see. I found the talk interesting because I saw some mathematics and computing that I know nothing about.
Keith Quinton
Keith talked about 130/30 funds. These in his view solve a few problems. In particular, they allow the fund manager to be more active, and allow bets on stocks going down (which tend to work better than positive bets) to be fully realized.
He emphasized that 130/30 is a portfolio construction device, it is not an alpha generator. In fact, if you don’t have alpha and you do 130/30, you get a whole lot more of nothing.
The 130/30 form has fallen out of favor with investors because performance has happened to be not so great. Keith thinks that a restart is called for.
Pat Burns
I talked about “3 realms of garch modelling” (pdf). A companion to the talk is the blog post “A practical introduction to garch modeling”. This points to another blog post on variance targeting, which is also discussed in the talk. The talk concludes with an argument for doing financial research in R.
High Frequency Trading Panel
There was a lively debate about the topic:
- Do we know what “high frequency trading” means? No.
- Does it provide liquidity? Maybe, maybe not.
- Is it dangerous? Various opinions.
I favor controlling against extreme risk. What a pity that it’s seemingly impossible. I liked civilization.
Bill Alpert
Barron’s
Perhaps I put it a bit strongly, there are plenty of sensible things you can do after an event to support stability. For example, if money gets distributed really badly then you can print lots more of it and wash the mistakes away!
My intended point was that it is verging on impossible to prevent a future financial crisis by deducing control rules from past historic data. Fancy stats are completely inappropriate, what is required is to make very simple deductions — eg that TBTFs, complexity and leverage are bad — and then move on to the more difficult task of taking steps to restrict them.
Or perhaps my precis was a little too concise.
It seems that I have slandered Alexios, the author of the
rugarch
package. The components garch model is already implemented. Since libel reform has not yet gone through in the UK, I could be at risk.Based on your Garch paper I think some of my concerns over Garch have been misplaced. At any point we get
1) a sensible estimate for both what vol will be in the absence of a new spike;
2) a probability for a new spike;
3) a T distribution on the size of the spike if it happens.
I think its data intensive mainly because spikes are rare. Because of its importance you might wish to shrink 3) towards something like the unconditional distribution… maybe.
I think that is a very interesting comment.
I recall suspecting that the distribution of the standardized residuals of real data didn’t match the estimated t distribution very well.
This would be a mechanism to explain such differences.
However, in Figure 3 of “Variability of garch estimates” we see that the degrees of freedom estimates for simulated data look pretty unbiased — they just aren’t overly precise.