Saturday, July 19, 2008

Antarctica

Watched "Encounters at the Bottom of the World". Also saw <
http://www.emilystone.net/uploads/Cold_Comfort.pdf >

Friday, June 13, 2008

A Remarkable Tornado Photo

"The Lede: A Remarkable Tornado Photo"

An Iowa woman snaps a stunning close-up of a looming funnel cloud.

Kalman filters

Consider these for noise rejection, especially when moving to continuous-time data.

Wednesday, May 07, 2008

FFT Window references

http://www.lds-group.com/docs/site_documents/AN014%20Understanding%20FFT%20Windows.pdf
http://www.bores.com/courses/advanced/windows/files/windows.pdf

The first link has a table with window characteristics:

Window / Best for these Signal Types / Frequency Resolution / Spectral Leakage / Amplitude Accuracy
Barlett: Random / Good / Fair / Fair
Blackman: Random or mixed / Poor / Best / Good
Flat top: Sinusoids / Poor / Good / Best
Hanning: Random / Good / Good / Fair
Hamming: Random / Good / Fair / Fair
Kaiser-Bessel: Random / Fair / Good / Good
None (boxcar): Transient & Synchronous Sampling / Best / Poor / Poor
Tukey: Random / Good / Poor / Poor
Welch: Random / Good / Good / Fair

A. O. Scott: Here Comes Everyboy, Again

"Mr. Sandler did not invent the archetype of the overgrown man-child, which has been around at least since the silent era. ... Nor has Mr. Sandler been alone, over the past 15 years or so, in turning male infantile aggression into the basis of a lucrative and long-running movie career. His rivals and confreres have included Jim Carrey, Jack Black and Will Ferrell — all of them different physical and temperamental types, but all of them committed to a brazen and unyielding refusal of maturity."

Add to this list the television comedies Seinfeld, Friends, Everyone Loves Raymond, and most recently, King of Queens. I had wondered about the provenance of this archetype, and the exact nature of the thing that so repulsed me about those shows, until this essay put a name to the phenomenon.

Thursday, April 03, 2008

DoG is used in wavelets

Just noticed that derivative of gaussian is also called "DoG" when used as a wavelet basis function. But I think Daubechies are more popular.
Are you using those in your price models?
Maybe in the future. One of my coworkers is using Daubechies wavelets in his model; compared with ordinary moving-average smoothing, doing a wavelet transform, cutting off the spectrum, and then transforming back seems to offer better noise reduction with less information. You still need a power-of-two data length, like for fast Fourier transforms.

Recently, I have been using techniques from "robust statistics": medians, trimmed data, etc. This reminded me of our early experiences with median filtering. Now I think that, while the technique may be nonlinear, the point of experimental science is to make a reliable measurement. If the number itself is what we want to measure (or even ordinary arithmetic transformations thereof), then I say filter. I think the only case where I wouldn't automatically reach for a trimmed mean (where the top and bottom x% of observations are excluded before taking the mean) is if I were doing transformation to a different space, such as FFT. But for dynamic pulling experiments, I suspect FFT is a rarely-used technique.

You probably know that the reporting of a standard deviation is only meaningful if the data are close to normal. Especially in the kind of data I have, I am plagued by fat-tailed distributions, and find that ordinary means and variances are too sensitive to outliers. The robust alternative is the MAD (median absolute deviation).

Wednesday, April 02, 2008

Database Monte Carlo (DBMC): A New Strategyfor Variance Reduction in Monte Carlo Simulation

Monday, March 31st,2008
10:00 am - 11:00 am

CNLS Conference Room (TA-3, Bldg 1690)

DataBase Monte Carlo (DBMC): A New Strategy for Variance Reduction in Monte Carlo Simulation

Professor Pirooz Vakili
Boston University

A well-known weakness of (ensemble) Monte Carlo is its slow rate of convergence. In general the rate of convergence cannot be improved upon, hence, since the inception of the MC method, a number of variance reduction (VR) techniques have been devised to reduce the variance of the MC estimator. All VR techniques bring some additional/external information to bear on the estimation problem and rely on the existence of specific problem features and the ability of the user of the method to discover and effectively exploit such features. This lack of generality has significantly limited the applicability of VR techniques.

We present a new strategy, called DataBase Monte Carlo (DBMC), which aims to address this shortcoming by divising generic VR techniques that can be generically applied. The core idea of the approach is to extract information at one or more nominal model parameters and use this information to gain estimation efficiency at neighboring parameters. We describe how this strategy can be implemented using two variance reduction techniques: Control Variates (CV) and Stratification (DBMC approach can be used more broadly and is not limited to these two techniques). We show that, once an initial setup cost of generating a database is incurred, this approach can lead to dramatic gains in computational efficiency. DBMC is quite general and easy to implement -- it can wrap existing ensemble MC codes. As such it has potential applications, among others, in ensemble weather prediction, hydrological source location, climate and ocean, optimal control, and stochastic simulations of biological systems.

We discuss connections of the DBMC approach with the resampling technique of Bootstrap and the analysis approach of Information Based Complexity.

LANL Host: Frank Alexander, ADTSC

Monday, February 11, 2008

Liars' Poker

After reading Michael Lewis's Liars' Poker, and Roger Lowenstein's When Genius Failed, I am struck with a thought: make a program to play Liars' Poker. Either decimal or dice systems. Bayesian estimation of probabilities. Python.