Andrew Burkard

Projects

PMTDB A searchable database of 26,000+ takes from Pardon My Take, scored for accuracy using baseball-style metrics. Built because nobody was holding podcast hosts accountable and someone had to.
March Madness Annual Kaggle competitor since 2019. Fully Bayesian hierarchical model in NumPyro/JAX — think KenPom, but probabilistic — adjusting for travel distance, altitude, and referee tendencies. Best finish: gold medal in 2021 (4th/451), the one year they didn't award a cash prize, so all I got was a water bottle.
Banana Browser A web browser that doesn't render HTML. It asks an AI to imagine what the page looks like instead. You click on the generated image, the AI interprets your click, and generates the next page. It's worse in every way and that's the point.
Language Models (2018) Trained ~100M parameter transformers on ProFootballTalk comments, then Reddit, before "GPT-1" was even a thing. Deployed as Twitter bots on t2.small CPU instances, generating surprisingly coherent replies in under 2 seconds. Elon Musk even followed one briefly — then he bought the platform, promising to ban all the bots.
sklearn-prg Precision-Recall-Gain curves for scikit-learn, because regular PR curves lie to you when your classes are imbalanced. pip install sklearn_prg.
What Do They Know About Me? See how any website can uniquely identify you without a single cookie.

old school projects

Examining NBA Crunch Time Modeled end-of-game basketball as a two-player zero-sum game to figure out whether a team down 4 should shoot a 2 or a 3. Ran thousands of simulations to mathematically confirm what every coach already believed.
Automatically Generating ProFootballTalk Comments Built a Markov chain and a pointer-generator network to auto-generate PFT comments — a corpus famously described as "didactic misspelling and arrogant disdain for critical thought." The neural net mostly learned to copy-paste usernames.
Bayesball Bayesian models for pitcher FIP components using Dirichlet-Multinomial conjugacy and Gibbs sampling. Beat the standard FIP at predicting future ERA, which is a very satisfying result for a very small audience.
Single Image Super Resolution Used sparse representations to upscale blurry satellite images. Turns out a dictionary trained on flowers works about as well as one trained on satellites, which is either profound or depressing.
Comparing League Strength via the Soccer Power Index Applied permutation tests to FiveThirtyEight's SPI to settle which European league is best. Found no significant difference between the top four, which will satisfy absolutely no one on soccer Twitter. Also confirmed that England's second division is statistically better than MLS.
GDELT News Analysis Analyzed 90 GB of global media event data to track sentiment around the 2016 election. Discovered that GDELT's sentiment algorithm thinks election night was a positive event for Hillary Clinton, which tells you most of what you need to know about sentiment analysis.