contact@mightymetrika.com

Boot War

August 25, 2023

Intuitive Learning Through Card Play

Boot War is a card game designed to enhance one's understanding of the nonparametric bootstrap test with pooled resampling methods. By diving deep into various game settings, players can derive rich insights. Let's break down the basic gameplay:


1. Select a Mode

Choose 't' for the independent t-test or 'pt' for the paired t-test. In the independent t-test, the effect size is derived from the mean difference between the player's cards and the computer's cards. For the paired t-test, the effect size is calculated as the mean of round-by-round differences between the player and the computer.


2. Define the Deck

Default: A standard 52 card deck with ranked suits. For a twist, employ the R anonymous function to craft an "anonymous deck". The game even supports a bring-your-own-deck feature, known as an "interleaved deck" in Boot War, where different decks are set for the player and the computer.


3. Select a Confidence Level

Set your desired confidence level for the test statistic and effect size. While the default is set at 0.95 (for 95% confidence intervals), it's a crucial component to experiment with.


4. Choose the Number of Bootstrap Resamples

Decide on the count of bootstraps for analysis, integral to the nonparametric bootstrap test with pooled resampling.


5. Select the Number of Rounds

 Opt for any round count between 1 to 26. However, a word of caution: selecting 1 to 3 rounds may lead to crashes. My personal recommendation lies between 5 to 12 rounds, especially if you're interested in observing outcomes in smaller sample sizes.


6. Set a Seed

Consistency is key. By setting a similar random number seed, expect consistent outcomes for identical inputs.


7. Play out the Rounds

Simply hit the 'Deal Card' button to progress through the rounds.


8. Score the Game

When you finish playing out the rounds, the game will employ the nonparametric bootstrap test with pooled resampling to compute the final score, which includes effect size and confidence interval, and to determine the winner.


Final Thoughts

For an enriching experience, I advise players to experiment with an anonymous deck (e.g., function(x) { rpois(20, 15) }) or an interleaved deck (e.g., function(x) { list(rpois(10, 15), rpois(10,10)) }). Once you've set your deck, play around with the mode, confidence level, number of bootstrap rounds, and total rounds to glean insights into the chosen distribution's performance. You might also find it enlightening to think about the side-by-side comparison of the bootstrap p-value and Welch's p-value as you play with the settings.


A tip for enthusiasts: Repetitively playing the game with consistent settings (except for the seed) mimics a manual simulation study, which can offer profound insights if you set the deck up to sample from a distribution that is meaningful to you.


August 19, 2024
Mighty Metrika focuses on statistical methods and mathematics for the analysis of small sample size data. As such, the project runs the risk of people with small sample sizes using tools and methods from mightymetrika.com and becoming over confident in their results because they used "small sample size methods." The long term rigorous goal to combat this disservice is to host citizen science projects, include simulation function in R packages, and share simulation results from the literature and from mightymetrika.com tools through blogs. A short and quick way to combat misuse is through the Who Said It Best series. The series will share some of the best warnings from the small sample size statistical literature. In the Conclusion section of Daniel McNeish's paper Challenging Conventional Wisdom for Multivariate Statistical Models With Small Samples he shares a clear and wonderfully worded warning:
June 25, 2024
This is a quick blog post to list some of the essential resources that I needed to get a citizen science app up and running. The app uses: R Shiny PostgreSQL Pool AWS EC2 The post is basically a way for me to bookmark resources that I found useful and also as a way to say thank you to the folks that put these resources up online.
June 10, 2024
In 'mmibain' v0.2.0, the unit tests are passing at the moment, but on r-devel-linux-x86_64-debian-clang it really seems to be hit or miss. I believe that when the test fails it is do to the new BFfe function which is a case-by-case type implementation of ' bain ' for linear models; however, I used a unit test which relies on a synthetic data set where I generated random numbers and then just used the rep() function to group observations by participants. As such, the data generating process does fit the statistical model and sometimes the random data set that is generated does not make it through bain::bain() without error. I have already changed the unit test and corresponding Roxygen2 documentation example on the Mighty Metrika GitHub and this blog post will walk through the new data and model. But just for further context, here is the original code that sometimes runs through and sometimes throws and error.
More Posts
Share by: