... I would run one to search through every program in numerical order until it found one that reproduced the entire content of the latest English-language Wikipedia pages-articles.xml.bz2 (the current state of every article text). Granted, extracting any sort of a usable AI out of the ensuing mess would almost certainly be harder than writing one from scratch in the first place... but at least it'd be interesting to see just how large it would be.
Comedy option: See if a program reproducing pages-meta-history.xml.7z (all article text and discussion, and all history of both) would be significantly bigger. Incidentally, that file (from the last time anyone actually managed to create one without the dumping process dying in horrible pain at some point) is apparently over five terabytes uncompressed, but 7z brings it down to 31 gigabytes. That's sort of quite impressive actually.