RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, const
The pull request is 3 attempts at 2 seconds each, but this is up for discussion. I do not know how much time is needed to get a near guarantee of directory deletion on Windows, but there is a decent chance we can run it for only 3 attempts at 1 second per. ...
I immediatelyembarked on an experiment which, even though iffailed toget rid ofthe ants, kept me fascinated fortwenty-four hours. I bound the base of the tree withsticky tape, makingit impossible for the ants to reach the aphides. The tape was sostick that they did not dare to cross ...