# Wolverine ## About Give your python scripts regenerative healing abilities! Run your scripts with Wolverine and when they crash, GPT-4 edits them and explains what went wrong. Even if you have many bugs it will repeatedly rerun until it's fixed. For a quick demonstration see my [demo video on twitter](https://twitter.com/bio_bootloader/status/1636880208304431104). ## Setup python3 -m venv venv source venv/bin/activate pip install -r requirements.txt cp .env.sample .env Add your openAI api key to `.env` _warning!_ by default this uses GPT-4 and may make many repeated calls to the api. ## Example Usage To run with gpt-4 (the default, tested option): python wolverine.py buggy_script.py "subtract" 20 3 You can also run with other models, but be warned they may not adhere to the edit format as well: python wolverine.py --model=gpt-3.5-turbo buggy_script.py "subtract" 20 3 If you want to use GPT-3.5 by default instead of GPT-4 uncomment the default model line in `.env`: DEFAULT_MODEL=gpt-3.5-turbo ## Future Plans This is just a quick prototype I threw together in a few hours. There are many possible extensions and contributions are welcome: - add flags to customize usage, such as asking for user confirmation before running changed code - further iterations on the edit format that GPT responds in. Currently it struggles a bit with indentation, but I'm sure that can be improved - a suite of example buggy files that we can test prompts on to ensure reliability and measure improvement - multiple files / codebases: send GPT everything that appears in the stacktrace - graceful handling of large files - should we just send GPT relevant classes / functions? - extension to languages other than python ## Star History [![Star History Chart](https://api.star-history.com/svg?repos=biobootloader/wolverine&type=Date)](https://star-history.com/#biobootloader/wolverine)