I Trained an AI to Beat Final Fight… Here’s What Happened [p]
![I Trained an AI to Beat Final Fight… Here’s What Happened [p]](/_next/image?url=https%3A%2F%2Fexternal-preview.redd.it%2FdWZSKa_lMUvycB0q8xwsIkTgDpHLe-W2-Q_S7RwWucQ.jpeg%3Fwidth%3D320%26crop%3Dsmart%26auto%3Dwebp%26s%3D6a0cfa02507091d2949c8f6b2fcd59a254a23929&w=3840&q=75)
| Hey everyone, I’ve been experimenting with Behavior Cloning on a classic arcade game (Final Fight), and I wanted to share the results and get some feedback from the community. The setup is fairly simple: I trained an agent purely from demonstrations (no reward shaping initially), then evaluated how far it could go in the first stage. I also plan to extend this with GAIL + PPO to see how much performance improves beyond imitation. A couple of interesting challenges came up:
The agent can already make some progress, but still struggles with consistency and survival. I’d love to hear thoughts on:
Here’s the code if you want to see the full process and results: Any feedback is very welcome! [link] [comments] |
Want to read more?
Check out the full article on the original site