AAMAS Conference 2017 Conference Paper
Dynamic Generalization Kanerva Coding in Reinforcement Learning for TCP Congestion Control Design
- Wei Li
- Fan Zhou
- Waleed Meleis
- Kaushik Chowdhury
Traditional reinforcement learning (RL) techniques often encounter limitations when solving large or continuous stateaction spaces. Training times needed to explore the very large space are impractically long, and it can be difficult to generalize learned knowledge. A compact representation of the state space is usually generated to solve both problems. However, simple state abstraction often cannot achieve the desired learning quality, while expert state representations usually involve costly hand-crafted strategies. We propose a new technique, generalization-based Kanerva coding, that automatically generates and optimizes state abstractions for learning. When applied to adapting the congestion window of the highly complex TCP congestion control protocol, a standard Internet protocol, this technique outperforms the current standard-TCP New Reno by 59. 5% in throughput and 6. 5% in delay. Our technique also achieves a 35. 2% improvement in throughput over the best previously proposed Kanerva coding technique when applied in the same context.