This project implements Quantum Reinforcement Learning using Parameterized Quantum Circuits (PQCs) as policies. It includes both RAW-PQC and SOFTMAX-PQC implementations and compares their performance on various Gymnasium environments.
The implementation supports the following environments:
- CartPole-v1
- MountainCar-v0
- Acrobot-v1
- Parameterized Quantum Circuit (PQC) implementation using Pennylane
- RAW-PQC and SOFTMAX-PQC policy implementations
- REINFORCE algorithm with value function baseline
- Experiment framework for comparing policy types
- Training visualization with confidence intervals
- Python 3.7+
- PyTorch
- Pennylane
- Gymnasium
- NumPy
- Matplotlib
- Clone this repository
- Install dependencies:
pip install -r requirements.txtTo run experiments on all environments:
python main.pyThis will:
- Train both RAW-PQC and SOFTMAX-PQC policies on each environment
- Run multiple independent trials
- Generate training curves with confidence intervals
- Save results as PNG files
pqc_architecture.py: Implementation of the quantum circuitpolicies.py: RAW-PQC and SOFTMAX-PQC policy implementationstrain.py: Training algorithm implementationmain.py: Experiment runnerrequirements.txt: Project dependencies
The training results for each environment will be saved as PNG files:
cartpole-v1_results.pngmountaincar-v0_results.pngacrobot-v1_results.png
This implementation is based on the quantum reinforcement learning architecture described in recent literature (NeurIPS 2021, Parameterized Quantum Policies for Reinforcement Learning), using parameterized quantum circuits for policy representation in reinforcement learning tasks.