Skip to content

DBarbosaDev/QRLIT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

QRLIT is a Quantum-classical algorithm version of SMARTIX. This version is based on the available source code in repository rl-db-indexing (https://github.com/Chotom/rl-db-indexing)

Cite this work with:

@Article{fi16120439,
	AUTHOR 		= {Barbosa, Diogo and Gruenwald, Le and D’Orazio, Laurent and Bernardino, Jorge},
	TITLE 		= {QRLIT: Quantum Reinforcement Learning for Database Index Tuning},
	JOURNAL 	= {Future Internet},
	VOLUME 		= {16},
	YEAR 		= {2024},
	NUMBER 		= {12},
	ARTICLE-NUMBER 	= {439},
	URL 		= {https://www.mdpi.com/1999-5903/16/12/439},
	ISSN 		= {1999-5903},
	ABSTRACT 	= {Selecting indexes capable of reducing the cost of query processing in database systems is a challenging task, especially in large-scale applications. Quantum computing has been investigated with promising results in areas related to database management, such as query optimization, transaction scheduling, and index tuning. Promising results have also been seen when reinforcement learning is applied for database tuning in classical computing. However, there is no existing research with implementation details and experiment results for index tuning that takes advantage of both quantum computing and reinforcement learning. This paper proposes a new algorithm called QRLIT that uses the power of quantum computing and reinforcement learning for database index tuning. Experiments using the database TPC-H benchmark show that QRLIT exhibits superior performance and a faster convergence compared to its classical counterpart.},
	DOI 		= {10.3390/fi16120439}
}

Setup and training with Docker (Based on https://github.com/Chotom/rl-db-indexing)

This should only be used for testing or if you are certain, that you can provide stable server performance.

  1. Start mysql server and client.

    docker compose up -d
    docker compose up -d --build # if docker images require to be rebuilt or created
  2. Generate data and load database to mysql_server from client container.

    docker compose exec client python3 /project/cli/initiate_environment.py
  3. You can start training by running script in client container

    docker compose exec client python3 /project/cli/run_quantum_train.py

Database indexes tuning with agent using reinforcement learning techniques.
TPC-H benchmark MySQL environment for reinforcement learning.

Source code for paper: https://doi.org/10.1007/978-3-031-42941-5_45

License


Setup and training

How to train an agent and set up an environment for TPC-H database.

  1. In db_env/tpch/tpch_tools directory paste tpc-h-tool.zip downloaded from TPC website (tested with version 3.0.0): download link.

  2. Define database size (in GB) and number of parallel streams in db_env/tpch/config.py.

    You can check recommended number of streams in TPC-H documentation here (page 96).

    SCALE_FACTOR = 0.1   # Should be >=1, however we used 0.1 because of limited resources
    STREAM_COUNT = 2     # 2 for 1 GB, we also used 2
  3. Follow further instruction:


Setup and training with Docker

This should only be used for testing or if you are certain, that you can provide stable server performance.

  1. Start mysql server and client.

    docker compose up -d
    docker compose up -d --build # if docker images require to be rebuilt or created
  2. Generate data and load database to mysql_server from client container.

    docker compose exec client python3 /project/cli/initiate_environment.py
  3. You can start training by running script in client container

    docker compose exec client python3 /project/cli/run_train.py

Server setup without Docker

  1. Install and configure mysql server (version 8).

  2. Allow remote root connection for database operations during benchmarking (or create different user):

    CREATE USER 'root'@'CLIENT_IP'  /* change CLIENT_IP to client IP address, or use 'root'@'%' for all addresses */
    IDENTIFIED WITH caching_sha2_password BY '1234';
    GRANT ALL PRIVILEGES ON *.* TO 'root'@'CLIENT_IP';  /* change CLIENT_IP */

Client setup without Docker

Project currently only supports Linux client, because of DBGen. However, it should be possible to compile DBGen on Windows as well.

  1. Update file .env with MySQL server data (user, password, server IP, port and name of the database, which will be used for benchmarking).

  2. Install gcc, make, python (minimal version is 3.8), pip and virtualenv.

  3. Move project to desired folder (for example /path/to/project).

  4. Create virtual environment (for example inside /path/to/project directory):

    virtualenv venv
  5. Save project path to $PYTHONPATH:

    echo 'export PYTHONPATH="${PYTHONPATH}:/path/to/project/"' >> venv/bin/activate
  6. Activate virtual environment:

    source venv/bin/activate
  7. Install requirements:

    pip install -r requirements.txt
  8. Patch DBGen:

    python cli/patch_dbgen.py
  9. Compile DBGen:

    make -C dbgen
  10. Initiate environment - populate database with generated data:

    python cli/initiate_environment.py

The environment is now ready and you can:

  • train agent, for example:

    nohup python cli/train_agent.py &> data/train.log &

    It is possible to stop training at any time (without need of resetting environment) by sending the script SINGINT (CTRL C). In case of example way of running training shown above, it can be done with:

    kill -2 PID   # change PID to process id (can be established using 'ps' command)
  • run single benchmark on current database configuration:

    python cli/run_benchmark.py
  • reset database (without generating data for benchmark again):

    python cli/reset_environment.py
  • reset index configuration to default state:

    python cli/reset_indexes.py
  • set specific index configuration (run python cli/set_index.py --help to see column order):

    python cli/set_index.py 100000000000000000000000000000000000000000000

Important notes

DBGen bug

In file BUGS in dbgen directory in official TPC-H Tools, a bug named "Problem #00062" is mentioned, which states the following:

bad update rollover after 1000 refreshes
This test uses tpcH scale 0.01. We've encountered
an situation in which dbgen doesn't generate
the correct data for delete files delete.1000 and
above. In particular, file delete.1000 contains
keys to be deleted that have never been loaded.
Because of this problem, keys that should have been
deleted never are causing duplicate unique values
to appear in the incremental loads after we cycle
from the 4000th incremental update back around starting
again with the 1st one.

The bug was closed, supposedly due to an unsupported scale factor. However, according to our observations, the bug still exists and affects every SCALE_FACTOR.

Due to this bug database is constantly growing in size after 999 refresh functions. In order to continue with experiments we had to implement a "fixed" way of generating data.

Our way of "fixing" data uses only 1998 refreshes, after which the database is in its initial state (as opposed to 4000, which would be official number if the bug hadn't existed).

Had there been any solution for that bug, ensure to update MAX_REFRESH_FILE_INDEX variable in the file db_env/tpch/config.py.

MAX_REFRESH_FILE_INDEX = 4000 # 1998 changed to 4000

and file cli/initiate_environment.py like so:

from db_env.tpch.TpchGenerator import TpchGenerator

if __name__ == '__main__':
    generator = TpchGenerator()
    generator.reset_db()
    generator.generate_data()
    generator.load_db()
    generator.generate_refresh_data()           # uncomment this line
    #generator.generate_fixed_refresh_data()    # comment or remove this line
How to confirm the bug
  1. Generate data with DBGen (note officially supported 1 GB SCALE_FACTOR):

    ./dbgen -vf -s 1
  2. Generate refresh data for 4000 refresh function pairs:

    ./dbgen -vf -U 4000 -s 1
  3. Delete files from 1000 onwards do not delete anything, which can be checked by:

    grep -F -f delete.1000 ./orders.tbl* 

This demonstrates that, after refreshes 1-999 (which are assumed to be correct), delete files don't actually remove anything. At the same time update files do add records, which causes database to grow in size, which messes up the benchmark.

Database state file

You should never interfere with file path/to/project/data/rf_db_index.txt.

Every executed refresh pair should increment number stored in that file.

After refresh pair number 4000 (or currently after refresh pair 1998) database should be in its initial state (note this has not been fully tested) - rf_db_index.txt should contain 1.

Otherwise, you will lose information about current database state, which may (and sooner or later will) lead to errors and force you to stop training and reset database.


Project structure

agent

Package with agent to train.

data

Directory to store project datafiles and agent data for analysis.

data_analysis

Directory to store scripts used for analysis of agent training data.

db_env

Package with database environment for agent.

shared_utils

Package with utilities functions and variables to use in project.

test

Run tests with code coverage (run pip install -r test/requirements.txt to install test requirements):

coverage run -m unittest
coverage report

Citing "Intelligent Index Tuning Using Reinforcement Learning"

If you use this repository in your research, please cite:

@inproceedings{10.1007/978-3-031-42941-5_45,
	title        = {Intelligent Index Tuning Using Reinforcement Learning},
	author       = {Matczak, Micha{\l} and Czocha{\'{n}}ski, Tomasz},
	year         = 2023,
	booktitle    = {New Trends in Database and Information Systems},
	publisher    = {Springer Nature Switzerland},
	address      = {Cham},
	pages        = {523--534},
	isbn         = {978-3-031-42941-5},
	editor       = {Abell{\'o}, Alberto and Vassiliadis, Panos and Romero, Oscar and Wrembel, Robert and Bugiotti, Francesca and Gamper, Johann and Vargas Solar, Genoveva and Zumpano, Ester},
	abstract     = {Index tuning in databases is a critical task that can significantly impact database performance. However, the process of manually configuring indexes is often time-consuming and can be inefficient. In this study, we investigate the process of creating indexes in a database using reinforcement learning. Our research aims to develop an agent that can learn to make optimal decisions for configuring indexes in a chosen database. The paper also discusses an evaluation method to measure database performance. The adopted performance test provides necessary documentation, database schema (on which experiments will be performed) and auxiliary tools such as data generator. This benchmark evaluates a selected database management system in terms of loading, querying and processing power of multiple query streams at once. It is a comprehensive test which results, calculated on measured queries time, will be used in the reinforcement learning algorithm. Our results demonstrate that used index technique requires repeatable benchmark with stable environment and high compute power, which cause cost and time demand. The replication package for this paper is available at GitHub: https://github.com/Chotom/rl-db-indexing.}
}