sourcebroker / deployer-extended-database
Deployer tasks to manage database synchronization between application instances.
Installs: 222 372
Dependents: 7
Suggesters: 0
Security: 0
Stars: 38
Watchers: 8
Forks: 13
Open Issues: 3
Requires
- deployer/deployer: ^7.0.0
- sourcebroker/deployer-instance: ^6.0.0 || dev-master
- sourcebroker/deployer-loader: ^4.0.0 || dev-master
- symfony/dotenv: ^3.3 || ^4 || ^5 || ^6
- dev-master
- 17.0.0
- 16.1.0
- 16.0.1
- 16.0.0
- 15.0.0
- 14.0.0
- 13.0.2
- 13.0.1
- 13.0.0
- 12.2.1
- 12.2.0
- 12.1.0
- 12.0.0
- 11.0.2
- 11.0.1
- 11.0.0
- 10.0.1
- 10.0.0
- 9.0.0
- 8.0.0
- 7.0.2
- 7.0.1
- 7.0.0
- 6.2.1
- 6.2.0
- 6.1.2
- 6.1.1
- 6.1.0
- 6.0.0
- 5.0.4
- 5.0.3
- 5.0.2
- 5.0.1
- 5.0.0
- 4.0.5
- 4.0.4
- 4.0.3
- 4.0.2
- 4.0.1
- 4.0.0
- 3.0.1
- 3.0.0
- 2.2.2
- 2.2.1
- 2.2.0
- 2.1.0
- 2.0.8
- 2.0.7
- 2.0.6
- 2.0.5
- 2.0.4
- 2.0.3
- 2.0.2
- 2.0.1
- 2.0.0
- 1.0.1
- 1.0.0
This package is auto-updated.
Last update: 2024-11-22 10:44:43 UTC
README
- What does it do?
- Installation
- Options
- Options for "db_databases"
- Examples
- Config with one database and database data read from .env file
- Config with two databases and database data read from .env file
- Config with one database and database config read from from different sources
- Config with one database and database config read from "my framework" file
- Config of truncate_tables, ignore_tables_out, post_sql_in_markers
- Tasks
- Changelog
What does it do?
The package provides additional tasks for deployer (deployer.org) for synchronizing databases between instances. Most useful are two tasks:
- task "db:pull [source-instance]" task which allows you to pull database from remote instance to local instance,
- task "db:push [target-instance]" task which allows you to push database from local to remote instance,
- task "db:copy [source-instance] --options=target:[target-instance]" which allows to copy database between remote instances.
Installation
Install package with composer:
composer require sourcebroker/deployer-extended-database
Put following lines in your deploy.php:
require_once(__DIR__ . '/vendor/sourcebroker/deployer-loader/autoload.php'); new \SourceBroker\DeployerLoader\Load([ ['path' => 'vendor/sourcebroker/deployer-instance/deployer'], ['path' => 'vendor/sourcebroker/deployer-extended-database/deployer'], ]);
IMPORTANT NOTE! Do not putrequire('/vendor/autoload.php')
inside your deploy.php because you can have dependency problems. Userequire_once(__DIR__ . '/vendor/sourcebroker/deployer-loader/autoload.php');
instead as suggested.
Create ".env" file in your project root (where you store deploy.php file). The .env file should be out of git because you need to store here information about instance name. Additionally put there info about database you want to synchronise. You can move the info about database data to other file later but for the tests its better to put it in .env file. Remember to protect .env file from downloading with https request if the root folder is readable from WWW server level.
INSTANCE="local" DATABASE_HOST="127.0.0.1" DATABASE_NAME="database_name" DATABASE_USER="database_user" DATABASE_PASSWORD="password"
The INSTANCE must correspond to
host()
name. You need to put the .env file with proper INSTANCE name and database access data on on each of you instances.Define "local" host and set the "db_databases" for it. Use following code:
(new \SourceBroker\DeployerExtendedDatabase\Driver\EnvDriver())->getDatabaseConfig()
which will read database data from .env file.
host('local') ->set('deploy_path', getcwd()) ->set('db_databases', [ 'database_default' => [ (new \SourceBroker\DeployerExtendedDatabase\Driver\EnvDriver())->getDatabaseConfig() ] ])
Add "db_databases" var for all other hosts. For example for live host it can be:
host('live') ->setHostname('my-server.example.com') ->setRemoteUser('deploy') ->set('deploy_path', '/var/www/myapplication/') ->set('db_databases', [ 'database_default' => [ (new \SourceBroker\DeployerExtendedDatabase\Driver\EnvDriver())->getDatabaseConfig() ] ])
Make sure all instances have the same
/vendors
folder withdeployer-extended-database
and the samedeploy.php
file. Most problems are because of differences indeploy.php
file between instances.Run
dep db:pull live
to test if all works.
Options
-
db_databases
default value: null
Databases to be synchronized. You can define more than one database to be synchronized. See db_databases for options available inside db_databases. Look for Examples for better understanding of structure.
-
db_storage_path_relative
default value: .dep/database/dumps
Path relative to "deploy_path" where you want to store database dumps produced during database synchro commands.
Options for "db_databases"
"db_databases" is an array of "database configurations" and "database configuration" is array of configuration parts. Configuration part can be array or string. If its string then its treated as absolute path to file which should return array of configuration. Each or array configuration parts is merged. Look for Examples for better understanding.
-
host
default value: null
Database host.
-
user
default value: null
Database user.
-
password
default value: null
Database user password.
-
dbname
default value: null
Database name.
-
truncate_tables
default value: null
Array of tables names that will be truncated with task db:truncate. Usually it should be some caching tables that will be truncated while deployment. The value is put between ^ and $ and treated as preg_match. For example you can write "cf_.*" to truncate all tables that starts with "cf_". The final preg_match checked is /^cf_.*$/i
-
ignore_tables_out
default value: null
Array of tables names that will be ignored while pulling database from target instance with task db:pull The value is put between ^ and $ and treated as preg_match. For example you can write "cf_.*" to truncate all tables that starts with "cf_". The final preg_match checked is /^cf_.*$/i
-
post_sql_in
default value: null
SQL that will be executed after importing database on local instance.
-
post_sql_in_markers
default value: null
SQL that will be executed after importing database on local instance. The diffrence over "post_sql_in" is that you can use some predefined markers. For now only marker is {{domainsSeparatedByComma}} which consist of all domains defined in->set('public_urls', ['https://live.example.com']);
and separated by comma. Having such marker allows to change active domain in database after import to other instance as some frameworks keeps domain names in database.
Examples
Below examples should illustrate how you should build your database configuration.
Config with one database and database data read from .env file
deploy.php file:
set('db_default', [ 'ignore_tables_out' => [ 'caching_.*' ] ]); host('live') ->setHostname('my-server.example.com') ->setRemoteUser('deploy') ->set('deploy_path', '/var/www/myapplication') ->set('db_databases', [ 'database_foo' => [ get('db_default'), (new \SourceBroker\DeployerExtendedDatabase\Driver\EnvDriver())->getDatabaseConfig() ], ] ); host('local') ->set('deploy_path', getcwd()) ->set('db_databases', [ 'database_foo' => [ get('db_default'), (new \SourceBroker\DeployerExtendedDatabase\Driver\EnvDriver())->getDatabaseConfig() ], ] );
Mind that because the db_* settings for all hosts will be the same then you can make the 'db_databases' setting global and put it out of host configurations. Look for below example where we simplified the config.
deploy.php file:
set('db_databases', [ 'database_foo' => [ 'ignore_tables_out' => [ 'caching_.*' ] (new \SourceBroker\DeployerExtendedDatabase\Driver\EnvDriver())->getDatabaseConfig() ], ] ); host('live') ->setHostname('my-server.example.com') ->setRemoteUser('deploy') ->set('deploy_path', '/var/www/myapplication/'); host('local') ->set('deploy_path', getcwd());
The .env file should look then like:
INSTANCE="[instance name]" DATABASE_HOST="127.0.0.1" DATABASE_NAME="database_name" DATABASE_USER="database_user" DATABASE_PASSWORD="password"
Config with two databases and database data read from .env file
deploy.php file:
set('db_databases', [ 'database_application1' => [ 'ignore_tables_out' => [ 'caching_.*' ] (new \SourceBroker\DeployerExtendedDatabase\Driver\EnvDriver())->getDatabaseConfig('APP1_') ], 'database_application2' => [ 'ignore_tables_out' => [ 'cf_.*' ] (new \SourceBroker\DeployerExtendedDatabase\Driver\EnvDriver())->getDatabaseConfig('APP2_') ], ] ); host('live') ->setHostname('my-server.example.com') ->setRemoteUser('deploy') ->set('deploy_path', '/var/www/myapplication/'); host('local') ->set('deploy_path', getcwd());
The .env file should look then like:
INSTANCE="[instance name]" APP1_DATABASE_HOST="127.0.0.1" APP1_DATABASE_NAME="database_name" APP1_DATABASE_USER="database_user" APP1_DATABASE_PASSWORD="password" APP2_DATABASE_HOST="127.0.0.1" APP2_DATABASE_NAME="database_name" APP2_DATABASE_USER="database_user" APP2_DATABASE_PASSWORD="password"
Config with one database and database config read from from different sources
In example we will use:
array,
'ignore_tables_out' => [ 'caching_*' ]
get() which returns array with database options,
get('db_default')
direct file include which returns array with database options
__DIR__ . '/.dep/database/config/additional_db_config.php
class/method which returns array with database options
(new \YourVendor\YourPackage\Driver\MyDriver())->getDatabaseConfig()
closure which returns array with database options
function() { return (new \YourVendor\YourPackage\Driver\MyDriver())->getDatabaseConfig()
}
Each of this arrays are merged to build final configuration for database synchro.
deploy.php file:
set('db_default', [ 'post_sql_in' => 'UPDATE sys_domains SET hidden=1;' ]); set('db_databases', [ 'database_foo' => [ [ 'ignore_tables_out' => [ 'caching_.*' ] ], get('db_default'), __DIR__ . '/databases/config/additional_db_config.php', (new \YourVendor\YourPackage\Driver\MyDriver())->getDatabaseConfig(), function() { return (new \YourVendor\YourPackage\Driver\MyDriver())->getDatabaseConfig(); } ], ] ); host('live') ->setHostname('my-server.example.com') ->setRemoteUser('deploy') ->set('deploy_path', '/var/www/myapplication/'); host('local') ->set('deploy_path', getcwd());
Config with one database and database config read from "my framework" file
Its advisable that you create you own special method that will return you framework database data. In below example
its call to \YourVendor\YourPackage\Driver\MyDriver()
. This way you do not need to repeat the data of database
in .env file. In that case .env file should hold only INSTANCE.
set('db_databases', [ 'database_default' => [ (new \YourVendor\YourPackage\Driver\MyDriver())->getDatabaseConfig() ], ] );
Config of truncate_tables, ignore_tables_out, post_sql_in_markers
Real life example for CMS TYPO3:
set('db_default', [ 'truncate_tables' => [ 'cf_.*' ], 'ignore_tables_out' => [ 'cf_.*', 'cache_.*', 'be_sessions', 'fe_sessions', 'sys_file_processedfile', 'tx_devlog', ], ]);
Tasks
db:backup
Backup database. In background, on target instance, two tasks are executed 'db:export' and 'db:compress'. Results are stored in "{{deploy_path}}/.dep/databases/dumps/".
If releases folder will be detected then it adds info about release in dumpcode name like in this example:
2017-12-04_00:20:22#server=live#dbcode=database_default#dumpcode=backup_for_release_160_ec77cb6bc0e941b0ac92e2109ad7b04e#type=structure.sql.gz
Example
dep db:backup local dep db:backup live dep db:backup live --options=dumpcode:mycode
db:compress
Compress dumps with given dumpcode stored in folder "{{deploy_path}}/.dep/databases/dumps/" on target instance.
There is required option --options=dumpcode:[value]
to be passed.
Look for config vars 'db_compress_suffix', 'db_compress_command', 'db_uncompress_command' for possible ways to overwrite standard gzip compression with your own.
Example
dep db:compress live --options=dumpcode:0772a8d396911951022db5ea385535f6
db:copy
This command allows you to copy database between instances.
dep db:copy [source-instance] --options=target:[target-instance]
In the background it runs several other tasks to accomplish this. Lets assume we want to copy database from live to dev instance. We will run following command on you local instance:
dep db:copy live --options=target:dev
Here are the tasks that will be run in background:
- In below description:
- source instance = live
- target instance = dev
- local instance = local
- First it runs
dep db:export live --options=dumpcode:123456
task on source instance. The dumps from export task are stored in folder "{{deploy_path}}/.dep/databases/dumps/" on target instance. - Then it runs
db:download live --options=dumpcode:123456
on local instance to download dump files from live instance from folder "{{deploy_path}}/.dep/databases/dumps/" to local instance to folder "{{deploy_path}}/.dep/databases/dumps/". - Then it runs
db:process local --options=dumpcode:123456
on local instance to make some operations directly on SQL dumps files. - Then it runs
db:upload dev --options=dumpcode:123456
on local instance. This task takes dump files with code:123456 and send it to dev instance and store it in folder "{{deploy_path}}/.dep/databases/dumps/". - Finally it runs
db:import dev --options=dumpcode:123456
on target instance. This task reads dumps with code:123456 from folder "{{deploy_path}}/.dep/databases/dumps/" on dev instance and import it to database. - At the very end it removes dumps it just imported in step 5 with command
db:rmdump dev --options=dumpcode:123456
Copy to instance defined in instance_live_name
(default live
) is special case.
If you copy to highest instance then by default you will be asked twice if you really want to.
You can disable asking by setting db_allow_copy_live_force
to true
.
You can also forbid copy to live instance by setting db_allow_copy_live
to false
.
db:decompress
Decompress dumps with given dumpcode stored in folder "{{deploy_path}}/.dep/databases/dumps/" on target instance.
There is required option --options=dumpcode:[value]
to be passed.
Look for config vars 'db_compress_suffix', 'db_compress_command', 'db_uncompress_command' for possible ways to overwrite standard gzip compression with your own.
Example
dep db:decompress live --options=dumpcode:0772a8d396911951022db5ea385535f6
db:download
Download database dumps with selected dumpcode from folder "{{deploy_path}}/.dep/databases/dumps/" on target instance
and store it in folder "{{deploy_path}}/.dep/databases/dumps/" on local instance.
There is required option --options=dumpcode:[value]
to be passed.
Example
dep db:download live --options=dumpcode:0772a8d396911951022db5ea385535f6
db:dumpclean
Clean database dump storage on target instance. By default it removes all dumps except last five but you can set your values and also change the values depending on instance.
Example
set('db_dumpclean_keep', 10); // keep last 10 dumps for all instances set('db_dumpclean_keep', [ 'live' => 10 // keep last 10 dumps for live instance dumps 'dev' => 5 // keep last 5 dumps for dev instance dumps '*' => 2 // keep last 5 dumps for all other instances dumps ]); dep db:dumpclean live
db:export
Dump database to folder on local instance located by default in "{{deploy_path}}/.dep/databases/dumps/".
Dumps will be stored in two separate files. One with tables structure. The second with data only.
There is option --options=dumpcode:[value]
that can be passed. If there is no dumpcode then its created and returned as
json structure.
Example
Example task call:
dep db:export live
Example output files located in folder {{deploy_path}}/.dep/databases/dumps/:
2017-02-26_14:56:08#server=live#dbcode=database_default#type=data#dumpcode=362d7ca0ff065f489c9b79d0a73720f5.sql 2017-02-26_14:56:08#server=live#dbcode=database_default#type=structure#dumpcode=362d7ca0ff065f489c9b79d0a73720f5.sql
Example task call with own dumpcode=
dep db:export live --options=dumpcode:mycode
Example output files:
2017-02-26_14:56:08#server=live#dbcode=database_default#type=data#dumpcode=mycode.sql 2017-02-26_14:56:08#server=live#dbcode=database_default#type=structure#dumpcode=mycode.sql
db:import
Import database dump files from local instance folder "{{deploy_path}}/.dep/databases/dumps/" to local database(s).
There is required option --options=dumpcode:[value]
to be passed.
Example
dep db:import dev --options=dumpcode:0772a8d396911951022db5ea385535f66
db:process
This command will run some defined commands on pure sql file as its sometimes needed to remove or replace some strings
directly on sql file before importing. There is required option --options=dumpcode:[value]
to be passed.
Example
dep db:process dev --options=dumpcode:0772a8d396911951022db5ea385535f66
db:pull
This command allows you to pull database from target instance to local instance. In the background it runs several other tasks to accomplish this.
Here is the list of tasks that will be done when you execute "db:pull":
- First it runs db:export task on target instance and get the "dumpcode" as return to use it in next commands.
- Then it runs db:download on local instance (with "dumpcode" value from first task).
- Then it runs db:process on local instance (with "dumpcode" value from first task).
- Then it runs db:import on local instance (with "dumpcode" value from first task).
Pull to instance defined in instance_live_name
(default live
) is special case.
If you pull to highest instance then by default you will be asked twice if you really want to.
You can disable asking by setting db_allow_pull_live_force
to true
.
You can also forbid pull to live instance by setting db_allow_pull_live
to false
.
Example
dep db:pull live
db:push
This command allows you to push database from local instance to remote instance. In the background it runs several other tasks to accomplish this.
Here is the list of tasks that will be done when you execute "db:push":
- First it runs db:export task on local instance and get the "dumpcode" as return to use it in next commands.
- Then it runs db:upload on local instance with remote as argument (with "dumpcode" value from first task).
- Then it runs db:process on remote instance (with "dumpcode" value from first task).
- Then it runs db:import on remote instance (with "dumpcode" value from first task).
Push to instance defined in instance_live_name
(default live
) is special case.
If you push to highest instance then by default you will be asked twice if you really want to.
You can disable asking by setting db_allow_push_live_force
to true
.
You can also forbid push to live instance by setting db_allow_push_live
to false
.
Example
dep db:push live
db:rmdump
This command will remove all dumps with given dumpcode (compressed and uncompressed).
There is required option --options=dumpcode:[value]
to be passed.
Example
dep db:rmdump live --options=dumpcode:0772a8d396911951022db5ea385535f66
db:truncate
This command allows you to truncate database tables defined in database config var "truncate_tables". No dumpcode is needed because it operates directly on database.
Example Truncate local instance databases tables.
dep db:truncate
Truncate live instance databases tables.
dep db:truncate live
db:upload
Upload database dumps with selected dumpcode from folder "{{deploy_path}}/.dep/databases/dumps/" on local instance and
store it in folder "{{deploy_path}}/.dep/databases/dumps/" on target instance.
There is required option --options=dumpcode:[value]
to be passed.
Example
dep db:upload live --options=dumpcode:0772a8d396911951022db5ea385535f6
Changelog
See https://github.com/sourcebroker/deployer-extended-database/blob/master/CHANGELOG.rst