READ CAREFULLY: Only attempt to migrate the databases between supervisors of the same version

    If you're doing this to install a new supervisor and want to run the latest version of Qube! on the new supervisor, match versions between the supervisors first, and then upgrade the new supervisor once the databases have been migrated over to it.

    If you're running Qube 6.0 or later, you will need to contact Sales at PipelineFX in order to get your Qube licenses moved over to the new machine. They'll need a MAC address for the new supervisor host.

    To migrate a Qube supervisor, you need to migrate both the MySQL databases and the job logs if the job logs are stored on the supervisor's local disk. If your job logs are stored on the network you will not have to move them, simply set the supervisor_logpath on the new supervisor to point to the same network directory as the old supervisor.

    It is recommended that you start moving the job logs before you start migrating the MySQL databases, as the job logs will take longer to move.

    Don't forget to duplicate any settings you've changed in the old supervisor's qb.conf.

    To migrate the MySQL databases: 

    The following instructions use the unix 'ls' command. You may want to install cygwin on the old supervisor if it's a Windows machine if you are following along, otherwise, substitute the equivalent Windows command line commands.

    Build the new supervisor up, install MySQL if it's a Linux host, then install the same version ofQube that you are running on the old supervisor, then stop the supervisor daemon on the new supervisor. Now the new supervisor is ready to have the databases copied over. The trick is to build a mysqldump command that will only dump the *qube databases, then send them over the network to the new server.

    Disable the Data Warehouse collectors on both the old and new supervisors

    This needs to be done at least 15 minutes before starting, as some of the tasks currently running on the old supervisor can take some time to finish on warehouses that contain several years' data.


    You'll need to temporarily disable all the scheduled tasks that start with the name Qube DataWarehouse Data Collector on both the old supervisor (so that they're not running while you dump the databases), and on the new supervisor (so that they're not running during the restore).  You can leave them disabled on the old supervisor, since you're shutting it down anyways, but remember to re-enable them on the new supervisor after everything's back up and running.



    Comment out all the tasks in the cron file /etc/cron.d/com.pipelinefx.DataWarehouse.cron on both supervisors.  

    OS X:

    Unload each of the launchd plists that run the data warehouse collectors

    sudo launchctl unload /Library/LaunchDaemons/com.pipelinefx.DataWarehouse.12h.plist
    sudo launchctl unload /Library/LaunchDaemons/com.pipelinefx.DataWarehouse.15m.plist
    sudo launchctl unload /Library/LaunchDaemons/com.pipelinefx.DataWarehouse.1d.plist
    sudo launchctl unload /Library/LaunchDaemons/com.pipelinefx.DataWarehouse.1h.plist
    sudo launchctl unload /Library/LaunchDaemons/com.pipelinefx.DataWarehouse.5m.plist

    Grant access to the root user from the old server.

    On the new server, at the mysql> prompt:


    GRANT ALL PRIVILEGES ON *.* TO 'root'@'';

    (replace the ip address in the command with the ip address of your old server)

    Test the access from the old supervisor:


    mysql -u root -h <newHostName> -B -e "SELECT USER()"


    It should return something like:


    USER() root@

    Now that you can access the MySQL server on the new supervisor from the MySQL server on the old supervisor, it's time to push the dump across from the old supervisor to the new one.

    On the old supervisor: 

    You need to build a list of all the *qube databases. The simplest way is to 'cd' into the mysql data directory and list them with "ls"


    cd /usr/local/mysql/data
    \ls -d *qube pfx_dw pfx_stats
    1qube  2qube  3qube  4qube  5qube  6qube  7qube  8qube pfx_dw pfx_stats qube

    Note the use of the leading backslash on the 'ls' command. It's to skip any aliases or formatting operators that you usually pass to 'ls', to avoid having slahes on the end of the directory names. The output from the 'ls' command should look like the above.

    Then the mysqldump command will look like:

    mysqldump --opt --databases `\ls -d *qube pfx_dw pfx_stats` | mysql -u root -h <linux ip address>

    Those are backticks around the "\ls -d *qube" command. The backtick is the left-leaning single-quote below the ~ character on the keyboard.


    1. Windows users must specify a port while running this command. Windows mysqlclient uses port 3300 by default whereas Linux and OS X assumes 3306. To specify a port, add "-P 3306" to the end of the above command
    2. If the old supervisor is running on Windows and you have not installed cygwin (or some other *nix shell utility), you cannot pipe commands to other commands, nor can you do command substitution with backticks. In this case you'll need to hand-enter the list of qube databases, send the output of mysqldump to a file on the old supervisor, then read that file on the new supervisor. The commands would look something like:

    	old supervisor: mysqldump --opt --databases 1qube 2qube 3qube ... qube > some_file.sql
    new supervisor: mysql -u root < some_file.sql

    The list of qube databases (1qube 2qube 3qube, etc) can be found by looking in c:\Program Files\pfx\qube\mysql\data for all files that contain the word "qube".


    If you get a error message:

    [ERROR] mysqld: Can't open file: './<database>/<table>.frm' (errno: 24)

    You need to increase your open_file_limits in my.cnf. On the old supervisor, add:

    open_files_limit = 100000

    to the [mysqld] section of my.cnf


    The mysqldump should take quite a while to finish, and will print no output. A way to check if it's doing the right thing while it's running is to login to the mysql server on linux machine, and run:



    You should start to see some qube databases with numbers in front of them, like 12qube. Run the command every minute or so, and the list of databases should be growing. Once the mysqldump command has finished on the old supervisor, you'll need to do a bit of maintenance on the new tables on the new supervisor.

    On the new supervisor

    mysqladmin -u root flush-privileges
    mysqlcheck -u root --repair --all-databases

    Once that's done, you should start the supervisor service.

    Verify that the jobs are present:

    qbjobs -u all

    Verify that you can see the logs for a random job:

    qbout <someJobId>
    • No labels