2)create the folder who ever is the creater of that database or owner of that database. I think the databse starts with username_ so i think i have to split the database name to get folders name . How to do that? backs up the databases belonging to that user.
I am attempting to write a backup script that will do the following:
1) lock and flush tables on a mysql db 2) dump the db to a file 3) unlock the tables 4) rsync the file to offsite storage
It all seems to be going well. However, obviously I don't want to setup ssh to the storage server on another network as the root user without a password. so I am attempting to su as the backup user inside of the script but when I try to run the script everything happens as it should until I try to so.. then it jumps out of the script .. akss me to login as the backup user.. proceeds to rsync to the offsite storage it does all this and then resumes execiting the script. it is not going to be setup as a cron job. it will be executed manually. assuming that is the case, how can I get the script to run without prompting for a password?
Here is what I've come up with so far... assuming that the script is run as root and the identity of the backup user will need to be assumed inside the script without perstering the user to enter the backup user's password.
I am using Joomla and a script within it. That Script is suppose to copy tables from one DB to another.
Code: Error(1) [1142] query [CREATE OR REPLACE VIEW test_users AS SELECT * FROM `localiz_master`.`jos_users`]. DB Error: CREATE VIEW command denied to user 'localiz_master'@'localhost' for table 'test_users' SQL=CREATE OR REPLACE VIEW test_users AS SELECT * FROM `localiz_master`.`jos_users` Error [1142] retrying query [CREATE OR REPLACE VIEW test_users AS SELECT * FROM `localiz_master`.`jos_users`]. DB Error: CREATE VIEW command denied to user 'localiz_master'@'localhost' for table 'test_users' SQL=CREATE OR REPLACE VIEW test_users AS SELECT * FROM `localiz_master`.`jos_users` localiz_master: name of the DB from which data is suppose to be copied. jos_users: name of the table form the main DB. localiz_master: user for the DB. localhost: host test_users: name of the table to be created.
We do have phpmyadmin but the user's table is hidden. What SSH command we might run to make localiz_master user have access to all the databases.
mysql> CREATE TABLE Shipment ( -> Shipment_ID INT(10) NOT NULL, -> Delivery_ID INT(10) NOT NULL, -> Customer_ID INT(10) NOT NULL, -> Driver_Employee_ID INT(10) NOT NULL, -> Consignee VARCHAR(255) NOT NULL, -> Payment_Method VARCHAR(255) NOT NULL, -> Payment_Received VARCHAR(1) NOT NULL, -> CONSTRAINT PRIMARY KEY (`Shipment_ID`) -> FOREIGN KEY (`Delivery_ID`) REFERENCES `Delivery`(`Delivery_ID`) -> FOREIGN KEY (`Customer_ID`) REFERENCES `Customer`(`Customer_ID`) -> FOREIGN KEY (`Driver_Employee_ID`) REFERENCES `Driver`(`Driver_Employee_ID`) -> )ENGINE=InnoDB DEFAULT CHARSET=latin1;
ERROR 1064 (42000): You have an error in your SQL syntax; check the manual that corresponds to your MySQL server version for the right syntax to use near 'FOREIGN KEY (`Delivery_ID`) REFERENCES `Delivery`(`Delivery_ID`) FOREIGN KEY (`C' at line 10)
Is there somebody knows about how to convert or changes tables on MySQL ? I was working a lot of months with MySQL with localhost. And now will setting up to MySQL database Cluster. I was success to setup the clustered with "MySQL-Cluster-gpl-7.0.16" on CentOS environment. But, when I tried to restore database from localhost to clustering system, I get this message error when do changes from MyISAM to NDBCLUSTER Tables "Error, Can't create table 'dbname.sql-3ef9_a8'(errno: 708).
I have a scheduled backup to run on our server at work and since the 7/12/09 it has be making 592k files instead of 10Mb files, In mysql-admin (the GUI tool) I have a stored connection for the user 'backup', the user has select and lock rights on the databases being backed up. I have a backup profile called 'backup_regular' and in the third tab along its scheduled to backup at 2 in the morning every week day. If I look at one of the small backup files generated I see the following:
Code:
-- MySQL Administrator dump 1.4 -- -- ------------------------------------------------------ -- Server version`
[code]....
It seems that MySQL can open and write to the file fine, it just can't dump
I installed mediawiki the other day and went with the default innodb option. However a week later something went wrong. And since I have scripts that nightly backup /var/ I just copied the backup of /var/lib/mysql/wikidb/ (as I've done with MyISAM). Then when I connect the wikidb database. I can see the tables (via "show tables"), but when I do any query with them (check table X, select * from X) I get:
Code:
Table 'wikidb.X' doesn't exist I've since read that can can't just copy the database directory like MyISAM, and there appears to be no way that I can find to restore or fix Innodb, without a dump of the data. And I never got a chance to do a mysqldump of the data. So has anybody got any idea how I can at least view the "page" table from the files I've backed up in /var/lib/mysql/wikidb/ ?
In my fresh FC12 install, I installed the MySQL server and it has a database called "mysql" with several tables. A couple of these tables, "plugin" and "server" in particular, appear to be corrupt. Anyone else experience this? Fix?
i am using liferay5.2(mysql included in the download pack) on fedora. while liferay is working fine but i cannot connect to mysql.i am getting the error
[ [root@localhost ~]# ln -s /var/lib/mysql/mysql.sock /tmp [root@localhost ~]# /etc/init.d/mysqld start Timeout error occurred trying to start MySQL Daemon.
[code]....
i didnt find the mysql.sock file in the location /var/lib/mysql/mysql.sock. i cannot find the portal-ext.properties file also to make intial settings.
I had a problem with my ubuntu installation, but I don't want to get into that...The OS wouldn't load...Now, that I have a ubuntu partition mounted on a live cd, is there a way I can access or backup my database without booting the system? I'm pretty sure the database is stored somewhere, but how can I find it and back it up?
How to backup from an old mysql in a hard drive?I have this HD which used to have an Ubuntu 7.04... However, the system crashed and it is not possible to boot anymore.I have moved the HD to another computer and made a copy for every important file and folder... But I am facing trouble to backup mysql... To copy /etc/lib/mysql doesn't work because it is probably in an older version from the mysql I use on Ubuntu 10.10
I'm trying to recover a compressed mysql backup. As the backup is extremely large, I dont wanna decompress it before importing. How can I make a mysql variable take effect before I load this compressed file into the database.
policy is to backup mysql with mysql-zrm.However at a certain stage it hangs forever. This is at the "flush logs". I tried this manually and it gave the same result. Even after restarting mysql and the host. After some googling and trying I found out "flush tables with read lock" gives the same result. The tables seem to be MyIsam. I tried with a mysqldump on one server and restore it on a test vm. I used the same config and flush logs still hangs. ALso I tried to change some configuration directives... but with the same result
Edit: btw, I checked the logfile and didn't found anything (/var/log/mysqld.log)
Edit2: I also did myisamchk -s *.MYI (in all direcoties with db files;actually did it with find command) and it did not return anything so datafiles seem ok.
I'm running: OpenSuSE 11.1 MySQL 5.0.67 MySQL Administrator 1.2.12
I have one database for my web application. It's using MyISAM tables. In MySQL admin, I went to Backup and created a scheduled backup project called backup1. On the advanced tab, I selected "Lock All Tables" and checked the box for "Backup Selected Database Completely". On the Schedule Backup tab I have "Execute this Project Daily" selected and the time 1AM. The target directory is set along with the connection name. The connection it's using is the same MySQL user that my web application uses in order to utilize the database. It has full access to that database. Now when the backup file is created, it backs up no data at all.
Instead in creates files that are 580bytes in size. When I open the files they contain: -- MySQL Administrator dump 1.4 -- -- ------------------------------------------------------ -- Server version5.0.67
/*!40101 SET @OLD_CHARACTER_SET_CLIENT=@@CHARACTER_SET_CLIENT */; /*!40101 SET @OLD_CHARACTER_SET_RESULTS=@@CHARACTER_SET_RESULTS */; /*!40101 SET @OLD_COLLATION_CONNECTION=@@COLLATION_CONNECTION */; /*!40101 SET NAMES utf8 */; /*!40014 SET @OLD_UNIQUE_CHECKS=@@UNIQUE_CHECKS, UNIQUE_CHECKS=0 */; /*!40014 SET @OLD_FOREIGN_KEY_CHECKS=@@FOREIGN_KEY_CHECKS, FOREIGN_KEY_CHECKS=0 */; /*!40101 SET @OLD_SQL_MODE=@@SQL_MODE, SQL_MODE='NO_AUTO_VALUE_ON_ZERO' */;
we've been trying to become a bit more serious about backup. It seems the better way to do MySQL backup is to use the binlog. However, that binlog is huge! We seem to produce something like 10Gb per month. I'd like to copy the backup to somewhere off the server as I don't feel like there is much to be gained by just copying it to somewhere else on the server. I recently made a full backup which after compression amounted to 2.5Gb and took me 6.5 hours to copy to my own computer ... So that solution doesn't seem practical for the binlog backup.Should we rent another server somewhere? Is it possible to find a server like that really cheap? Or is there some other solution? What are other people's MySQL backup practices?
I am currently using a script to backup my Ubuntu 10.04.1 system. The mySQL databases are backed up separately from the the system / data.
My problem is with the mySQL incremental / binary log backups.
The problem is that the binary log file(s) are always named xxxx-bin.1.
Up to about a month ago the binary logs were named xxxx-bin.000001, xxxx-bin.000002, etc.
I did make some changes at about the time that this change in file naming ocurred, but I can not identify what, if any, setting I may have changed that has caused all of the binary log files to always have the same name.
My back up script uses both mysqldump and mysqladmin flush-logs to create the binary logs.
All of the setting for mysqldump and mysqladmin are contained in the my.cnf file.
The my.cnf file contents that are relavent are as follows:
Code:
The statements in the backup script that do the backup are:
mysqladmin flush-logs
or
mysqldump | gzip > $DB_BACKUP_DIR/$ARCHIVE_FILE #Note: delete-master-logs in my.cnf
I'm using ArchLinux and I have an IP tables rule that I know works (from my other server), and it's in /etc/iptables/iptables.rules, it's the only rule set in that directory. I run, /etc/rc.d/iptables save, then /etc/rc.d/iptables/restart, but when I do "iptables --list", I get ACCEPTs on INPUT,FORWARD & OUTPUT.
# Generated by iptables-save v1.4.8 on Sat Jan 8 18:42:50 2011 *filter :INPUT DROP [0:0] :FORWARD DROP [0:0]
How to configure IP tables in opensuse?My situation is,1. My network is connected in this way...ISP -> Hub -> Firewall Router -> Switches -> DNS ->LANI need to insert a linux system with iptables configured. I am expecting it that it should be between DNS and LAN. For that I am in search of "How to configure IP tables on opensuse10.3
translationables,i.e,camp_generic_transcamp_generic_trans_defcamp_generic_trans_epcthese tables exist in cntr1/cntr1@camABC1I have to copy these tables from here to staging environment.I need to first check if any of the tables exist in staging.If yes,then delete and create new.There are 4 staging env from 5 to 8.I have to pass the staging environment as parameter to ask the user which environment does he want the tables to be copied
Everyday logwatch sends me the following information from IPTables and it seems odd, can anyone tell me are these entries odd, they all seem to be to two ports 28960 and 28964