General :: Huge Directory Into Tar Gz Or Bz2?
Dec 26, 2010
suggest or advise the best practice of bz2 or tar gz. i have a directory /var/opt/axigen which has size 33gb on daily basis as per the schedule we need to take back.i want to know pros and crons of below commands, say best compression and decompresstar cvzf /var/opt/bkup_axigen/axigen_bkup_1.tar.gz /var/opt/axigenortar -jvzf9 /var/opt/bkup_axigen/axigen_bkup_1.tar.gz /var/opt/axigen
View 2 Replies
ADVERTISEMENT
Jan 2, 2011
EDIT, 3 January 2010: SOLVED: In fact, it was the fonts that were being rendered larger and pushing everything out. Altering DPI in xorg.conf solved the problem, see details at bottom.ORIGINAL POST:I used jockey-kde to activate the nVidia (closed-source) driver in order to fix some full-screen problems I was having.It has fixed that, but now everything is really big. I have my resolution set to 1680x1050 but it appears to be lower than it was before - the K menu, for instance, takes up about 1/6 of the screen when I open it. I took a screenshot here:http://i.imgur.com/94lyN.jpgI know this isn't much information, but can anyone tell me why - while the desktop appears to be larger with a higher resolution - applications are actually appearing as though the resolution is lower?
View 7 Replies
View Related
Oct 31, 2010
I got a apple PowerBook 3400c with upgraded ram. I would like to boot linux on it but the ram is still not huge (96MB). I burned PowerPup to a CD but it would not boot and kept booting macOs 9.
View 2 Replies
View Related
Mar 4, 2011
Since a couple of months, the machine is hanging itself. I can't connect in any way to it. I have it in a Datacenter far away my office, so when it happens I have to ask the provider to reboot it manually because it always happen in working hours and my clients gets mad on that moments. It happens every 15 days (more or less). It is very difficult to get the procesess status, but today hangs again and I could connect (a very very slow connection) and save a "ps aux".
[Code]...
There are some scripts that you can see as "php /var/www/html/call_engine.php ..." that are consuming CPU more than I expect and I will correct, but I would like to know about the other processes like many crond or the "/usr/sbin/sendmail -FCronDaemon -i -odi -oem -oi -t". Or maybe the whole problem would be the php script?
View 2 Replies
View Related
Jul 8, 2011
I want to extract a huge .tar.gz file but when I do extract it stalls the server. The server is write heavy and extracting seems to choke the disk. Is there a nice way to extract without stopping the world? I've tried the 'nice' and 'cpulimit' command but they don't seem to do the trick.
View 2 Replies
View Related
Jul 13, 2011
I am running the command on a Mac but due to it being a generic unix command and a command line query.. I thought I can write on this forum.. I am running the command
Code:
df -h | grep '/dev/'
I get
Code:
/dev/disk0s2 389Gi 62Gi 327Gi 16% /
/dev/disk0s3 76Gi 24Gi 52Gi 32% /Volumes/Backup
/dev/disk3s2 500Gi 47Gi 453Gi 10% /Volumes/Misc
Note the huge space between the 1st and 2nd Column..
This is because currently I have some NAS drives mounted which are not showing due to grep. When they are not mounted. The output is fine with equal spaces between each column (like between col 2 and 3.. or 3 and 4). I want to do a (dare I say) sed or awk or something to reduce the space between 1st and 2nd col. So that it has space like between col 3 and 4.. or 2 and 3. This is because I am showing this output somewhere and because of the space its not showing up correctly.Also I hope the command will still work when the NAS drives (afp) are not mounted.. basically consistency. The spaces are not showing properly in the quote tag. Changed it to CODE tag.
View 12 Replies
View Related
Apr 23, 2011
I have a huge log file of around 3.5 GB and would like to sample random sections in the middle of say 10 MB for the purpose of debugging what my application is doing.
I could use head or tail commands to get the beginning or end of the file, how can I grab an arbitrary portion from the middle of the file? I guess I could do something like head -n 1.75GB | tail -n 10MB but that seems clumsy and I'd need to determine line numbers for the midpoint of the file to get 1.75GB and 10MB line counts.
View 2 Replies
View Related
May 19, 2010
Is there any software in Linux to view huge .txt files, say, over 10 megas? I'm now using default "gedit", version 2.28.0, which seems to not be able to open huge .txt files. It's the same case for Windows default .txt browser, but in Windows, "Win Word" seems to work fine. software under Linux to browse huge .txt files?
View 5 Replies
View Related
May 30, 2011
I've a file with a size of 6GB. I would like to compress this file and split them into smaller files. I was also thinking in use bzip2 to compress it, because if offers a good compression rate. How can I split this file into small ones to compress it?
View 5 Replies
View Related
Jan 9, 2011
Please why my scanning is always creating huge 50Mb to 100Mb PDF files ?Each A4 Pnm file is of 6.5Mbytes by resolution of 150.If I decrease the resolution lower than 100, then it starts to be unreadable my text ...
View 11 Replies
View Related
Jan 24, 2011
I use a program which makes a large image which I have to scroll to view. The program has no way to save the image, and I have no access to the source to modify it. The only way I have to get the image from the program is by screenshot. My goal is to save the full size image without having to piece together individual screenshots. I'm using this script to try taking a screenshot:
#!/bin/bash
window=$(wmctrl -l | grep "Program$" | awk '{print $1}')
wmctrl -v -i -r $window -e '0,0,0,6030,5828'
wmctrl -i -a $window
import -window $window ~/Desktop/screenshot.png
This uses wmctrl to get the window id ($window) for a window named "Program". It then tries to resize the window to the desired dimensions. It uses imagemagick (import) to save a screenshot.png on the user's Desktop. All of this works except the resize step. I can resize the window using wmctrl -r -e, but sizes greater than the screen size don't work. I'm using Ubuntu 10.04 and the Gnome Desktop. I run two monitors, but I've tried this with one of them disabled. Is there a way to resize the window larger than my screen to get a huge screenshot?
Part II: I tried using xrandr to set up screen panning, so as to have a bigger desktop than my monitor. xrandr --output LVDS --panning 2600x2500 This command makes the laptop screen pan over a 2600x2500 size desktop, even though it can only show 1440x900 at one time. To turn off the panning, I can use a similar command to set total size and with zeroes for the panning section. This gives me back my original laptop display behavior. xrandr --fb 1440x900 --output LVDS --panning 0x0 This is all done with xrandr, and does not require any Xorg.conf changes (my Ubuntu system doesn't even have an Xorg.conf).
My video card seems to only allow about 6.5 million pixels, even though the maximum dimensions are 8192x8192. That maximum seems to be the maximum for either dimension, but there is a limit to how many pixels can be drawn, which is the width multiplied by the height. Once I did the screen resize, I tried my script again and got a screenshot. The screenshot however is totally scrambled. I'm not sure if it's unable to take a screenshot of an off-screen window or if it is unable to handle the large dimensions of the window. With the panning display, the window should think it is visible, and the window manager should think it is on-screen. So there is a pixel buffer somewhere with those pixels in it, so there should be a way to get a screenshot.
View 1 Replies
View Related
Dec 15, 2009
I am a both rhel5 and fedora user.I can not configure my Samsung syncmaster 632nw monitor to display full screen at 1360X768.There is huge black space both left and right side of the monitor. I have tried many times to solve it but unable to do it.The max screen resolution is 1024*768 and minimum is 640*480.
Here is the xorg.conf file:
View 2 Replies
View Related
Jun 16, 2010
I am a student studying computer science course.
Well, I am facing problem when doing lab questions.
I must use DLXLinux bundled in Bochs (bochs.sourceforge.net).
I am required to use the /usr/local directory.
In /usr directory, there is no directory named 'local' but there is one thing called 'local@'. So, when I try to use mkdir command to create 'local' directory in /usr , there are error "cannot make directory.....".
Look at my screenshot at [url].
View 9 Replies
View Related
Jan 18, 2010
How do you perform a long directory listing of all files in the /bin directory that have exactly three characters in their name?
View 1 Replies
View Related
Jan 29, 2010
Is there a way, on Linux, to cause all new files created in a directory to be owned by the directory's group instead of the creating user's group?
View 2 Replies
View Related
Oct 13, 2010
I want to run a cronjob every 15 minutes that checks a directory for files. If the directory contains more than ten files I want it to send an email to me.
All I have is this...
*/15 * * * * ls -l | wc -l | [filename] | mail -s "This is just a test" [email address]
I would rather not write a bash script. Is there an easier way to do this? I was looking into some commands like find and grep.
View 6 Replies
View Related
Jul 13, 2011
I'm quite new to linux but I have configured a simple ftp server and it's working great. I have a FTP-Shared folder with upload and download subfolders. Under upload's and download's I have identical category subfolders like mp3's, movies, software etc. in both. As the guy's upload, I would like to create a line crontab where I can move all the content under /FTP-Shared/upload/mp3/* older than 14 day's to FTP-Shared/downloads/mp3/ recursively (Like in cp command), but the timestamp must be searched on the first directory and not sub files example: /mp3/Club Dance/CD1/Hallo world.mp3This is how far I got:[root@clients ~]# /usr/bin/find /FTP_Shared/upload/Mp3s/ -depth -mindepth 1 -mtime +14 -type d -exec mv -f {} /FTP_Shared/download/Mp3s/ ;This command moves the directory and files, but it is not recursively
View 4 Replies
View Related
Apr 12, 2010
I've just installed Ubuntu 10.04, trying out linux for the first time. I'm wondering if it's possible to link my documents folder in Ubuntu to my documents folder in Windows, so that when I open the folder in Ubuntu, I am linked to the folder in Windows.
View 1 Replies
View Related
Aug 23, 2011
I'd like to mount the directory /var/www/mysite to the directory /home/daniel/mysite, but also have the user of the mounted files mapped from the original user (www-data) to my own user (daniel). So that the file /var/www/mysite/index.php who's user is www-data will appear in the mounted directory as /home/daniel/mysite/index.php and be owned by daniel - and alternatively, if I create a file /home/daniel/mysite/test.php with my own user, it will be created in the original directory under the user www-data Is it possible? If not, what alternatives do I have so I can use an IDE and still make sure all the files belong to the HTTP server's user?
View 2 Replies
View Related
May 26, 2011
I need to create a directory named just like his parent.
Example:
I tried this but because there is "Space Characters" in the name of Parent directory, my script fails.
View 8 Replies
View Related
May 13, 2011
I am in my current directory. I want to copy a directory somewhere else into this current directory. Lets say I want to take it from direc1/direc2 and the directory I want to take is called demo.
Code:
That is what it shows in the man pages, but when I do that, it says cp: no match
View 5 Replies
View Related
Feb 11, 2011
If I have a directory /foo with a few files in it, how do I symlink each entry in /foo into /bar/? For instance, if /foo has the files a, b and c, I want to create three symlinks:
/bar/a -> /foo/a
/bar/b -> /foo/b
/bar/c -> /foo/c
View 3 Replies
View Related
Apr 7, 2010
How can i auto mount more than directory in the same directory ? i want to automount 2 home directories in the /home and still be able to enter the other home directories !
The problem that i've another account on the system with home directory joe when the user1 home directory auto mounted i become unable to enter joe home directory !
View 1 Replies
View Related
May 15, 2011
Just a simple question
Code:
$ cat -
do in the Bash? I know 'cd -' changes directory to the previous working directory, but 'cat - ' does not seem to do the same thing.
View 3 Replies
View Related
Apr 13, 2010
I'm using Mac OS X's Terminal.app shell to compile and run Fortran programs. One such program resides outside of my home directory (it is in the Applications folder, which resides on my hard drive but seems to be outside of my home folder). How can I navigate into this directory using Terminal.app to run the programs that reside there?
View 7 Replies
View Related
Mar 10, 2010
i am trying to transfer a file from my live linux machine to remote linux machine it is a mail server and single .tar.gz file include all data. but during transfer it stop working. how can i work and trouble shooot the matter. is there any better way then this to transfer huge 14 gb file over network,vpn,wan transfer. the speed is 1mbps,rest of the file it copy it.
rsync -avz --stats bkup_1.tar.gz root@10.1.1.22:/var/opt/bkup
[root@sa1 logs_os_backup]# less remote.log
Wed Mar 10 09:12:01 AST 2010
building file list ... done
bkup_1.tar.gz
deflate on token returned 0 (87164 bytes left)
rsync error: error in rsync protocol data stream (code 12) at token.c(274)
building file list ... done
code....
View 1 Replies
View Related
Jul 7, 2011
So I noticed today that my machines root hard drive had almost no space on it.
I did a disk usage analyzer and I found out my var/log folder is 95GB.
The large logs are:
can I just delete them? also how can I stop this from happening again?
View 1 Replies
View Related
May 24, 2010
Using slackware current and I'm really digging KDE 4.4.3. It's been way more stable on my machine than 4.3.x and it performs MUCH better. I think slackware 13.1 is going to be a really good release. Much better than 13.0. Looking forward to upgrading all of my hosts though it will probably take me a few months given how many I manage.
View 3 Replies
View Related
May 13, 2010
I have a directory /sanvol/ifr/ which contains 90,000 files and 2 subdirectories containing 60,000 files. I need to copy the contents of /sanvol/ifr to /sanvol/edm/. The cp command will not work because there are too many files.
I tried the following: find /sanvol/ifr/ type f -name '*' -exec cp -p {} /sanvol/edm/. ;
but this copies all files including whats in the subdirectories to /sanvol/edm/
I need to replicate exactly the contents of /sanvol/ifr/ to /sanvol/edm/ and preserve file attributes and dates.
View 3 Replies
View Related
Jan 29, 2016
I installed Debian on my laptop and now I had warning that my log files inside var are getting out of hands...Following files...
36G daemon.log
48G daemon.log.1
41G kern.log
55G kern.log.1
31G messages
42G messages.1
8.2G syslog
17G syslog.1
How can I clear this and set up properly so they don't take so much space?
View 6 Replies
View Related