EXT4 File System @ Server Side & Ext3 External Storage

Hi guys

We are urgently moving Oracle application to a backup server for some patching etc. The backup server is installed with Oracle 6.6 64Bit linux with default partitions. We found that the system has ext4 file system, while our external storage through HBA interface/IBM is formatted with ext3 file system.
Hence, please let us know what kind of performances issues we may come across, or the same could be ignored?

regards


Similar Content



Linux File Server

Hello Everyone
I have installed Ubuntu Server on a 500 GB drive formatted as EXT4.

Of of my Media, movies, music and pictures are from Windows PC's formatted in NTFS.
As all of the computers connecting to the server will be windows based will NTFS be fine or will I have to reformat them to work. The two media drives are 2 x 2GB Sata drives.
Could you tell me if this setup would work or will I have to reformat and and transfer the files to a new filesystem.
Also as a file server whats the best linux file system to use so it will work with windows PC's in a lan flawlessly.
Thanks for you expertise.
Regards

I Set Up An Ext4 Partition On External HD To Store Media Files. Permission Issues

I recently bought a WD external hard drive for storing file of several types. Using gparted I made two partitions, one ntfs for windows files and an ext 4 for linux files. Strangely, I have complete access to ntfs partition from linux side of duel boot system, but do not have permission to access ext4 partition. My root password does not work when I use su to gain root access. It works fine on built in hard drive.

Parallel File System Check

Reboot of Linux server consuming more time when there is a file system check (fsck)
Most of our server are using ext3 file system
Is there any possibility to make parallel fsck during reboot ?

.tar.gz 111GB Extracting Fails

Hi guys

We are trying to move Oracle applications database tier archive, that is 111GB (over Linux) to a USB external drive using cp. Though the file successfully gets transferred to the external drive, trying to extract the file from a 2nd machine always fails, saying the archive is corrupt.
We have checked the integrity of the archive using 7-zip, reporting no errors. However totally frustrated as our last few attempts were totally futile.

The interesting part is, if we do scp to transfer the file to 2nd machine, extraction doesn't fail.

Please let us know, how we can successfully move this archive to the 2nd machine which is at a remote location and no possibilities of setting up a FTP for such a huge size file.

Both the source and destination Linux distros are RHEL 5 Enterprise, 64Bit, ext3 file systems.


regards,

Re-assembled Software Raid5 Now With No File System

Hi All

Please could someone assist me with a software raid 5 array issue i am having?

About a week or so ago i was asked to have a look at a mate of mine's Company Server that had crashed after a power failure - no ups and close to 9GB data with no backup.
HP Proliant mini server running OpenMediaVault v1.9. with a software raid5 setup consisting of 4x 3TB Western Digital Hard drives


The 80GB boot drive crashed (hardware malfunction) so i replaced it and installed the server with OpenMediaVault 1.9 as was previously installed. When starting up however i noticed that none of the shares were available and upon closer inspection discovered that the raid to had crashed. I therefore re-assembled the raid but now have no file system nor any partitions on any of the 4x drives.

This is about the extent of my knowledge base when it comes to Linux, and again i don't know that i'd want to risk it with no backup of the data.

I have run fsck /dev/md1, but this reports an error 2 while executing fsck.linux_raid_member not found.
Please i am desperate for some advise!

Thanks
Wayne

An Error Occurred During The File System Check

Hi all!! Im new in this forum and also new at linux !!

I am installing an application that includes automatically the instalation of CentOS 6.3.

The server that hosts this application and OS had a power outage and then when i turn on the server again i get the error that you can see in this pictu http://postimg.org/image/7opm0xzel/

I have 2 disks in the system, 1 for the application and SO and other to store call recordings (thats what the application do, record calls!).

If I re-install the system it has no problem!! But i dont want to re-install the system every time i have a power outage because i lost too much and important information.

I have installed this application in other server and had no problem with power outage. In this server the diference is that it originally had raid and i "remove it" frome the BIOS. Maybe here is the problem??

Anyone can help me?? Remember im really new in linux !!

HOW-TO: FAT32/NTFS/ext3 For Windows/Linux Cross-compatible

I own a NAS D-Link DNS-320 running fun_plug 0.7 & transmissionBT.

I just bought a Transcend 2.5 inch 2TB USB HDD which is preformatted to NTFS.

Just wondering whether my NAS(linux) can write to my USB HDD for a reliable storage for transmissionBT.

Otherwise, if writing to NTFS is unstable in linux, how should I format my USB HDD. I know I could format as FAT32... but FAT32 has a 4GB file size limit.
Is it possible to format USB HDD as ext3 (linux file system)... and still compatible/read/write by Windows XP/7/8?

Password Protect Ext3 Partition In A USB External Drive

Hi folks,

as my title says, i need to protect an ext3 partition in my external 500GB toshiba usb hard diks. it is a 250gb Partition. so if i try to open the partition on any linux or windows system (using a utility), it should ask for a password to open the partition. pls advise.

How To Remount Root Partition On System Booted From External Usb Drive

Hello,
I have a setup with arch linux installed on external ssd as normal (non-live) installation.
I had to do that after my main laptop was broken. I don't want to put my drive into the new hardware, so I run it as external one for until the main will be repaired.

It boots as usual, everything is working fine until the usb drive reconnects for some reason (may be the physical connection problem). It probably mounts itself as read-only, the system freaks out and dies slowly from I/O errors process by process. I cannot issue commands from shell, even reboot. I have to physically reset the machine.

Is there a way to change its behaviour to auto remount as rw after reconnecting the USB and just keep running as usual?

I think the fstab is the file that i should modify, but cannot find a way how to do it properly. Please correct me if I'm wrong. My current fstab:
Code:
# 
# /etc/fstab: static file system information
#
# <file system>	<dir> <type> <options> <dump> <pass>
# /dev/sdb2
UUID=145b3c51-6531-48f3-a79b-fb1ae238c7ee / ext4 rw,relatime,data=ordered 0 1

# /dev/sdb1
UUID=AFC4-6899 /boot vfat rw,relatime,fmask=0022,dmask=0022,codepage=437,iocharset=iso8859-1,shortname=mixed,errors=remount-ro 0 2

/swapfile none swap defaults 0 0

Backups And External Drives

Hello everyone,

I recently had an issue where I lost my whole backup server due to an electrical overload causing my server to literally explode and fried all 4 of my terabyte drives.... needless to say, I have no more backups because of this, and everywhere I read about backups said that setting up a raid array would allow me to keep good backups.... boy did I learn this lesson the hard way in needing to have some sort of external backup option, which brings me to this post and my questions:

I'm using Ubuntu 14.04 LTS server on an older Dell Poweredge 600sc, and I was thinking of using WD Passport 1Tb external drives to be used as my "offsite" backup option. I don't have a lot of data, and my current backup schedule is only a weekly backup, so thinking that if I have two of these passport drives so that I can have one drive offsite and one attached to the server, and rotate them every 4 weeks so as not to loose all my data.

Here's my question: Ideally, I would love to just be able to unplug the current drive, plug in the new drive and have everything work. However, I don't see this actually working, but if there's a way to do this, that would be totally awesome.... ;-)

So, realistically, I know I will have to unmount the one drive, unplug it, then plug in the new drive and mount it on the system. Is there a way to mount this to the same mount point automatically so that I don't have to rewrite my backup script each time I swap drives out so that the backups go to the same mount point? Or will the UUID's get messed up each time I do this?

Hopefully this makes sense and an easy solution can be found to accomodate this idea.....

Thanks again for all your help. This site is awesome for newbies such as myself........

Mikey