It's scanning the HD to try to recover any data it could have lost, I think...
^^^^ what he said. fsck is trying to fix a corrupted filesystem, meaning if it cannot, chances are strong you've just lost all your data on that disk. maybe the power dropped out on that box and it fooked the filesystem, fsck is usually run on reboot...
Thanks guys, this was acutally a planned downtime and they are talking about "no suspected loss of data" ... If they knew this would be possible what good reason could they have for not backing up all data before they went through the process?.
Yea, $65 free ... It's just a shame this doesn't cover my lost profits..
I'm still happy with zone.net other than this. In my opinion, this is not a fuckup on their part, rather a hardware failure of sort. Have you seen anything to imply otherwise? If so, I have no problems sticking with them regardless of this catastrophic downtime...
I dont know the company, nor ever used it, but a 350 day uptime is damn promising. Shit happened... I suggest you let it go...
Yeah. This rivals my SoftLayer 18 hour downtime because they re-assigned my IP block to another customer -twice- but I can't complain, isolated to one 18 hour window -and- 40% off next month's bill...
Still down ... getting a bit frustrated obviously. I'm going to get back online and find that google has deindexed all of my sites (like 20% of my total sites). I understand that they are working on the problem, but this alone is enough to make me go elsewhere even though I'm happy with them up till now .... I guess, however, this could easily happen on any server. It's just disasterous and no good can come of it at this point...
In case you missed it: Zone.net - Horrible Experience - Over 24 Hours Down and Counting - Web Hosting Talk - The largest, most influential web hosting community on the Internet..
I know that whenever it gets this far it is usually followed by a great deal of lost data. The storyline goes backups were corrupted, disk was bad, fine print says we are not liable, company stops responding...
2.5 days and you haven't start moving your stuff?..
Thanks for the link subigo ... hadn't read that..
Server just got back online but from what I understand the service is spotty now as EVERYONE is taking backups to move to a different host ... overloading the CPU.
Just checked 4 of my smaller sites and not a single indexed page on any of them. To me, that's really worse news than having to relocate to a different server..
Since I've weathered this storm (and prepaid until Feb 09) I'll probably hang tight and take my time looking around for a different host. I paid what I consider a premium for this VPS ($65) as there are many cheaper on the market ... customer support sucks either way you look at it..
I'm going to stop posting on this thread now that things are back to normal, thanks for all the input. I just wanted to comment real quick that this can happen to any server at any time..
BACKUPS ARE CRITICAL!!!.
I had them, but figured I'd wait out the storm before shifting my bets around the table. I'm 2nd guessing this logic now but hindsight is always 20/20 .. probably should have moved this stuff on Friday afternoon. But it really would have been jumping the gun to retreat after 12 hours down..
Regardless, they are comping 3 months and giving me a free hot swap. At least they are stepping up to the plate and taking responsibility for this issue. I'm sure this hurts financially, ooh well, not my buck :].
For the record, zone.net is not shit .. they just have bad luck..
You were doing backups but didnt move them?.
If your not already doing them, here's a quick automated way to get that going. I use a cheap arse Dreamhost account, cron, scp (ssh), and zip (or tar) for a Unix box..
- Setup a trust with ssh from your box over to the DH account (loads of tutorials out there, it's just the simple matter of copying a key over from DH).
- Setup a cron job to tar (or zip) up your files.
- Then have the cron scp over your files to your DH account..
Really cheap, quick, and easy - DH has gob's of disk space to dump crap on.....
Shit. That's pretty bad man..
We got DDoSd the weekend. Little fucker waited until Friday evening, Euro time, so there was well and truly no one in the office to take care of shit for a good few hours..
The boss was kicking himself about not having a 2nd server..
If this was scheduled downtime, I'd be asking two things.
1) How come it's taking so much longer than expected.
2) Why wouldn't they host your shit on another box temporarily and redirect as appropriate?.
They don't have fail-over or cascade setups at all?..