IWETHEY v. 0.3.0 | TODO
1,095 registered users | 0 active users | 0 LpH | Statistics
Login | Create New User
IWETHEY Banner

Welcome to IWETHEY!

New rofl microsoft
http://www.pcworld.c...e_smartphone.html
his week, Microsoft announced that they had lost all Sidekick user data including t-mobile sidekickpictures, contacts, calendars and other information from the Danger's servers. Since the devices sync with the servers, the devices also lost the data. The Sidekick data services had amazingly been out over a week
New Check how the business press reports it
http://www.reuters.c...N1213255320091012


Microsoft said in an emailed statement that the recovery
process has been "incredibly complex" because it suffered a
confluence of errors from a server failure that hurt its main
and backup databases supporting Sidekick users.


Whiny SOBs, ain't they.
New Blame shifted to Hitachi
http://www.channelre...sidekick_hitachi/
It appears that the direct cause of the Sidekick data loss may have been storage area network remedial work outsourced to Hitachi Data Systems.

But I don't think they can blame Hitachi for this:
There was however, no backed-up or replicated data set...

A synced replica could have been vaporized if the damage was replicated to it, but no backup whatsoever??
New I was afriad of that....
Hitachi would NOT have gone forward without the "GO AHEAD" from the operators of the SAN... Danger/Microsoft.

No matter that problem, Microsoft is the one that screwed up here with *NO* backups.

I am sure though that T-Mobile will get the brunt of the push back and they will be the ones losing the customers.

The blame on this SQUARELY lies on the person(s) that were supposedly backing up this data. It is the *TASK* of the backups people to test a restore from time to time to make sure things are good. This tells me its Microsoft's subsidiary Danger.
New not so fast, with SAN connectivity getting smarter
backups are done on the SAN with snapshots not offline storage UNLESS there is a DR requirement that requires offline storage. Many folks are resistant to that because of cost and speed of same and dont do it. Getting more common in data centers
New Not so fast yourself...
Redundancy IS NOT a replacement for backups.

RAID IS NOT a replacement for backups.

SNAPSHOTS ON THE SAME SAN (and yes I have used them to restore from) ARE NOT a replacement for backups

Even though I have all that... I still have a backup solution... which then also replicates to a geo-redundant storage.

Tis no excuse to "rely" on a sure thing, that has historically been shown to fail. RAID fails regularly, Redundancy fails regularly, SNAPSHOTS can and have been proven bad regularly, backups have proven to be bad, geo-redundant storage has been proven to be bad in some cases (if the backup source is bad).

If you don't test... you'll never know.

New YOU do I ADVOCATE that, doesnt mean that it happens
New Re: YOU do I ADVOCATE that, doesnt mean that it happens
So, that means lost data eventually.

Plan for worst case and implement worst scenario recovery plans.

Plus, if you don't test regularly...

Backups are like Expensive insurance. You hate to spend money on it, hate to devote time to it, hate to have to have. But when all else fails and backups are there and save all of your customer's data... You'll be glad you had it.

Sort of like going with "no-fault" only on a 2010 Mercedes Benz 500SEL. If you crash it, you are stuck fixing it yourself and slowly and at great cost. If you had full coverage, you might even get a new 2010 to replace it. Almost as if nothing happened.

First time you have catastrophic Data loss, you better make sure you've archived *ALL* of your advice and be able to prove people off-put it, other wise... its your ass.
New Saved in several places
New inexcusable
Basic risk analysis would point this potential weakness out. We won't TOUCH a datastore or database server without a well-defined maintenance window with risks and contingencies identified, and the first step after taking the database offline is a backup. The risk of making the backup is infinitesimal compared to what they're facing now.

Also, any fool who'd design a service like that without a) at least one georedundant data center, and b) a separate geo-redundant backup system should be hung from the nearest yardarm. Likewise, anyone who'd create this task without a known-good backup as part of a fleshed out contingency plan should join him.

I've seen a fair number of articles (mostly written by idiots) claiming this is yet another reason that cloud computing is teh bad. Personally, I don't think you can apply that label here because this service is apparently only slightly better equipped than the high school webgenius with a stack of servers in a back bedroom cooled with a box fan and served from wifi stolen from a neighbor.

As a company that sells services based on cloud and grid computing, that kind of reporting is what pisses me off. Yes, we're georedundant on backups and content delivery, and will be on application serving within a year. And no we don't have on the order of 11MM users, but we apparently take access and security of our client data a hell of a lot more seriously than Danger.
New Little harsh there, don't you think?
You make it sound like computer systems are notoriously unreliable, perhaps due to their inherent complexity. Do you really think the recent history of computing bears out such a pessimistic outlook?
--

Drew
New I think it was right on...
I personally feel that the entire stack of people responsible for this screw up:

The people that gave the OK to Hitachi. FIRST!

The people in charge of backups (tested backups).

The people that planned this event without properly recognizing the risks and there fore having the contingency plan all setup.

The SAN Operators that allowed this to proceed with out proper SNAPSHOT saved off to another SAN, unless they were (provably) strong-armed.

Last but not least the CTO and IT Managers, possibly even the CEO as the ultimate responsibility is his.



And Yes, Computers are KNOWN to be that un-reliable. Google doesn't even fix broken machines in data centers. They have a policy that *IF* a machine stops responding, a reboot request to the "console setup" they have. If the machine comes back, it is automatically re-imaged and made "better". If it fails to respond, it is shutdown and left to decay in place until the "rack" itself is removed.

They figure its going to cost them a minimum of $600 to send a warm body out there, find the machine, reboot it and (possibly) fix it. When the cost of a new machine for them is so low... its not even worth the time and effort to deal with failures any other way than to ignore them.
New Well played, sir. :-)
New Oh dude... did he zing me or what?
Of course, I live it. So its hard to see the sarcasm when you are so close to it.
New That should be "'hanged' from the nearest yardarm" . . .
. . hung is something else.
New rman to a different array than database
I insisted on an rman to virt but was overruled. Its gonna happen sooner or later do have georedundancy for a different application in place, those folks understood
New Test? You're supposed to test?
Few years back, when employer was using tape backup, 3000 tapes per back, I was advocating changing. Told them that even with a 99.995% good rate, there was still going to be bad tapes. Nah never happen.

Did a Disaster recovery test. Had to go back 3 or 4 weeks of tapes before we found a complete set...

Much better now, still long way from what I'd call reliable.
New Yup, saw it coming
Point and click vendor techs are the most dangerous thing you can let loose in your server room.
New I have no direct experience with Hitachi...
... but their equipment is at the very high end of the scale. I doubt they would be sending in point and click techs to work on it.

(Closest I came was in '05, back in Belgium, working for a city hospital. We put out a public bid for a fully redundant SAN installation. We were drooling over the specs of their entry. Hitachi was then trying to get beyond the mainframe/supercomputer market, so the price was absolute killer. Unfortunately for them (and us), their partner vendor submitted the bid a day late...)
New EMC is as good as hitachi
and yes they do send in point and click techs
New I have buddy that for for HDS
He wasn't in implementation or service, but always spoke highly of their field staff. I got the impression there were some pretty impressively smart people working there.
     rofl microsoft - (boxley) - (20)
         Check how the business press reports it - (crazy)
         Blame shifted to Hitachi - (scoenye) - (18)
             I was afriad of that.... - (folkert) - (17)
                 not so fast, with SAN connectivity getting smarter - (boxley) - (12)
                     Not so fast yourself... - (folkert) - (11)
                         YOU do I ADVOCATE that, doesnt mean that it happens -NT - (boxley) - (9)
                             Re: YOU do I ADVOCATE that, doesnt mean that it happens - (folkert) - (1)
                                 Saved in several places -NT - (boxley)
                             inexcusable - (Steve Lowe) - (6)
                                 Little harsh there, don't you think? - (drook) - (3)
                                     I think it was right on... - (folkert)
                                     Well played, sir. :-) -NT - (Another Scott) - (1)
                                         Oh dude... did he zing me or what? - (folkert)
                                 That should be "'hanged' from the nearest yardarm" . . . - (Andrew Grygus)
                                 rman to a different array than database - (daemon)
                         Test? You're supposed to test? - (jbrabeck)
                 Yup, saw it coming - (crazy) - (3)
                     I have no direct experience with Hitachi... - (scoenye) - (2)
                         EMC is as good as hitachi - (boxley)
                         I have buddy that for for HDS - (Steve Lowe)

This isn't AOL, despite the best efforts of a nameless few.
131 ms