- “And nobody expressed any concern that this might not be right?“
- “Bill, you are building a system, not a collection of parts.“
Category Archives: Distributed Computing
Reliability
Steve Vinoski talking about
To achieve reliability, you have to accept the fact that failure will occur, Once you accept that, then other things fall into place: you need to be able to restart things quickly, and to do that, processes need to be cheap. If something fails, you don’t want it taking everything else with it, so you need to at least minimize, if not eliminate, sharing, which leads you to message passing. You also need monitoring capabilities that can detect failed processes and restart them
Amen to that!
23 January 2008
Towards Robust Distributed Systems – Eric Brewer’s slidedeck that kicked off the wholeCAP Theorem debate atPODC in 2000.
18 Jan 2008
- London
Erlang Meeting – “How do you test large systems written in Erlang?” on Monday (I presume) Jan 21. - Map Reduce a major step back. Michael
Stonebraker and David J.DeWitt let forth on why they dislike Google’sMapReduce infrastructure and how it is poor relation to standardrelational database technology. Relational Database Experts Jump The MapReduce Shark – a reply to Stonebraker and DeWitt that I pretty broadly agree with.- MapReducing 20
petabytes per day – Greg Linden commenting on the recent paper “MapReduce: Simplified Data Processing on Large Clusters” by Jeff Dean and Sanjay Ghemawat.
Eventual Consistency
Werner has a new piece on
Amazon DB and Erlang
The world and it’s dog is jumping on the
I am keen to see how the DB heads respond to this. I also want to see how how long it takes for people to get their heads around the
July 23 2007
- Interesting job spec from
Vogels …
My favourite line: “You have at least once tried to understandPaxos by reading the original paper.“. It all sounds like ancient greek to me!
Thankgod for chaos!
It’s always nice to get an Ebay perspective on things – especially when you totally agree with what is being said. For a long time I have felt that if we really want to embrace the power of large scale distributed systems we have to accept that we just do not know what is going on – life is non-determinstic, messy, chaotic and generally random. Most tech orientated people are used to thinking in a single linear fashion. First the processor does this, then it does that. Nice – but not the way the world works. Things actually happen at the same time. Just not neccasarily on your processor or in your address space. For many people dealing with that is a problem and something that should be somehow hidden away or, at the very least, forced into a 2PC harness. That might work in a small system but not when you reach internet scale systems. Forget all about a linear world and move to thinking about multiple agents doing multiple things at the same time.
This is a thought process that is going to increasingly dominant the way we build systems.
Collaborative Filtering – Google style
I have always had a soft spot for
Via Greg
The river is flowing….
The
Great stuff!