Why NoSQL is better at “scaling out” than RDBMS?

So I’ve been trying to figure out the real bottom-line when it comes to NoSQL vs RDBMS myself, and always end up with a response that doesn’t quite cut it. In my search there are really 2 primary differences between NoSQL and SQL, with only 1 being a true advantage.

  1. ACID vs BASE – NoSQL typically leaves out some of the ACID features of SQL, sort of ‘cheating’ it’s way to higher performance by leaving this layer of abstraction to the programmer. This has already been covered by previous posters.

  2. Horizontal Scaling – The real advantage of NoSQL is horizontal scaling, aka sharding. Considering NoSQL ‘documents’ are sort of a ‘self-contained’ object, objects can be on different servers without worrying about joining rows from multiple servers, as is the case with the relational model.

Let’s say we want to return an object like this:

post {
    id: 1
    title: 'My post'
    content: 'The content'
    comments: {
      comment: {
        id: 1
      }
      comment: {
        id: 2
      }
      ...

    views: {
      view: {
        user: 1
      }
      view: {
        user: 2
      }
      ...
    }
}

In NoSQL, that object would basically be stored as is, and therefore can reside on a single server as a sort of self-contained object, without any need to join with data from other tables that could reside on other DB servers.

However, with Relational DBs, the post would need to join with comments from the comments table, as well as views from the views table. This wouldn’t be a problem in SQL ~UNTIL~ the DB is broken into shards, in which case ‘comment 1’ could be on one DB server, while ‘comment 2’ yet on another DB server. This makes it much more difficult to create the very same object in a RDBMS that has been scaled horizontally than in a NoSQL DB.

Would any DB experts out there confirm or argue these points?

Leave a Comment