Reputation: 2141
I've built a nice website system that caters to the needs of a small niche market. I've been selling these websites over the last year by deploying copies of the software using Capistrano to my web server.
It occurs to me that the only difference in these websites is the database, the CSS file, and a small set of images used for the individual client's graphic design.
Everything else is exactly the same, or should be... Now that I have about 20 of these sites deployed, it is getting to be a hassle to keep them all updated with the same code. And this problem will only get worse.
I am thinking that I should refactor this system, so that I can use one set of deployed ruby code, dynamically selecting the correct database, etc, by the URL of the incoming request.
It seems that there are two ways of handling the database:
The multiple database approach would be the simplest for me at the moment, since I wouldn't have to refactor every model in my application to add the client_id field to all CRUD operations.
However, it would be a hassle to have to run 'rake db:migrate' for tens or hundreds of different databases, every time I want to migrate the database(s). Obviously this could be done by a script, but it doesn't smell very good.
On the other hand, every client will have 20K-50K items in an 'items' table. I am worried about the speed of fulltext searches when the items table has a half million or million items in it. Even with an index on the client_id field, I suspect that searches would be faster if the items were separated into different client databases.
If anyone has an informed opinion on the best way to approach this problem, I would very much like to hear it.
Upvotes: 5
Views: 2006
Reputation: 2141
I have decided to go with the multiple database approach. This is the easiest path for me, since I don't have to rework the entire application.
What I'm going to do is to add a before_filter in application_controller, so it applies to all controllers... something like this:
before_filter :client_db # switch to client's db
Then, in application_controller.rb
, I'll include something like this:
def client_db
@client = Client.find(params[:client_id])
spec = Client.configurations[RAILS_ENV]
new_spec = spec.clone
new_spec["database"] = @client.database_name
ActiveRecord::Base.establish_connection(new_spec)
end
Then, a URL like example.com?client_id=12345
will select the correct database.
Since I am using Apache as a proxy in front of Mongrel, Apache will add the correct client_id to all requests, based on the client's website URL. So the client_id won't actually be part of the URL that users see. It will only be passed between Apache and Mongrel. I'm not sure if I'm explaining this properly, but it works and keeps things clean and simple.
If I decide I need to use a single database in the future, I can refactor all the code then. At the moment, this seems to be the simplest approach.
Upvotes: 4
Reputation: 2589
I would go for a single database, using client IDs - you should be able to make the refactoring less painful by using some form of base model, and a named scope to scope any actions to that client's ID.
You could use an indexing library, such as Ferret, or something along those lines, to deal with the issue of full-text searches becoming slow. That's going to be an issue anyway once a single client's database becomes to big, so you may need to implement that either way.
Upvotes: 1
Reputation: 110489
There are advantages to using separate DBs (including those you already listed):
All that being said, the single DB solution is probably better given:
Upvotes: 2