Reputation: 1046
When I use R open source, if not using a specific package, it's not possible handle data sets bigger than RAM memory. So I would like to know if it's possible handle big data sets applying PL/R functions inside PostgreSQL.
I didn't found any documentation about this.
Upvotes: 5
Views: 1815
Reputation: 125
As mentioned by Hong Ooi, PL/R loads an R interpreter into the PostgreSQL backend process. So your R code is running "in database".
There is no universal way to deal with memory limitations, but there are least two possible options:
See PL/R docs here: http://www.joeconway.com/plr/doc/index.html
I am guessing what you would really like to have is a data.frame in which the data is paged to and from an underlying database cursor transparently to your R code. This is on my long term TODO, but unfortunately I have not been able to find the time to work it out. I have been told that Oracle's R connector has this feature, so it seems it can be done. Patches welcomed ;-)
Upvotes: 11
Reputation: 57696
No. PL/R just starts up a separate R process to run your R code. This uses exactly the same binaries and executables as what you'd use from the command line, so all the standard limitations still apply.
Upvotes: 1