I am writing an app for basically internal use for a client for their
billing process. There are a few large files which need to be uploaded
or
added to the system as in input to this process. I am wondering whether
it
would be better to set up ftp or whether rails running on any server
(may be
just Mongrel as this is an internal app that will probably have one or a
few
users max) would handle this file size gracefully? Does it simply have
to do
with the memory of the server or are there other factors? Since this is
an
internal process it does not have to be slick like something public
facing,
most important is reliability and the files getting uploaded right the
first
time.
Check out paperclip. It will allow you to upload to and download from
your local file system (or S3). In the local file system, it stores the
docs in public/system. I am using it for smaller (up to 30Mb) files and
using Pat Shaunessey’s version to save the data to db tables (not
recommended for files of the magnitude you mentioned). It works fine
and is really fast to incorporate into your code.
On Tue, Jan 11, 2011 at 6:07 PM, Donald R. Ziesig [email protected]
wrote:
David,
Check out paperclip. It will allow you to upload to and download from your
local file system (or S3). In the local file system, it stores the docs in
public/system. I am using it for smaller (up to 30Mb) files and using Pat
Shaunessey’s version to save the data to db tables (not recommended for
files of the magnitude you mentioned). It works fine and is really fast to
incorporate into your code.
Thanks, that is probably at least worth a shot to see how it performs. I
had
not used it with files this size so was just curious if it was
contraindicated.
This forum is not affiliated to the Ruby language, Ruby on Rails framework, nor any Ruby applications discussed here.