A while ago I wrote a plugin to limit uploads in Mongrel. Mongrel does
save uploads to a tempfile on disk - if the upload is bigger than about
12 KB - using a Ruby TempFile-object. (which stores in the $TMPDIR, /tmp
on most systems).
The request is handed over to Rails after it’s fully received by
Mongrel.
I’m not sure if this “saving to disk” works the same with chunked
uploads (uploads without a Content-Length header).
Maybe my plugin can help you: http://slasaus.netsend.nl/articles/show/7
(warning: it works, but is not very elegant).
Gl,
Tim
----- Original Message ----
From: Daniel B. [email protected]
To: [email protected]
Sent: Monday, October 15, 2007 2:00:20 PM
Subject: [Mongrel] POST with huge HTTP body
Hi,
I found the following page describing how to stream data from the
server to the client:
http://api.rubyonrails.com/classes/ActionController/Streaming.html
I want to do the opposite, streaming data from the client to the
server,
letting the controller saving the data while it’s being received.
In particular, I want to upload large files to the RailsDav
(http://www.liverail.net/railsdav) plugin, without having to store
the entire file in memory during the operation.
This seems to be a Mongrel problem, so I started looking at this
plugin:
http://mongrel.rubyforge.org/docs/upload_progress.html
How do I get Mongrel to store the incoming POST data in a tempfile,
passing a handler to that file as the raw_post field in the Rails
request?
Is it even possible?
How is the “request” parameter to the process() method related to the
“request” object in the Rails controllers?
/Daniel
Mongrel-users mailing list
[email protected]
http://rubyforge.org/mailman/listinfo/mongrel-users
____________________________________________________________________________________
Shape Yahoo! in your own image. Join our Network Research Panel today!
http://surveylink.yahoo.com/gmrs/yahoo_panel_invite.asp?a=7