Victor Muslin
victor at prodigy.net
Fri Apr 13 02:35:02 EST 2001
At 10:18 AM 4/12/2001 +1000, you wrote: Wouldn't it be better just to buffer the output and then just use one output to send the entire buffer ? Isn't that what Grisha was suggesting in the first place ? yes, this is what he is suggesting and exactly what I was trying to avoid in the first place. Also, while not being a CGI/Web guru, I can imagine a number of potential problems with having to buffer all of the output before sending it to the browser. First of all the output may be large and it may not be practical to assemble it in memory. Second it may be useful to allow browser to start rendering some output before it is completely created if it takes a long time to create dynamic output. Third, if the browser cancelled the request the only way to find out is to try to send the reply and get some sort of bad status. Imagine a script that does a set of time-consuming database queries to create the output. it would be useful to test whether the request was cancelled after each query by attempting to send something back (a space character perhaps) to see whether the socket is still open before doing the rest of the queries. Perhaps somebody could suggest how these scenarios could be handled with mod_python? I don't understand the problem with redirecting stdout. My understanding of Mod-Python is that it just keeps the intertpreter running so it reduces start up time. I understood that each CGI session is still, as always, a separate session that fires up it's own instance of the code. Is this not true ? I am using the "publisher" capability of mod_python. This is how I imagine it works (not having had the time to go through the code). There is a function -- call it handler() -- that handles requests. Let's say there are two identical concurrent requests. Both are handled by the same instance of Python interpeter that calls handler() for each one. All variables instantiated inside the handler() function are local to the function and, therefore, each request has its own instance of these variables. Variables that are module level (of the module where the handler() function is) of class-level are global, i.e. there is one instance of them in the interpreter and, therefore, they are shared by the requests. If this weren't the case you couldn't open a database connection once, for example, and keep it open instead of re-opening it for each request. sys.stdout happens to be a global variable and, therefore, shared by multiple instances of handler() and consequently by the requests. If code in one request reassigns it, the code in the other concurrent requests is affected. Secondly the idea of just rewriting something that works doesn't seem to be a good idea to me. I would say rewrite if you're finding you are doing a lot of mainatainence on existing code however if the code works well and has been tested, deployed etc, it would be better to interfere with it as little as possible unless you already know that current requirements will make a rewrite inevitable at some stage in the future (obviously I'm not just talking about a few lines of python here ). I think you are making my point here. I did not want to re-write anything. Wilson "Gregory (Grisha) Trubetskoy" wrote: > Victor - > > Rather than invent ways to deal with legacy CGI code, I would bite the > bullet and rewrite the code without the use of "print". There are too many > subtle gotchas with simulating CGI... > > Grisha > > On Sun, 8 Apr 2001, Victor Muslin wrote: > > > > > Sorry for a long message, but this requires a bit of explanation. I > > appreciate your patience in advance. > > > > I have a bunch of python legacy code that used to be part of a large > > CGI-based system. This code simply used print statements to output HTML as > > follows: > > > > def foo(): > > print 'html1' > > print 'html2' > > > > Now I want to convert CGI to mod_python, but I would like to re-use the > > legacy code with as little re-writing as possible (obviously the legacy > > code is a lot lengthier and more complicated than the example above). I am > > using the publisher module, which requires my code to return a string > > containing all of the HTML. So I thought I would be clever and do something > > like this: > > > > import sys, cStringIO > > def handler(req): > > out = sys.stdout = StringIO() > > foo() > > return out > > > > This works great as long as the second request does not arrive before the > > first one is done. Otherwise, the output gets screwed up. Since "out" is a > > local variable, each request has its own instance, but sys.stdout is a > > global. When the second request arrives, sys.stdout gets reassigned and the > > rest of the output produced by print statements in the foo() function goes > > to the new StringIO object. For example, if the second request arrives and > > gets executed between the two print statements of the first request, then > > the first request's output could be 'html1\n' and the second request's > > output could be 'html2\nhtml1\nhtml2\n'. > > > > Has anyone dealt with such a situation? Any clever suggestion would be > > appreciated as I hate to have to go into all the legacy code and change it > > to something like this: > > > > def foo(): > > out = 'html1\n' > > out = out + 'html2\n' > > return out > > > > def handler(req): > > return foo() > > > > Thanks in advance. > > __________________________________________________________________________________ Victor Muslin The power of accurate observation is frequently called cynicism by those who don't have it. - George Bernard Shaw -------------- next part -------------- An HTML attachment was scrubbed... URL: http://mailman.modpython.org/pipermail/mod_python/attachments/20010413/96c1a4d3/attachment-0003.htm
|