curl-and-python
RE: Function to retrieve mulltiple URLs asyncrously
Date: Fri, 11 Mar 2005 14:16:50 +0100
ok, then perhaps that's the source of the problem the original poster
encountered (as the code that was posted had a select call with a timeout).
i'll change the select call to require a timeout to be set, and update the
example and test code in pycurl to use this.
curl_multi_timeout will surely be handy, in the meanwhile a 1.0 sounds like
a good number to use :)
- kjetil
-----Original Message-----
From: curl-and-python-bounces_at_cool.haxx.se
[mailto:curl-and-python-bounces_at_cool.haxx.se] On Behalf Of Daniel Stenberg
Sent: 11 March 2005 14:01
To: curl stuff in python
Cc: 'gf gf'
Subject: RE: Function to retrieve mulltiple URLs asyncrously
On Fri, 11 Mar 2005, Kjetil Jacobsen wrote:
> that code is part of an outer loop which subsequently calls info_read
> and select. this whole procedure (perform, info_read, select) is done
> until there is no more work to do.
Ah, ok. I missed that. info_read should however only be necessary if there
actually is one or more transfers that completed.
> one thing though -- the code does a select(..) without at timeout, so
> unless there is activity on the file descriptors this will block
> infinitely. is this harmless or does the multi api assume that select
> times out periodically?
You need to call it periodically to allow it to do its timers and things.
An future libcurl release will feature a curl_multi_timeout() that'll let
you know the longest possible time you should wait before calling libcurl
again (unless action was detected on the sockets of course).
-- Daniel Stenberg -- http://curl.haxx.se -- http://daniel.haxx.se Dedicated custom curl help for hire: http://haxx.se/curl.html _______________________________________________ http://cool.haxx.se/mailman/listinfo/curl-and-python _______________________________________________ http://cool.haxx.se/mailman/listinfo/curl-and-pythonReceived on 2005-03-11