pld-builder.new: PLD_Builder/request_fetcher.py, PLD_Builder/rpm_builder.py...
arekm
arekm at pld-linux.org
Tue Sep 1 11:28:49 CEST 2009
Author: arekm Date: Tue Sep 1 09:28:49 2009 GMT
Module: pld-builder.new Tag: HEAD
---- Log message:
Correctly use urllib2 API.
---- Files affected:
pld-builder.new/PLD_Builder:
request_fetcher.py (1.27 -> 1.28) , rpm_builder.py (1.106 -> 1.107)
---- Diffs:
================================================================
Index: pld-builder.new/PLD_Builder/request_fetcher.py
diff -u pld-builder.new/PLD_Builder/request_fetcher.py:1.27 pld-builder.new/PLD_Builder/request_fetcher.py:1.28
--- pld-builder.new/PLD_Builder/request_fetcher.py:1.27 Tue Sep 1 11:16:39 2009
+++ pld-builder.new/PLD_Builder/request_fetcher.py Tue Sep 1 11:28:44 2009
@@ -41,8 +41,9 @@
signal.signal(signal.SIGALRM, alarmalarm)
signal.alarm(300)
try:
- headers = urllib.urlencode( { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' } )
- f = urllib2.urlopen(control_url + "/max_req_no", headers)
+ headers = { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' }
+ req = urllib2.Request(url=control_url + "/max_req_no", headers=headers)
+ f = urllib2.urlopen(req)
count = int(string.strip(f.readline()))
signal.alarm(0)
except Exception, e:
@@ -60,8 +61,9 @@
socket.setdefaulttimeout(240)
signal.alarm(300)
try:
- headers = urllib.urlencode( { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' } )
- f = urllib2.urlopen(control_url + "/queue.gz", headers)
+ headers = { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' }
+ req = urllib2.Request(url=control_url + "/queue.gz", headers=headers)
+ f = urllib2.urlopen(req)
signal.alarm(0)
except Exception, e:
signal.alarm(0)
================================================================
Index: pld-builder.new/PLD_Builder/rpm_builder.py
diff -u pld-builder.new/PLD_Builder/rpm_builder.py:1.106 pld-builder.new/PLD_Builder/rpm_builder.py:1.107
--- pld-builder.new/PLD_Builder/rpm_builder.py:1.106 Tue Sep 1 11:16:39 2009
+++ pld-builder.new/PLD_Builder/rpm_builder.py Tue Sep 1 11:28:44 2009
@@ -5,7 +5,6 @@
import atexit
import time
import string
-import urllib
import urllib2
from config import config, init_conf
@@ -58,8 +57,9 @@
b.log_line("checking if we should skip the build")
while not good:
try:
- headers = urllib.urlencode( { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' } )
- f = urllib2.urlopen(src_url, headers)
+ headers = { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' }
+ req = urllib2.Request(url=src_url, headers=headers)
+ f = urllib2.urlopen(req)
good = True
except urllib2.HTTPError, error:
return False
@@ -81,8 +81,9 @@
good = False
while not good:
try:
- headers = urllib.urlencode( { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' } )
- f = urllib2.urlopen(src_url, headers)
+ headers = { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' }
+ req = urllib2.Request(url=src_url, headers=headers)
+ f = urllib2.urlopen(req)
good = True
except urllib2.HTTPError, error:
# fail in a way where cron job will retry
================================================================
---- CVS-web:
http://cvs.pld-linux.org/cgi-bin/cvsweb.cgi/pld-builder.new/PLD_Builder/request_fetcher.py?r1=1.27&r2=1.28&f=u
http://cvs.pld-linux.org/cgi-bin/cvsweb.cgi/pld-builder.new/PLD_Builder/rpm_builder.py?r1=1.106&r2=1.107&f=u
More information about the pld-cvs-commit
mailing list