本文整理汇总了Python中mechanize.install_opener函数的典型用法代码示例。如果您正苦于以下问题:Python install_opener函数的具体用法?Python install_opener怎么用?Python install_opener使用的例子?那么恭喜您, 这里精选的函数代码示例或许可以为您提供帮助。
在下文中一共展示了install_opener函数的20个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于我们的系统推荐出更棒的Python代码示例。
示例1: retrieve_product_data
def retrieve_product_data(self, product_link):
cookies = mechanize.CookieJar()
opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(cookies))
opener.addheaders = [('User-agent', 'Mozilla/5.0 (MyProgram/0.1)'),
('From', '[email protected]')]
mechanize.install_opener(opener)
browser = mechanize.Browser()
product_data = browser.open(product_link).get_data()
soup = BeautifulSoup(product_data)
product_name = soup.find('title').string.encode('ascii', 'ignore')
product_prices = soup.find('div', 'price').contents
try:
cash_price = int(clean_price_string(product_prices[4]))
product_data = ProductData()
product_data.custom_name = product_name
product_data.price = cash_price
product_data.url = product_link
product_data.comparison_field = product_link
return product_data
except IndexError:
return None
开发者ID:vkhemlan,项目名称:solotodo,代码行数:26,代码来源:global_mac.py
示例2: _retrieve_product
def _retrieve_product(cls, url):
cookies = mechanize.CookieJar()
opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(cookies))
opener.addheaders = [('User-agent', 'Mozilla/5.0 (MyProgram/0.1)'),
('From', '[email protected]')]
mechanize.install_opener(opener)
browser = mechanize.Browser()
product_data = browser.open(url).get_data()
soup = BeautifulSoup(product_data)
product_name = soup.find('h1').string.encode('ascii', 'ignore')
product_price = soup.find('span', {'id': 'product_price'})
product_price = Decimal(clean_price_string(product_price.string))
payment_methods = ['cash', 'deposit', 'wire_transfer']
additional_data = soup.find('td', 'descr').findAll('h3')
if not additional_data:
payment_methods.extend(['debit_card', 'credit_card'])
elif additional_data[0].string and 'Contado' not in \
additional_data[0].string:
payment_methods.extend(['debit_card', 'credit_card'])
prices = {}
for p in payment_methods:
prices[p] = product_price
return [product_name, prices]
开发者ID:SoloTodo,项目名称:storescrapper,代码行数:29,代码来源:global_mac.py
示例3: __init__
def __init__(self, username="RP\\12345", password="abcdef"):
self.username = "RP\\"+username
self.password = password
self.password_manager = urllib2.HTTPPasswordMgrWithDefaultRealm()
ntlm_auth = HTTPNtlmAuthHandler.HTTPNtlmAuthHandler(self.password_manager)
opener = mechanize.build_opener(ntlm_auth)
mechanize.install_opener(opener)
开发者ID:Dino198,项目名称:PythonLeo,代码行数:7,代码来源:PythonLeo.py
示例4: customizeUserAgent
def customizeUserAgent():
import mechanize
cookies = mechanize.CookieJar()
opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(cookies))
# Pretend to be Chrome to avoid getting the mobile site.
opener.addheaders = [("User-agent", "Chrome/16.0.912.63")]
mechanize.install_opener(opener)
开发者ID:georgevdd,项目名称:bank-account-reporting,代码行数:7,代码来源:Halifax.py
示例5: init
def init(self):
br = mechanize.Browser()
br.set_handle_robots(False)
self.cj = mechanize.LWPCookieJar()
br.set_cookiejar(self.cj)
br.set_handle_refresh(mechanize._http.HTTPRefreshProcessor(),
max_time=1)
br.open("https://www.tumblr.com/login")
br.select_form(nr=0)
br['user[email]'] = ""
br['user[password]'] = ""
url, data, hdrs = br.form.click_request_data()
br.open("https://www.tumblr.com/login", data)
self.nf = 0
opener = mechanize.build_opener(
mechanize.HTTPCookieProcessor(self.cj))
mechanize.install_opener(opener)
self._fetch()
开发者ID:non117,项目名称:boxnya-plugin,代码行数:25,代码来源:tumblrnotify.py
示例6: retrieve_product_links
def retrieve_product_links(self):
cookies = mechanize.CookieJar()
opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(cookies))
opener.addheaders = [('User-agent', 'Mozilla/5.0 (MyProgram/0.1)'),
('From', '[email protected]')]
mechanize.install_opener(opener)
url_base = 'http://www.globalmac.cl/'
browser = mechanize.Browser()
url_extensions = [
['Distribuidor-Apple-Chile/MacBook-Air', 'Notebook'],
['Distribuidor-Apple-Chile/MacBook-Pro', 'Notebook'],
['Hardware-Mac-PC/Discos-Duros-Notebook-SATA-2.5', 'StorageDrive'],
['Hardware-Mac-PC/Discos-Duros-SATA-3.5', 'StorageDrive'],
['Hardware-Mac-PC/Discos-Duros-SSD-SATA-2.5', 'StorageDrive'],
]
product_links = []
for url_extension, ptype in url_extensions:
url = url_base + url_extension
base_data = browser.open(url).get_data()
soup = BeautifulSoup(base_data)
for item in soup.findAll('div', 'name'):
product_links.append([item.find('a')['href'], ptype])
return product_links
开发者ID:vkhemlan,项目名称:solotodo,代码行数:29,代码来源:global_mac.py
示例7: _checkStoredInjections
def _checkStoredInjections(self):
for r in self.results:
# At this state injections in Result obj are not
# compacted yet so it will only be 1st injected param
url, data = r.target.getPayloadedUrl(r.first_param, "")
# In case of proxy
if self.engine.getOption('http-proxy') is not None:
proxy = ProxyHandler({'http': self.engine.getOption('http-proxy')})
opener = build_opener(proxy)
install_opener(opener)
# Some headers
if self.engine.getOption('ua') is not None:
if self.engine.getOption('ua') is "RANDOM":
headers = {'User-Agent': random.choice(USER_AGENTS)}
else:
headers = {'User-Agent': self.engine.getOption('ua')}
else:
headers = {}
if self.engine.getOption("cookie") is not None:
headers["Cookie"] = self.engine.getOption("cookie")
# Build the request
req = Request(url, data, headers)
try:
to = 10 if self.engine.getOption('http-proxy') is None else 20
response = urlopen(req, timeout=to)
except HTTPError, e:
self._addError(e.code, r.target.getAbsoluteUrl())
continue
except URLError, e:
self._addError(e.reason, r.target.getAbsoluteUrl())
continue
开发者ID:Drx51,项目名称:Framework,代码行数:34,代码来源:scanner.py
示例8: _performInjections
def _performInjections(self, target):
# Check every parameter
for k, v in target.params.iteritems():
pl = Payload(taint=True)
url, data = target.getPayloadedUrl(k, pl.payload)
# In case of proxy
if self.engine.getOption('http-proxy') is not None:
proxy = ProxyHandler({'http': self.engine.getOption('http-proxy')})
opener = build_opener(proxy)
install_opener(opener)
# Some headers
if self.engine.getOption('ua') is not None:
if self.engine.getOption('ua') is "RANDOM":
headers = {'User-Agent': random.choice(USER_AGENTS)}
else:
headers = {'User-Agent': self.engine.getOption('ua')}
else:
headers = {}
if self.engine.getOption("cookie") is not None:
headers["Cookie"] = self.engine.getOption("cookie")
# Build the request
req = Request(url, data, headers)
try:
to = 10 if self.engine.getOption('http-proxy') is None else 20
response = urlopen(req, timeout=to)
except HTTPError, e:
self._addError(e.code, target.getAbsoluteUrl())
return
except URLError, e:
self._addError(e.reason, target.getAbsoluteUrl())
return
开发者ID:Drx51,项目名称:Framework,代码行数:33,代码来源:scanner.py
示例9: readUrl
def readUrl(inUrl):
tryCount = 0
while tryCount < 5 :
# print "Create CookieJar"
cookies = mechanize.CookieJar()
# print "Build Opener"
opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(cookies))
# print "Add Headers"
opener.addheaders = [("User-agent", "Mozilla/5.0 (compatible; MyProgram/0.1)"),("From", "[email protected]")]
# print "Install Opener"
mechanize.install_opener(opener)
try:
# print "Open URL"
response = mechanize.urlopen(inUrl)
tryCount = 99
except:
tryCount += 1
print "******** Error on urlopen ***********"
print "URL: ", inUrl
print "Trying Again....", tryCount
# print response.read()
# html = urllib.urlopen(inUrl).read()
# print "Reading Response"
html = response.read()
# print "Response Read:", html[0:100]
root = lxml.html.fromstring(html)
# print "Root created: ", root
return root
开发者ID:carriercomm,项目名称:scraperwiki-scraper-vault,代码行数:31,代码来源:gptest02_1.py
示例10: test_cookies
def test_cookies(self):
import urllib2
# this test page depends on cookies, and an http-equiv refresh
#cj = CreateBSDDBCookieJar("/home/john/db.db")
cj = CookieJar()
handlers = [
HTTPCookieProcessor(cj),
HTTPRefreshProcessor(max_time=None, honor_time=False),
HTTPEquivProcessor(),
HTTPRedirectHandler(), # needed for Refresh handling in 2.4.0
# HTTPHandler(True),
# HTTPRedirectDebugProcessor(),
# HTTPResponseDebugProcessor(),
]
o = apply(build_opener, handlers)
try:
install_opener(o)
try:
r = urlopen(urljoin(self.uri, "/cgi-bin/cookietest.cgi"))
except urllib2.URLError, e:
#print e.read()
raise
data = r.read()
#print data
self.assert_(
data.find("Your browser supports cookies!") >= 0)
self.assert_(len(cj) == 1)
# test response.seek() (added by HTTPEquivProcessor)
r.seek(0)
samedata = r.read()
r.close()
self.assert_(samedata == data)
开发者ID:Almad,项目名称:Mechanize,代码行数:35,代码来源:functional_tests.py
示例11: slurp_with_login_and_pwd
def slurp_with_login_and_pwd():
import sys
import mechanize
# sys.path.append('ClientCookie-1.0.3')
# from mechanize import ClientCookie
# sys.path.append('ClientForm-0.1.17')
# import ClientForm
# Create special URL opener (for User-Agent) and cookieJar
cookieJar = mechanize.CookieJar()
opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(cookieJar))
opener.addheaders = [("User-agent","Mozilla/5.0 (compatible)")]
mechanize.install_opener(opener)
fp = mechanize.urlopen("http://login.yahoo.com")
forms = mechanize.ParseResponse(fp)
fp.close()
# print forms on this page
for form in forms:
print "***************************"
print form
form = forms[0]
form["login"] = "yahoo-user-id" # use your userid
form["passwd"] = "password" # use your password
fp = mechanize.urlopen(form.click())
fp.close()
fp = mechanize.urlopen("https://class.coursera.org/ml-003/lecture/download.mp4?lecture_id=1") # use your group
fp.readlines()
fp.close()
开发者ID:yz-,项目名称:ut,代码行数:31,代码来源:slurping.py
示例12: GetHtml
def GetHtml(url):
opener = mechanize.build_opener()
opener.addheaders = [("User-Agent", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10.6; rv:2.0) Gecko/20100101 Firefox/4.0")]
mechanize.install_opener(opener)
request = mechanize.urlopen(url)
html = request.read()
request.close()
return html
开发者ID:yanigisawa,项目名称:PyComicSyndicator,代码行数:8,代码来源:htmlFetch.py
示例13: setup_mechanize
def setup_mechanize():
"""
Set up user agent for all mechanize calls.
"""
cookies = mechanize.CookieJar()
opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(cookies))
homepage = "http://github.com/aszlig/picfetcher"
opener.addheaders = [("User-agent", "PicFetcher/0.1.0 (+%s)" % homepage)]
mechanize.install_opener(opener)
开发者ID:aszlig,项目名称:picfetcher,代码行数:9,代码来源:picturefetch.py
示例14: openUrl
def openUrl(url, cookie=None, login=False):
"""
Opens a given url through mechanize.
If there is no cookie (string path) passed in or if there is a cooke path
passed in but the login parameter is False (signifying to open the url with
cookie saved in the cookie path), the html from the opened url is returned
as a string.
If a cookie path is passed in and the login parameter is True, then the
Mechanize.Broswer object is returned to perform a yogaglo login through
a form submission.
"""
browser = mechanize.Browser()
browser.addheaders = [
('User-Agent',
'Mozilla/5.0 (Macintosh; Intel Mac OS X 10.9; rv:24.0) Gecko/20100101 Firefox/24.0'),
('Accept',
'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8'),
('Accept-Language', 'en-gb,en;q=0.5'),
('Accept-Charset', 'ISO-8859-1,utf-8;q=0.7,*;q=0.7'),
('Keep-Alive', '115'),
('Connection', 'keep-alive'),
('Cache-Control', 'max-age=0'),
]
#Experimental?
# browser.set_handle_gzip(True)
browser.set_handle_redirect(True)
browser.set_handle_referer(True)
browser.set_handle_robots(False)
browser.set_handle_refresh(mechanize._http.HTTPRefreshProcessor(), max_time = 1)
if not cookie is None:
cj = cookielib.LWPCookieJar()
browser.set_cookiejar(cj)
opener = mechanize.build_opener(HTTPCookieProcessor(cj))
mechanize.install_opener(opener)
# trying to login, no cookie, must return browser so it can follow the
# login url
if login is True:
browser.open(url)
return browser
# can't set to expire, can't read when this particular cookie expires
cj.load(cookie , ignore_discard=True)
return browser.open(url).read()
开发者ID:jacobono,项目名称:xbmc-yogaglo-plugin,代码行数:50,代码来源:http.py
示例15: setUp
def setUp(self):
mechanize._testcase.TestCase.setUp(self)
self.test_uri = urljoin(self.uri, "test_fixtures")
self.server = self.get_cached_fixture("server")
if self.no_proxies:
old_opener_m = mechanize._opener._opener
old_opener_u = urllib2._opener
mechanize.install_opener(mechanize.build_opener(
mechanize.ProxyHandler(proxies={})))
urllib2.install_opener(urllib2.build_opener(
urllib2.ProxyHandler(proxies={})))
def revert_install():
mechanize.install_opener(old_opener_m)
urllib2.install_opener(old_opener_u)
self.add_teardown(revert_install)
开发者ID:kovidgoyal,项目名称:mechanize,代码行数:15,代码来源:test_functional.py
示例16: fillform
def fillform(self, form, choice,questionid,sessionid,charturl,user,password):
if choice != "Random":
for i in range(1, 5):
form[questionid + str(i)] = [choice]
else:
for i in range(1, 5):
form[questionid + str(i)] = [str(random.randint(1, 5))]
data = form.click().get_data()
charturl += sessionid + "&questionid=" + questionid + "&qtype=" + "LS"
opener = self.addAuthentication(charturl, user, password)
mechanize.install_opener(opener)
req = mechanize.Request(charturl, data)
req.add_header('User-Agent', 'Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US; rv:1.9.1.6) Gecko/20091201 Firefox/3.5.6')
req.add_header('Accept', 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8')
req.add_header('Accept-Encoding', 'gzip,deflate')
开发者ID:Dino198,项目名称:PythonLeo,代码行数:15,代码来源:AutoEvaluation_v4.py
示例17: __init__
def __init__(self, username, password):
mechanize.Browser.__init__(self)
cj = mechanize.LWPCookieJar()
self.set_cookiejar(cj)
self.set_handle_equiv(True)
self.set_handle_redirect(True)
self.set_handle_referer(True)
self.set_handle_robots(False)
self.addheaders = [('User-agent', 'Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.0.1) Gecko/2008071615 Fedora/3.0.1-1.fc9 Firefox/3.0.1')]
self.open(self.base_url)
self.username = username
self.password = password
self.login()
opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(cj))
mechanize.install_opener(opener)
开发者ID:zacharydenton,项目名称:BoredAtBot,代码行数:17,代码来源:BoredAtBot.py
示例18: acm
def acm(query_str):
acm_url = u"http://dl.acm.org/"
cookieJar = mechanize.CookieJar()
opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(cookieJar))
opener.addheaders = [("User-agent","Mozilla/5.0 (compatible)")]
mechanize.install_opener(opener)
fp = mechanize.urlopen(acm_url)
forms = mechanize.ParseResponse(fp, backwards_compat=False)
fp.close()
#doc = fetch(acm_url)
form = forms[0]
form['query'] = query_str
fp = mechanize.urlopen(form.click())
doc = fp.read()
with open("acm.html", 'wb') as fo:
fo.write(doc)
fp.close()
开发者ID:pebbie,项目名称:BIBINT,代码行数:19,代码来源:scrape.py
示例19: _product_urls_and_types
def _product_urls_and_types(cls, product_types):
cookies = mechanize.CookieJar()
opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(cookies))
opener.addheaders = [('User-agent', 'Mozilla/5.0 (MyProgram/0.1)'),
('From', '[email protected]')]
mechanize.install_opener(opener)
url_base = 'http://www.globalmac.cl/'
browser = mechanize.Browser()
url_extensions = [
['MacBook/', 'Notebook'],
['MacBook-Pro/', 'Notebook'],
['Monitores-LCD/', 'Monitor'],
['Cinema-Display/', 'Monitor'],
['Disco-Duro-SATA-2.5/', 'StorageDrive'],
['Discos-Duros-SATA/', 'StorageDrive'],
]
if 'Ram' in product_types:
memory_catalog_url = url_base + 'Memorias/'
base_data = browser.open(memory_catalog_url).get_data()
soup = BeautifulSoup(base_data)
subcats = soup.findAll('span', 'subcategories')
for subcat in subcats:
link = subcat.find('a')['href'].replace(url_base, '')
url_extensions.append([link, 'Ram'])
product_links = []
for url_extension, ptype in url_extensions:
if ptype not in product_types:
continue
base_data = browser.open(url_base + url_extension).get_data()
soup = BeautifulSoup(base_data)
titles = soup.findAll('a', 'product-title')
for title in titles:
product_links.append([title['href'], ptype])
return product_links
开发者ID:SoloTodo,项目名称:storescrapper,代码行数:42,代码来源:global_mac.py
示例20: __init__
def __init__(self, login=login.facebook):
super(KaggLoader, self).__init__()
self.login = login
self.set_handle_equiv(True)
self.set_handle_robots(False)
if not os.path.exists(self.BASE_DIR):
os.makedirs(self.BASE_DIR)
if not os.path.exists(self.COOKIE_PATH):
with open(self.COOKIE_PATH, 'w') as f:
f.write('#LWP-Cookies-2.0')
self.cj = mechanize.LWPCookieJar()
self.cj.load(self.COOKIE_PATH, ignore_discard=False, ignore_expires=False)
opener = mechanize.build_opener(mechanize.HTTPCookieProcessor(self.cj))
mechanize.install_opener(opener)
self.set_cookiejar(self.cj)
开发者ID:Sundrique,项目名称:kaggloader,代码行数:21,代码来源:kaggloader.py
注:本文中的mechanize.install_opener函数示例由纯净天空整理自Github/MSDocs等源码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。 |
请发表评论