| 
 | 
 
马上注册,结交更多好友,享用更多功能^_^
您需要 登录 才可以下载或查看,没有账号?立即注册  
 
x
 
我爬取boss直聘,返回的是空列表,不知道怎么回事,还有就是我想要将第一个页面到第十个页面分别写入十个csv文件该怎么写?- import urllib.request
 
 - import urllib.parse
 
 - from lxml import etree
 
 - def creat_request(my_page):
 
 -     base_url = 'https://www.zhipin.com/c101270100/?'
 
 -     headers = {
 
 -         'Accept': ' text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.9',
 
 -         # 'Accept-Encoding': ' gzip, deflate, br',
 
 -         'Accept-Language': ' zh,en-US;q=0.9,en;q=0.8,zh-CN;q=0.7',
 
 -         'Connection': ' keep-alive',
 
 -         'Cookie': ' acw_tc=0bcb2f0716324609378892612e49ac132e87df9e2107cbb14b08e84e449455; __g=-; Hm_lvt_194df3105ad7148dcf2b98a91b5e727a=1632460940; lastCity=100010000; __l=l=%2Fwww.zhipin.com%2Fc101270100%2F%3Fka%3Dsel-city-101270100&r=https%3A%2F%2Fcn.bing.com%2F&g=&s=3&friend_source=0&s=3&friend_source=0; __c=1632460940; __a=34040189.1632460940..1632460940.9.1.9.9; Hm_lpvt_194df3105ad7148dcf2b98a91b5e727a=1632461742; __zp_stoken__=f68bdWDVKZBhBNRMPdCExbkphQAJsIzQwI3ROPh9gXUpCAQYhd3ENZnFRIh87IjdMJkx%2BFx06Dy43OjZAYwE9CB9LDXIeIWlAQR5kD1QNLx9LKzFwTQZrYCAHenNLDxEMZHVMP2BODQ1gdEY%3D',
 
 -         'Host': ' www.zhipin.com',
 
 -         'sec-ch-ua': ' "Chromium";v="94", "Google Chrome";v="94", ";Not A Brand";v="99"',
 
 -         'sec-ch-ua-mobile': ' ?0',
 
 -         'sec-ch-ua-platform': ' "Windows"',
 
 -         'Sec-Fetch-Dest': ' document',
 
 -         'Sec-Fetch-Mode': ' navigate',
 
 -         'Sec-Fetch-Site': ' none',
 
 -         'Sec-Fetch-User': ' ?1',
 
 -         'Upgrade-Insecure-Requests': ' 1',
 
 -         'User-Agent': ' Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/94.0.4606.61 Safari/537.36',
 
 -     }
 
 -     data = {
 
 -         'page' : my_page,
 
 -         'ka' : 'page-'+ str(my_page),
 
 -     }
 
 -     data = urllib.parse.urlencode(data)
 
 -     url = base_url + data
 
 -     print(url)
 
 -     request = urllib.request.Request(url=url,headers=headers)
 
 -     return request
 
  
- def get_content(request):
 
 -     response = urllib.request.urlopen(request)
 
 -     content = response.read().decode('utf-8')
 
 -     return content
 
  
- def down_load(content):
 
 -     tree = etree.HTML(content)
 
 -     name_list = tree.xpath('//*[@id="main"]/div/div[2]/ul/li//div/span/a/text()')
 
 -     place_list = tree.xpath('//*[@id="main"]/div/div[2]/ul/li//div/span[2]/span/text()')
 
 -     for i in range(len(name_list)):
 
 -         name = name_list[i]
 
 -         place = place_list[i]
 
 -     print(place_list)
 
  
 
 
- if __name__ == '__main__':
 
 -     start_page = int(input("请输入起始页码:"))
 
 -     end_page = int(input("请输入结束页码:"))
 
 -     for my_page in range(start_page,end_page+1):
 
 -         request = creat_request(my_page)
 
 -         content = get_content(request)
 
 -         down_load(content)
 
  复制代码 |   
 
 
 
 |