关于Python课程第54讲爬虫有道,只能爬到网页代码。
代码完全是按照小甲鱼的来弄的,也去掉了_o,就是得不到小甲鱼的效果。 放代码 wp231957 发表于 2021-8-15 11:29放代码
import urllib.request
import urllib.parse
url = "https://fanyi.youdao.com/translate?smartresult=dict&smartresult=rule"
data = {}
data["i"]:" i love fishc.com"
data["from"]: "AUTO"
data["to"]: "AUTO"
data["smartresult"]: "dict"
data["client"]: "fanyideskweb"
data["salt"]: "16289478560301"
data["sign"]: "50f56d58392d0eeb8554f16b02746b3d"
data["lts"]: "1628947856030"
data["bv"]: "eda468fc64295ecf2810ab8a672c2db1"
data["doctype"]: "json"
data["version"]: "2.1"
data["keyfrom"]: "fanyi.web"
data["action"]: "FY_BY_CLICKBUTTION"
data = urllib.parse.urlencode(data).encode("utf-8")
response = urllib.request.urlopen(url,data)
html = response.read().decode("utf-8")
print(html)
本帖最后由 白two 于 2021-8-15 16:26 编辑
没有加请求头,ua和cookie 白two 发表于 2021-8-15 16:25
没有加请求头和cookies
谢谢大佬解答,我去查阅一下相关资料 白two 发表于 2021-8-15 16:25
没有加请求头,ua和cookie
还是不行呢
from urllib import request
import urllib.parse
url = "https://fanyi.youdao.com/translate?smartresult=dict&smartresult=rule"
header = {}
data = {}
header["User-Agent"]="Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/92.0.4515.131 Safari/537.36"
header["Cookie"]="OUTFOX_SEARCH_USER_ID=-1404491642@10.108.160.101; OUTFOX_SEARCH_USER_ID_NCOO=847436334.4984028; JSESSIONID=aaaai22Oo32k9_PctWhTx; ___rl__test__cookies=1629016383535"
data["i"]=" i love fishc.com"
data["from"]= "AUTO"
data["to"]= "AUTO"
data["smartresult"]= "dict"
data["client"]= "fanyideskweb"
data["salt"]= "16289478560301"
data["sign"]= "50f56d58392d0eeb8554f16b02746b3d"
data["lts"]= "1628947856030"
data["bv"]= "eda468fc64295ecf2810ab8a672c2db1"
data["doctype"]= "json"
data["version"]= "2.1"
data["keyfrom"]= "fanyi.web"
data["action"]= "FY_BY_CLICKBUTTION"
data = urllib.parse.urlencode(data).encode("utf-8")
response = request.Request(url,headers = header)
res = request.urlopen(response)
html = res.read().decode("utf-8")
print(html)
from urllib import request
import urllib.parse
url = "https://fanyi.youdao.com/translate?smartresult=dict&smartresult=rule"
header = {}
data = {}
header[
"User-Agent"] = "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/92.0.4515.131 Safari/537.36"
header[
"Cookie"] = "OUTFOX_SEARCH_USER_ID=-1404491642@10.108.160.101; OUTFOX_SEARCH_USER_ID_NCOO=847436334.4984028; JSESSIONID=aaaai22Oo32k9_PctWhTx; ___rl__test__cookies=1629016383535"
data["i"] = " i love fishc.com"
data["from"] = "AUTO"
data["to"] = "AUTO"
data["smartresult"] = "dict"
data["client"] = "fanyideskweb"
data["salt"] = "16289478560301"
data["sign"] = "50f56d58392d0eeb8554f16b02746b3d"
data["lts"] = "1628947856030"
data["bv"] = "eda468fc64295ecf2810ab8a672c2db1"
data["doctype"] = "json"
data["version"] = "2.1"
data["keyfrom"] = "fanyi.web"
data["action"] = "FY_BY_CLICKBUTTION"
data = urllib.parse.urlencode(data).encode("utf-8")
response = request.Request(url,data, headers=header)
res = request.urlopen(response)
html = res.read().decode("utf-8")
print(html)
emmm,第29行,你data根本没提交
页:
[1]