這篇文章將為大家詳細(xì)講解有關(guān)Python如何爬取Json數(shù)據(jù),小編覺(jué)得挺實(shí)用的,因此分享給大家做個(gè)參考,希望大家閱讀完這篇文章后可以有所收獲。
十多年的巨野網(wǎng)站建設(shè)經(jīng)驗(yàn),針對(duì)設(shè)計(jì)、前端、開(kāi)發(fā)、售后、文案、推廣等六對(duì)一服務(wù),響應(yīng)快,48小時(shí)及時(shí)工作處理。全網(wǎng)整合營(yíng)銷推廣的優(yōu)勢(shì)是能夠根據(jù)用戶設(shè)備顯示端的尺寸不同,自動(dòng)調(diào)整巨野建站的顯示方式,使網(wǎng)站能夠適用不同顯示終端,在瀏覽器中調(diào)整網(wǎng)站的寬度,無(wú)論在任何一種瀏覽器上瀏覽網(wǎng)站,都能展現(xiàn)優(yōu)雅布局與設(shè)計(jì),從而大程度地提升瀏覽體驗(yàn)。成都創(chuàng)新互聯(lián)從事“巨野網(wǎng)站設(shè)計(jì)”,“巨野網(wǎng)站推廣”以來(lái),每個(gè)客戶項(xiàng)目都認(rèn)真落實(shí)執(zhí)行。
該地址返回的響應(yīng)內(nèi)容為Json類型,其中紅框標(biāo)記的項(xiàng)即為AI流轉(zhuǎn)率值:
實(shí)現(xiàn)代碼如下:
import requests import json import csv # 爬蟲(chóng)地址 url = 'https://databank.yushanfang.com/api/ecapi?path=/databank/crowdFullLink/flowInfo&fromCrowdId=3312&beginTheDate=201810{}&endTheDate=201810{}&toCrowdIdList[0]=3312&toCrowdIdList[1]=3313&toCrowdIdList[2]=3314&toCrowdIdList[3]=3315' # 攜帶cookie進(jìn)行訪問(wèn) headers = { 'Host':'databank.yushanfang.com', 'Referer':'https://databank.yushanfang.com/', 'Connection':'keep-alive', 'User-Agent':'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.84 Safari/537.36', 'Cookie':'_tb_token_=iNkDeJLdM3MgvKjhsfdW; bs_n_lang=zh_CN; cna=aaj1EViI7x0CATo9kTKvjzgS; ck2=072de851f1c02d5c7bac555f64c5c66d; c_token=c74594b486f8de731e2608cb9526a3f2; an=5YWo5qOJ5pe25Luj5a6Y5pa55peX6Iiw5bqXOnpmeA%3D%3D; lg=true; sg=\"=19\"; lvc=sAhojs49PcqHQQ%3D%3D; isg=BPT0Md7dE_ic5Ie3Oa85RxaMxbLK3UqJMMiN6o5VjH8C-ZRDtt7aRXb3fXGEAVAP', } rows = [] for n in range(20, 31): row = [] row.append(n) for m in range (21, 32): if m < n + 1: row.append("") else: # 格式化請(qǐng)求地址,更換請(qǐng)求參數(shù) reqUrl = url.format(n, m) # 打印本次請(qǐng)求地址 print(url) # 發(fā)送請(qǐng)求,獲取響應(yīng)結(jié)果 response = requests.get(url=reqUrl, headers=headers, verify=False) text = response.text # 打印本次請(qǐng)求響應(yīng)內(nèi)容 print(text) # 將響應(yīng)內(nèi)容轉(zhuǎn)換為Json對(duì)象 jsonobj = json.loads(text) # 從Json對(duì)象獲取想要的內(nèi)容 toCntPercent = jsonobj['data']['interCrowdInfo'][1]['toCntPercent'] # 生成行數(shù)據(jù) row.append(str(toCntPercent)+"%") # 保存行數(shù)據(jù) rows.append(row) # 生成Excel表頭 header = ['AI流轉(zhuǎn)率', '21', '22', '23', '24', '25', '26', '27', '28', '29', '30', '31'] # 將表頭數(shù)據(jù)和爬蟲(chóng)數(shù)據(jù)導(dǎo)出到Excel文件 with open('D:\\res\\pachong\\tmall.csv', 'w', encoding='gb18030') as f : f_csv = csv.writer(f) f_csv.writerow(header) f_csv.writerows(rows)
import csv import json import ssl import urllib.request # 爬蟲(chóng)地址 url = 'https://databank.yushanfang.com/api/ecapi?path=/databank/crowdFullLink/flowInfo&fromCrowdId=3312&beginTheDate=201810{}&endTheDate=201810{}&toCrowdIdList[0]=3312&toCrowdIdList[1]=3313&toCrowdIdList[2]=3314&toCrowdIdList[3]=3315' # 不校驗(yàn)證書 ssl._create_default_https_context = ssl._create_unverified_context # 攜帶cookie進(jìn)行訪問(wèn) headers = { 'Host':'databank.yushanfang.com', 'Referer':'https://databank.yushanfang.com/', 'Connection':'keep-alive', 'User-Agent':'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.84 Safari/537.36', 'Cookie':'_tb_token_=iNkDeJLdM3MgvKjhsfdW; bs_n_lang=zh_CN; cna=aaj1EViI7x0CATo9kTKvjzgS; ck2=072de851f1c02d5c7bac555f64c5c66d; c_token=c74594b486f8de731e2608cb9526a3f2; an=5YWo5qOJ5pe25Luj5a6Y5pa55peX6Iiw5bqXOnpmeA%3D%3D; lg=true; sg=\"=19\"; lvc=sAhojs49PcqHQQ%3D%3D; isg=BPT0Md7dE_ic5Ie3Oa85RxaMxbLK3UqJMMiN6o5VjH8C-ZRDtt7aRXb3fXGEAVAP', } rows = [] n = 20 while n <31: row = [] row.append(n) m =21 while m <32: if m < n + 1: row.append("") else: # 格式化請(qǐng)求地址,更換請(qǐng)求參數(shù) reqUrl = url.format(n, m) # 打印本次請(qǐng)求地址 print(reqUrl) # 發(fā)送請(qǐng)求,獲取響應(yīng)結(jié)果 request = urllib.request.Request(url=reqUrl, headers=headers) response = urllib.request.urlopen(request) text = response.read().decode('utf8') # 打印本次請(qǐng)求響應(yīng)內(nèi)容 print(text) # 將響應(yīng)內(nèi)容轉(zhuǎn)換為Json對(duì)象 jsonobj = json.loads(text) # 從Json對(duì)象獲取想要的內(nèi)容 toCntPercent = jsonobj['data']['interCrowdInfo'][1]['toCntPercent'] # 生成行數(shù)據(jù) row.append(str(toCntPercent) + "%") m = m+1 rows.append(row) n = n+1 # 生成Excel表頭 header = ['AI流轉(zhuǎn)率', '21', '22', '23', '24', '25', '26', '27', '28', '29', '30', '31'] # 將表頭數(shù)據(jù)和爬蟲(chóng)數(shù)據(jù)導(dǎo)出到Excel文件 with open('D:\\res\\pachong\\tmall.csv', 'w', encoding='gb18030') as f : f_csv = csv.writer(f) f_csv.writerow(header) f_csv.writerows(rows)
導(dǎo)出內(nèi)容如下:
關(guān)于Python如何爬取Json數(shù)據(jù)就分享到這里了,希望以上內(nèi)容可以對(duì)大家有一定的幫助,可以學(xué)到更多知識(shí)。如果覺(jué)得文章不錯(cuò),可以把它分享出去讓更多的人看到。