python爬虫实战

import requests
import json
yes=input('输入页数:')
yes=int(yes)

headers = {
    "accept": "application/json, text/plain, */*",
    "accept-language": "zh-CN,zh;q=0.9",
    "content-type": "application/json",
    "origin": "https://www.gaokao.cn",
    "referer": "https://www.gaokao.cn/",
    "sec-ch-ua": "\"Google Chrome\";v=\"123\", \"Not:A-Brand\";v=\"8\", \"Chromium\";v=\"123\"",
    "sec-ch-ua-mobile": "?0",
    "sec-ch-ua-platform": "\"Windows\"",
    "sec-fetch-dest": "empty",
    "sec-fetch-mode": "cors",
    "sec-fetch-site": "cross-site",
    "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/123.0.0.0 Safari/537.36"
}
url = "https://api.zjzw.cn/web/api/"
ff=open('x.csv','w')
for paga in range(1,yes):
    params = {
        "big_min": "750",
        "keyword": "",
        "local_batch_id": "",
        "local_province_id": "32",
        "local_type_id": "2073",
        "page": str(paga),
        "province_id": "",
        "size": "20",
        "small_min": "0",
        "type": "",
        "uri": "apidata/api/gk/score/province",
        "year": "2023",
        "zslx": "0",
        "signsafe": "7c6422ab7db981d40fe7115c1b439de2"
    }
    data = {
        "big_min": 750,
        "keyword": "",
        "local_batch_id": "",
        "local_province_id": "32",
        "local_type_id": "2073",
        "page": paga,
        "province_id": "",
        "signsafe": "7c6422ab7db981d40fe7115c1b439de2",
        "size": 20,
        "small_min": 0,
        "type": "",
        "uri": "apidata/api/gk/score/province",
        "year": "2023",
        "zslx": "0"
    }
    data = json.dumps(data, separators=(',', ':'))
    response = requests.post(url, headers=headers, params=params, data=data)
    res=response.json()
    for i in res['data']['item']:
        a=i['city_name']
        b=i['county_name']
        c=i['dual_class_name']
        d=i['local_batch_id']
        e=i['local_batch_name']
        f=i['local_province_name']
        g=i['local_type_name']
        h=i['name']
        j=i['nature_name']
        k=i['province_name']
        l=i['sg_info']
        m=i['zslx_name']
        n=str(i['min'])
        ab=a+b
        sg_name='专业组'+i['sg_name']
        ff.write(','.join([h,a,b,g,e,c,sg_name,l,n,d,f,m,j,k,'\n']))
ff.close()

相关推荐

  1. Python爬虫实战案例

    2024-05-05 02:16:01       63 阅读
  2. Python爬虫实战-1

    2024-05-05 02:16:01       32 阅读
  3. python实现数据爬虫

    2024-05-05 02:16:01       42 阅读
  4. python实现网络爬虫

    2024-05-05 02:16:01       32 阅读

最近更新

  1. docker php8.1+nginx base 镜像 dockerfile 配置

    2024-05-05 02:16:01       94 阅读
  2. Could not load dynamic library ‘cudart64_100.dll‘

    2024-05-05 02:16:01       100 阅读
  3. 在Django里面运行非项目文件

    2024-05-05 02:16:01       82 阅读
  4. Python语言-面向对象

    2024-05-05 02:16:01       91 阅读

热门阅读

  1. 初识Flask

    2024-05-05 02:16:01       27 阅读
  2. 【C++基础】缺省参数

    2024-05-05 02:16:01       29 阅读
  3. MySQL商城数据表(70-79)

    2024-05-05 02:16:01       29 阅读
  4. PCL 点云中的平面点云提取

    2024-05-05 02:16:01       31 阅读
  5. 在C#中使用s7Netplus实现与西门子PLC的数据通信

    2024-05-05 02:16:01       28 阅读
  6. 数据结构-堆

    2024-05-05 02:16:01       30 阅读
  7. 数据存储-SharedPreferences

    2024-05-05 02:16:01       39 阅读
  8. 【C语言】命令行参数

    2024-05-05 02:16:01       34 阅读