爬虫,处理好了防盗链还是出现文章已下线

#在学习到防盗链的某案例以后想在教学基础上改进到多视频爬取
#在爬取娱乐版时发现好像是我现在处理不了的动态加载啥的
#然后我就想先把源码里有的14个视频爬出来,但是处理了防盗链后输出的json还是出现已下线
泪了,有哥可以帮忙看看吗

img

import requests
import re
url = "https://www.pearvideo.com/category_4"

headers = {
    "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/98.0.4758.102 Safari/537.36",
}

resp = requests.get(url,headers=headers)

obj = re.compile('.*?<a href="video_(?P<id>.*?)" class="vervideo-lilink actplay">',re.S)

url_ids = obj.finditer(resp.text)

id_lists = []
for url_id in url_ids:
        group = url_id.groupdict()
        groupstr = str(group)
        newgroup = eval(groupstr)
        id_lists.append(newgroup['id'])
        
a=2
for id_list in id_lists:
    url = f"https://www.pearvideo.com/video_{id_list[a]}"    
    
    headers = {
        "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/98.0.4758.102 Safari/537.36",
        #防盗链处理,在看到cookie的下面
        "Referer": url
    }

    #抓包工具里的用来发送请求的地址

    video_url = f"https://www.pearvideo.com/videoStatus.jsp?contId={id_list[a]}&mrd=0.6826852473664105"
    resp = requests.get(video_url, headers=headers)

    #想要获取的下载地址是preview里的json格式
    dic = resp.json()
    print(dic)
    #已json格式查找要寻找的视频下载地址,以及需要替换的东西
    srcUrl = dic['videoInfo']['videos']['srcUrl']
    systemTime = dic['systemTime']
    #对找到的视频地址进行替换加工,得到真正的下载地址
    video_download_url = srcUrl.replace(systemTime,f"cont-{id_list[a]}")

    with open (f"{id_list[a]}梨视频下载视频.mp4",mode="wb") as f:
        f.write(requests.get(video_download_url).content)
        print(id_list[a],"over")
    a+=1