香港云主机最佳企业级服务商!

ADSL拨号VPS包含了中国大陆(联通,移动,电信,)

中国香港,国外拨号VPS。

当前位置:云主机 > python >

电信ADSL拨号VPS
联通ADSL拨号VPS
移动ADSL拨号VPS

python抓取搜狗微信公众号文章


时间:2022-04-02 10:26 作者:admin610456


初学python/' target='_blank'>python,抓取搜狗微信公众号文章存入mysql/' target='_blank'>mysql

MySQL表:

代码:

import requestsimport jsonimport reimport pymysql # 创建连接conn = pymysql.connect(host='你的数据库地址', port=端口, user='用户名', passwd='密码', db='数据库名称', charset='utf8')# 创建游标cursor = conn.cursor()cursor.execute("select * from hd_gzh")effect_row = cursor.fetchall()from bs4 import BeautifulSoupsocket.setdefaulttimeout(60)count = 1headers = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:65.0) Gecko/20100101 Firefox/65.0'}#阿布云ip代理暂时不用# proxyHost = "http-cla.abuyun.com"# proxyPort = "9030"# # 代理隧道验证信息# proxyUser = "H56761606429T7UC"# proxyPass = "9168EB00C4167176"# proxyMeta = "http://%(user)s:%(pass)s@%(host)s:%(port)s" % {#  "host" : proxyHost,#  "port" : proxyPort,#  "user" : proxyUser,#  "pass" : proxyPass,# }# proxies = {#   "http" : proxyMeta,#   "https" : proxyMeta,# }#查看是否已存在数据def checkData(name):  sql = "select * from gzh_article where title = '%s'"  data = (name,)  count = cursor.execute(sql % data)  conn.commit()  if(count!=0):    return False  else:    return True#插入数据def insertData(title,picture,author,content):  sql = "insert into gzh_article (title,picture,author,content) values ('%s', '%s','%s', '%s')"  data = (title,picture,author,content)  cursor.execute(sql % data)  conn.commit()  print("插入一条数据")  return  for row in effect_row:  newsurl = 'https://weixin.sogou.com/weixin?type=1&s_from=input&query=' + row[1] + '&ie=utf8&_sug_=n&_sug_type_='  res = requests.get(newsurl,headers=headers)  res.encoding = 'utf-8'  soup = BeautifulSoup(res.text,'html.parser')  url = 'https://weixin.sogou.com' + soup.select('.tit a')[0]['href']  res2 = requests.get(url,headers=headers)  res2.encoding = 'utf-8'  soup2 = BeautifulSoup(res2.text,'html.parser')  pattern = re.compile(r"url \+= '(.*?)';", re.MULTILINE | re.DOTALL)  script = soup2.find("script")  url2 = pattern.search(script.text).group(1)  res3 = requests.get(url2,headers=headers)  res3.encoding = 'utf-8'  soup3 = BeautifulSoup(res3.text,'html.parser')  print()  pattern2 = re.compile(r"var msgList = (.*?);$", re.MULTILINE | re.DOTALL)  script2 = soup3.find("script", text=pattern2)  s2 = json.loads(pattern2.search(script2.text).group(1))  #等待10s  time.sleep(10)    for news in s2["list"]:    articleurl = "https://mp.weixin.qq.com"+news["app_msg_ext_info"]["content_url"]    articleurl = articleurl.replace('&','&')    res4 = requests.get(articleurl,headers=headers)    res4.encoding = 'utf-8'    soup4 = BeautifulSoup(res4.text,'html.parser')    if(checkData(news["app_msg_ext_info"]["title"])):      insertData(news["app_msg_ext_info"]["title"],news["app_msg_ext_info"]["cover"],news["app_msg_ext_info"]["author"],pymysql.escape_string(str(soup4)))    count += 1    #等待5s    time.sleep(10)    for news2 in news["app_msg_ext_info"]["multi_app_msg_item_list"]:      articleurl2 = "https://mp.weixin.qq.com"+news2["content_url"]      articleurl2 = articleurl2.replace('&','&')      res5 = requests.get(articleurl2,headers=headers)      res5.encoding = 'utf-8'      soup5 = BeautifulSoup(res5.text,'html.parser')      if(checkData(news2["title"])):        insertData(news2["title"],news2["cover"],news2["author"],pymysql.escape_string(str(soup5)))      count += 1      #等待10s      time.sleep(10)cursor.close()conn.close()print("操作完成")

以上就是本文的全部内容,希望对大家的学习有所帮助,也希望大家多多支持脚本之家。

(责任编辑:admin)






帮助中心
会员注册
找回密码
新闻中心
快捷通道
域名登录面板
虚机登录面板
云主机登录面板
关于我们
关于我们
联系我们
联系方式

售前咨询:17830004266(重庆移动)

企业QQ:383546523

《中华人民共和国工业和信息化部》 编号:ICP备00012341号

Copyright © 2002 -2018 香港云主机 版权所有
声明:香港云主机品牌标志、品牌吉祥物均已注册商标,版权所有,窃用必究

云官方微信

在线客服

  • 企业QQ: 点击这里给我发消息
  • 技术支持:383546523

  • 公司总台电话:17830004266(重庆移动)
  • 售前咨询热线:17830004266(重庆移动)