脚本之家,脚本语言编程技术及教程分享平台!
分类导航

Python|VBS|Ruby|Lua|perl|VBA|Golang|PowerShell|Erlang|autoit|Dos|bat|

服务器之家 - 脚本之家 - Python - Python实战快速上手BeautifulSoup库爬取专栏标题和地址

Python实战快速上手BeautifulSoup库爬取专栏标题和地址

2022-02-13 14:06小旺不正经 Python

BeautifulSoup是爬虫必学的技能,BeautifulSoup最主要的功能是从网页抓取数据,Beautiful Soup自动将输入文档转换为Unicode编码,输出文档转换为utf-8编码

BeautifulSoup库快速上手

 

安装

pip install beautifulsoup4
# 上面的安装失败使用下面的 使用镜像
pip install beautifulsoup4 -i https://pypi.tuna.tsinghua.edu.cn/simple

使用PyCharm的命令行

Python实战快速上手BeautifulSoup库爬取专栏标题和地址

 

解析标签

from bs4 import BeautifulSoup
import requests
url="https://blog.csdn.net/weixin_42403632/category_11076268.html"
headers={"User-Agent":"Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:93.0) Gecko/20100101 Firefox/93.0"}
html=requests.get(url,headers=headers).text
s=BeautifulSoup(html,"html.parser")
title =s.select("h2")
for i in title:
    print(i.text)

第一行代码:导入BeautifulSoup库

第二行代码:导入requests

第三、四、五行代码:获取url的html

第六行代码:激活BeautifulSoup库 "html.parser"设置解析器为HTML解析器

第七行代码:选取所有<h2>标签

Python实战快速上手BeautifulSoup库爬取专栏标题和地址

 

解析属性

BeautifulSoup库 支持根据特定属性解析网页元素

 

根据class值解析

from bs4 import BeautifulSoup
import requests
url="https://blog.csdn.net/weixin_42403632/category_11076268.html"
headers={"User-Agent":"Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:93.0) Gecko/20100101 Firefox/93.0"}
html=requests.get(url,headers=headers).text
s=BeautifulSoup(html,"html.parser")
title =s.select(".column_article_title")
for i in title:
    print(i.text)

Python实战快速上手BeautifulSoup库爬取专栏标题和地址

 

根据ID解析

from bs4 import BeautifulSoup
html="""<div class="crop-img-before">
         <img src=""      </div>
        <div id="title">
        测试成功
        </div>
      <div class="crop-zoom">
         <a href="javascript:;" rel="external nofollow"  rel="external nofollow"  rel="external nofollow"  rel="external nofollow"  class="bt-reduce">-</a><a href="javascript:;" rel="external nofollow"  rel="external nofollow"  rel="external nofollow"  rel="external nofollow"  class="bt-add">+</a>
      </div>
      <div class="crop-img-after">
         <div  class="final-img"></div>
      </div>"""
s=BeautifulSoup(html,"html.parser")
title =s.select("#title")
for i in title:
    print(i.text)

Python实战快速上手BeautifulSoup库爬取专栏标题和地址

 

多层筛选

from bs4 import BeautifulSoup
html="""<div class="crop-img-before">
         <img src=""      </div>
        <div id="title">
        456456465
        <h1>测试成功</h1>
        </div>
      <div class="crop-zoom">
         <a href="javascript:;" rel="external nofollow"  rel="external nofollow"  rel="external nofollow"  rel="external nofollow"  class="bt-reduce">-</a><a href="javascript:;" rel="external nofollow"  rel="external nofollow"  rel="external nofollow"  rel="external nofollow"  class="bt-add">+</a>
      </div>
      <div class="crop-img-after">
         <div  class="final-img"></div>
      </div>"""
s=BeautifulSoup(html,"html.parser")
title =s.select("#title")
for i in title:
    print(i.text)
title =s.select("#title h1")
for i in title:
    print(i.text)

 

提取a标签中的网址

title =s.select("a")
for i in title:
    print(i["href"])

Python实战快速上手BeautifulSoup库爬取专栏标题和地址

 

实战-获取博客专栏 标题+网址

Python实战快速上手BeautifulSoup库爬取专栏标题和地址

from bs4 import BeautifulSoup
import requests
import re
url="https://blog.csdn.net/weixin_42403632/category_11298953.html"
headers={"User-Agent":"Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:93.0) Gecko/20100101 Firefox/93.0"}
html=requests.get(url,headers=headers).text
s=BeautifulSoup(html,"html.parser")
title =s.select(".column_article_list li a")
for i in title:
    print((re.findall("原创.*?
(.*?)
",i.text))[0].lstrip())
    print(i["href"])

Python实战快速上手BeautifulSoup库爬取专栏标题和地址

到此这篇关于Python实战快速上手BeautifulSoup库爬取专栏标题和地址的文章就介绍到这了,更多相关Python BeautifulSoup库内容请搜索服务器之家以前的文章或继续浏览下面的相关文章希望大家以后多多支持服务器之家!

原文链接:https://blog.csdn.net/weixin_42403632/article/details/120852131

延伸 · 阅读

精彩推荐