为了结果直观,做的简单

网页地址:

​https://search.51job.com/list/180200,000000,0000,00,9,99,%25E6%25AD%25A6%25E6%25B1%2589,2,1.html?lang=c&postchannel=0000&workyear=99&cotype=99&degreefrom=99&jobterm=99&companysize=99&ord_field=0&dibiaoid=0&line=&welfare=​

 

# -*- coding:utf-8 -*-
import requests
from lxml import etree
from wordcloud import WordCloud# 写入csv
def write_csv(name, row):
reload(sys)
sys.setdefaultencoding("utf-8")
# 统计出现频率
def get_count(text):
wordlist_jieba = jieba.cut(text)
# jieba分词
def chinese_jieba(text):
wordlist_jieba = jieba.cut(text)
text_jieba = " ".join(wordlist_jieba)
return text_jieba
# 生成词云图
def get_ciyun(text):
text = chinese_jieba(text)
print(text)
# mask_pic = numpy.array(Image.open(os.path.join(cur_path, "bit.jpg")))
# print(text)
# 背景颜色 词数量 词字体大小 字体文件路径(需要放到和py文件同一个路径下) 去掉的词 遮罩层
font_path = path.join(d, 'fonts', 'Symbola', 'Symbola.ttf')
image.show()
#获取数据
def get_data(url):
result = ''
headers={'Host':'search.51job.com','Upgrade-Insecure-Requests':'1','User-Agent':'Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko)\ Chrome/63.0.3239.132 Safari/537.36'}

return result# 主程序
def main():
url = 'https://search.51job.com/list/180200,000000,0000,00,9,99,%25E6%25AD%25A6%25E6%25B1%2589,2,1.html?lang=c&postchannel=0000&workyear=99&cotype=99°reefrom=99&jobterm=99&companysize=99&ord_field=0&dibiaoid=0&line=&welfare='
text = get_data(url)
#print("获取完毕,分词生成词云")
get_ciyun(text)
get_count(text) main()

网页:

python 爬虫 抓取 前程无忧 51job-代码_爬虫

结果

1 抓取结果

python 爬虫 抓取 前程无忧 51job-代码_html_02

2 词云图

python 爬虫 抓取 前程无忧 51job-代码_词云_03

3 统计词频

python 爬虫 抓取 前程无忧 51job-代码_5e_04