我正在制作一个网络爬虫,当我尝试抓取一页数据时,它会不断加载相同的信息。from urllib.request import urlopen as uReqfrom bs4 import BeautifulSoup as soup my_url = 'https://www.realtor.com/realestateagents/phoenix_az'#opening up connection, grabbing the pageuClient = uReq(my_url)#read page page_html = uClient.read()#close pageuClient.close()#html parsingpage_soup = soup(page_html, "html.parser")#finds all realtors on page containers = page_soup.findAll("div",{"class":"agent-list-card clearfix"})for container in containers: name = page_soup.find('div', class_='agent-name text-bold') agent_name = name.text.strip() number = page_soup.find('div', class_='agent-phone hidden-xs hidden-xxs') agent_number = number.text.strip() print("name: " + agent_name) print("number: " + agent_number)
添加回答
举报
0/150
提交
取消