我正在制作一个网络爬虫,当我尝试抓取一页数据时,它会不断加载相同的信息。
from urllib.request import urlopen as uReq
from bs4 import BeautifulSoup as soup
my_url = 'https://www.realtor.com/realestateagents/phoenix_az'
#opening up connection, grabbing the page
uClient = uReq(my_url)
#read page
page_html = uClient.read()
#close page
uClient.close()
#html parsing
page_soup = soup(page_html, "html.parser")
#finds all realtors on page
containers = page_soup.findAll("div",{"class":"agent-list-card clearfix"})
for container in containers:
name = page_soup.find('div', class_='agent-name text-bold')
agent_name = name.text.strip()
number = page_soup.find('div', class_='agent-phone hidden-xs hidden-xxs')
agent_number = number.text.strip()
print("name: " + agent_name)
print("number: " + agent_number)
红颜莎娜
相关分类