如何获取每个URL文本值,我只获取最后一个URL值

2024-09-24 22:23:04 发布

您现在位置:Python中文网/ 问答频道 /正文

我是Python新手。通过这段代码,我只需要最后一个URL值,但我希望每个URL都能抓取内容

contents = []
with open('c:\\users\\thegl\\documents\\datab.csv','r') as csvf: # Open file in read mode
    urls = csv.reader(csvf)
    for url in urls:
        contents.append(url) # Add each url to list contents

for url in contents:  # Parse through each url in the list.
    page = urlopen(url[0]).read()
    soup = BeautifulSoup(page, "html.parser")
for List in soup.find_all('ol',class_='breadcrumb'):
    for listext in List.find_all('li'):
        print(listext.text)

file datab.csv contain following ULS: https://www.dumpstool.com/1Y0-371-exam.htmlhttps://www.dumpstool.com/TK0-201-exam.htmlhttps://www.dumpstool.com/C9510-401-exam.html


Tags: csvinhttpscomurlforhtmlwww
1条回答
网友
1楼 · 发布于 2024-09-24 22:23:04

您应该缩进最后一个for,以便对每个url执行它

contents = []
with open('c:\\users\\thegl\\documents\\datab.csv','r') as csvf: # Open file in read mode
    urls = csv.reader(csvf)
    for url in urls:
        contents.append(url) # Add each url to list contents

for url in contents:  # Parse through each url in the list.
    page = urlopen(url[0]).read()
    soup = BeautifulSoup(page, "html.parser")
    for List in soup.find_all('ol',class_='breadcrumb'):
        for listext in List.find_all('li'):
            print(listext.text)

相关问题 更多 >