生活随笔
收集整理的這篇文章主要介紹了
python Chrome + selenium自动化测试与python爬虫获取网页数据
小編覺得挺不錯(cuò)的,現(xiàn)在分享給大家,幫大家做個(gè)參考.
一、使用Python+selenium+Chrome 報(bào)錯(cuò): selenium.common.exceptions.SessionNotCreatedException: Message: session not created: This version of ChromeDriver only supports Chrome version 78 說明,這個(gè)chrom驅(qū)動(dòng)支持78版本
谷歌瀏覽器版本
進(jìn)入驅(qū)動(dòng)網(wǎng)址:http://npm.taobao.org/mirrors/chromedriver/,下載對(duì)應(yīng)版本的谷歌驅(qū)動(dòng) 下載后放在與python解析器同一文件夾下面 重新運(yùn)行程序,運(yùn)行成功!!!
二、實(shí)例測試: 使用python selenium自動(dòng)化測試模塊結(jié)合python爬蟲獲取網(wǎng)頁數(shù)據(jù)。
使用谷歌瀏覽器模擬打開https://www.hao123.com/網(wǎng)頁,瀏覽器點(diǎn)擊百度銜接,模擬輸入python進(jìn)行百度搜索,在關(guān)閉子網(wǎng)頁,最后在https://www.hao123.com/網(wǎng)頁獲取精選新聞信息。
1.python控制谷歌瀏覽器 main.py
"""=== coding: UTF8 ==="""
from selenium
. webdriver
import Chrome
from selenium
. webdriver
. common
. keys
import Keys
import time
"""
========================================
主函數(shù)功能測試
========================================
"""
if __name__
== '__main__' : web
= Chrome
( ) web
. get
( "https://www.hao123.com/" ) web
. find_element_by_xpath
( '//*[@id="userCommonSites"]/ul/li[1]/div/a' ) . click
( ) time
. sleep
( 1 ) web
. switch_to
. window
( web
. window_handles
[ - 1 ] ) web
. find_element_by_xpath
( '//*[@id="kw"]' ) . send_keys
( "python" , Keys
. ENTER
) time
. sleep
( 1 ) web
. close
( ) web
. switch_to
. window
( web
. window_handles
[ - 1 ] ) a_list
= web
. find_elements_by_xpath
( '//*[@id="topzixun-over"]/div/div[2]/p' ) for a
in a_list
: print ( a
. find_element_by_xpath
( './a[1]' ) . text
) web
. close
( )
2.python使谷歌瀏覽器在后臺(tái)運(yùn)行,爬取數(shù)據(jù)(即無頭瀏覽器) main.py
"""=== coding: UTF8 ==="""
from selenium
. webdriver
import Chrome
from selenium
. webdriver
. common
. keys
import Keys
from selenium
. webdriver
. chrome
. options
import Options
import time
"""
========================================
主函數(shù)功能測試
========================================
"""
if __name__
== '__main__' : opt
= Options
( ) opt
. add_argument
( "--headless" ) opt
. add_argument
( "--disbale-gpu" ) web
= Chrome
( options
= opt
) web
. get
( "https://www.hao123.com/" ) web
. find_element_by_xpath
( '//*[@id="userCommonSites"]/ul/li[1]/div/a' ) . click
( ) time
. sleep
( 1 ) web
. switch_to
. window
( web
. window_handles
[ - 1 ] ) web
. find_element_by_xpath
( '//*[@id="kw"]' ) . send_keys
( "python" , Keys
. ENTER
) time
. sleep
( 1 ) web
. close
( ) web
. switch_to
. window
( web
. window_handles
[ - 1 ] ) a_list
= web
. find_elements_by_xpath
( '//*[@id="topzixun-over"]/div/div[2]/p' ) for a
in a_list
: print ( a
. find_element_by_xpath
( './a[1]' ) . text
) web
. close
( )
關(guān)注公眾號(hào),獲取更多資料
總結(jié)
以上是生活随笔 為你收集整理的python Chrome + selenium自动化测试与python爬虫获取网页数据 的全部內(nèi)容,希望文章能夠幫你解決所遇到的問題。
如果覺得生活随笔 網(wǎng)站內(nèi)容還不錯(cuò),歡迎將生活随笔 推薦給好友。