通过url获取网页源码,我们一般分为这几种情况
网页是静态网页,直接使用
- public static string HttpPost(string url, string paraJsonStr)
- {
- WebClient webClient = new WebClient();
- webClient.Headers.Add("Content-Type", "application/x-www-form-urlencoded");
- byte[] postData = System.Text.Encoding.UTF8.GetBytes(paraJsonStr);
- byte[] responseData = webClient.UploadData(url, "POST", postData);
- string returnStr = System.Text.Encoding.UTF8.GetString(responseData);
- return returnStr;
-
- }
就可以获取网页页面的源码
Request unsuccessful. Incapsula incident ID: 89300071...
这种情况通常是因为网站有反爬虫的处理
解决办法:
2.1.请求添加Headers里面的信息

通常需要添加 “User-Agent”的值,不行的话,可以多添加几个header的属性和 值进去
常用的 “User-Agent”如下:
"Mozilla/5.0 (compatible; MSIE 9.0; Windows NT 6.1; Win64; x64; Trident/5.0; .NET CLR 3.5.30729; .NET CLR 3.0.30729; .NET CLR 2.0.50727; Media Center PC 6.0)",
"Mozilla/5.0 (compatible; MSIE 8.0; Windows NT 6.0; Trident/4.0; WOW64; Trident/4.0; SLCC2; .NET CLR 2.0.50727; .NET CLR 3.5.30729; .NET CLR 3.0.30729; .NET CLR 1.0.3705; .NET CLR 1.1.4322)",
"Mozilla/4.0 (compatible; MSIE 7.0b; Windows NT 5.2; .NET CLR 1.1.4322; .NET CLR 2.0.50727; InfoPath.2; .NET CLR 3.0.04506.30)",
"Mozilla/5.0 (Windows; U; Windows NT 5.1; zh-CN) AppleWebKit/523.15 (KHTML, like Gecko, Safari/419.3) Arora/0.3 (Change: 287 c9dfb30)",
"Mozilla/5.0 (X11; U; Linux; en-US) AppleWebKit/527+ (KHTML, like Gecko, Safari/419.3) Arora/0.6",
"Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.8.1.2pre) Gecko/20070215 K-Ninja/2.1.1",
"Mozilla/5.0 (X11; Linux i686; U;) Gecko/20070322 Kazehakase/0.4.5",
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/101.0.4951.54 Safari/537.36 Edg/101.0.1210.39",
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/104.0.5112.81 Safari/537.36 Edg/104.0.1293.54"
例如:
- public static string HttpPost(string url, string paraJsonStr = "")
- {
- HttpWebRequest request = (HttpWebRequest)HttpWebRequest.Create(url);
- request.Headers["User-Agent"] = "Mozilla/5.0 (Linux; Android 6.0; Nexus 5 Build/MRA58N) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/104.0.5112.81 Mobile Safari/537.36 Edg/104.0.1293.54";
- request.Headers["Cookie"] = "visid_incap_161858=BlNpjA+qS9ucNza1";
- HttpWebResponse response = (HttpWebResponse)request.GetResponse();
- Stream responseStream = response.GetResponseStream();
- StreamReader streamReader = new StreamReader(responseStream, Encoding.UTF8);
- string str = streamReader.ReadToEnd();
- return str;
- }
2.2.在访问页面时加上时间间隔
在Response中无法获取到完整的页面源代码,跟element的元素不同。这种情况使用Selenium插件的办法
- import re
- import time
- import json
- import random
- from turtle import pd
- from xml.dom.minidom import Element
- import requests
- import csv
- from selenium import webdriver
- # from selenium.webdriver.chrome.options import Options
- from selenium.webdriver.edge.options import Options
- from time import strftime, localtime
-
- #path to webdriver
- //需要下载浏览器的driver,括号里面式msedgedriver.exe的下载地址
- driver = webdriver.Edge(executable_path = "C:\\...\\msedgedriver.exe")
- //加载时窗口的大小
- driver.maximize_window()
- //url时加载时的网址,获取信息的具体网页的地址
- driver.get(url)
- #The sleep time is the time for the web page to dynamically obtain information. Depending on the speed of the network, different adjustments can be made.
- time.sleep(10)
- //网页自动往下滑动的大小
- driver.execute_script("window.scrollBy(0,{})".format(1500))
- time.sleep(2)
- #The page has a button, click to get the information of the whole page
-
- //具体网页button的xpath路径
- if(driver.find_element_by_xpath("//div[@class='articlePage_button']/span[@class='buttonText']")):
- driver.find_element_by_xpath("//div[@class='articlePage_button']/span[@class='buttonText']").click()
- time.sleep(5)
-
- #Store data to file
- f=open("C:\\...\\a.csv","a+",encoding='utf-8',newline="")
- a_writer=csv.writer(f)
-
- #parsing html information
- aText=driver.find_element_by_xpath("//h1[@class='pageTitle']").text
- print("aText:",aText)
- a_writer.writerow(["name",aText])
- #输出空行需要这样
- a_writer.writerow(["",""])
-
-
- driver.quit()
前面两种办法,都可以通过Response中查看到完整的页面代码
