IP | Country | PORT | ADDED |
---|---|---|---|
68.71.247.130 | 4145 | 30 minutes ago | |
68.71.254.6 | 4145 | 30 minutes ago | |
72.195.114.184 | us | 4145 | 31 minutes ago |
103.216.49.233 | kh | 8080 | 31 minutes ago |
128.140.113.110 | de | 4145 | 31 minutes ago |
101.71.143.237 | cn | 8092 | 31 minutes ago |
50.55.52.50 | us | 80 | 31 minutes ago |
221.231.13.198 | cn | 1080 | 31 minutes ago |
203.95.199.159 | kh | 8080 | 31 minutes ago |
98.152.200.61 | us | 8081 | 31 minutes ago |
161.35.70.249 | de | 3128 | 31 minutes ago |
183.247.199.51 | cn | 30001 | 31 minutes ago |
49.207.36.81 | in | 80 | 31 minutes ago |
67.201.33.10 | us | 25283 | 31 minutes ago |
72.205.0.93 | us | 4145 | 31 minutes ago |
101.71.72.253 | cn | 52300 | 31 minutes ago |
209.97.150.167 | us | 3128 | 31 minutes ago |
70.166.167.55 | us | 57745 | 31 minutes ago |
178.128.86.216 | sg | 50001 | 31 minutes ago |
209.141.45.119 | us | 56666 | 31 minutes ago |
Our proxies work perfectly with all popular tools for web scraping, automation, and anti-detect browsers. Load your proxies into your favorite software or use them in your scripts in just seconds:
Connection formats you know and trust: IP:port or IP:port@login:password.
Any programming language: Python, JavaScript, PHP, Java, and more.
Top automation and scraping tools: Scrapy, Selenium, Puppeteer, ZennoPoster, BAS, and many others.
Anti-detect browsers: Multilogin, GoLogin, Dolphin, AdsPower, and other popular solutions.
Looking for full automation and proxy management?
Take advantage of our user-friendly PapaProxy API: purchase proxies, renew plans, update IP lists, manage IP bindings, and export ready-to-use lists — all in just a few clicks, no hassle.
PapaProxy offers the simplicity and flexibility that both beginners and experienced developers will appreciate.
And 500+ more tools and coding languages to explore
The easiest option is to use ready-made online proxy checkers. For example, Hidemy.name, which shows the type of protocol used. Or you can simply run Speedtest - this will show you the bandwidth and response speed (ping).
In PHP, you can generate JSON data using the json_encode function, and in Swift (iOS/macOS), you can parse it using JSONSerialization or Codable depending on your needs.
Here's an example of generating JSON in PHP and parsing it using NSJSONSerialization in Swift
PHP (Generate JSON):
'John Doe',
'age' => 25,
'city' => 'New York',
'is_student' => true
);
// Encode data to JSON
$jsonData = json_encode($data);
// Output JSON
echo $jsonData;
?>
In this PHP script, the json_encode function is used to convert the PHP associative array into a JSON string.
Swift (Parse JSON using NSJSONSerialization):
import Foundation
// Sample JSON data as a string
let jsonString = """
{
"name": "John Doe",
"age": 25,
"city": "New York",
"is_student": true
}
"""
// Convert JSON string to Data
if let jsonData = jsonString.data(using: .utf8) {
do {
// Parse JSON data using NSJSONSerialization
if let jsonObject = try JSONSerialization.jsonObject(with: jsonData, options: []) as? [String: Any] {
// Access parsed JSON data
let name = jsonObject["name"] as? String ?? ""
let age = jsonObject["age"] as? Int ?? 0
let city = jsonObject["city"] as? String ?? ""
let isStudent = jsonObject["is_student"] as? Bool ?? false
// Print parsed data
print("Name: \(name)")
print("Age: \(age)")
print("City: \(city)")
print("Is Student: \(isStudent)")
}
} catch {
print("Error parsing JSON: \(error.localizedDescription)")
}
}
In this Swift code, the JSONSerialization class is used to parse the JSON string (converted to Data) into a Swift dictionary ([String: Any]). You can then access individual values from the parsed JSON data.
Note: Ensure that the JSON structure in your PHP script and Swift code aligns, and handle errors appropriately during parsing. Additionally, consider using Codable in Swift for a more convenient way to work with JSON data if your data structure matches your Swift model.
To scrape all HTML content from a website using Scrapy, you need to create a spider that visits each page of the website and extracts the HTML content. Here's a simple example:
Create a Scrapy Project:
If you haven't already, create a Scrapy project by running the following commands in your terminal or command prompt:
scrapy startproject myproject
cd myproject
Define a Spider:
Open the spiders directory in your project and create a spider (e.g., html_spider.py). Edit the spider file with the following content:
import scrapy
class HtmlSpider(scrapy.Spider):
name = 'html_spider'
start_urls = ['http://example.com'] # Start with the main page of the website
def parse(self, response):
# Extract HTML content and yield it
html_content = response.text
yield {
'url': response.url,
'html_content': html_content
}
# Follow links to other pages (if needed)
for next_page_url in response.css('a::attr(href)').extract():
yield scrapy.Request(url=next_page_url, callback=self.parse)
This spider, named html_spider, starts with the main page (start_urls) and extracts the HTML content. It then follows links (a::attr(href)) to other pages and extracts their HTML content as well.
Run the Spider:
Run your spider using the following command:
scrapy crawl html_spider -o output.json
This command will execute the html_spider and save the output in a JSON file named output.json. Each item in the JSON file will contain the URL and HTML content of a page.
In the messenger settings, go to "Data and Drive". Click on "Proxy settings", and then, enabling the "Use proxy settings" tab, enter the server, port, username and password in the specially highlighted fields. If you are going to make settings in the Desktop version, you will need to go to the menu. There, in the "Connection method" item, click on "TSP via Socks5" and enter the required data.
In the "System Settings" section, open the "Network" tab, and then, when you highlight the active connection, click "Advanced". Here, in the "Proxies" tab, tick only the HTTP proxy if you do not intend to use other types of proxies temporarily. Enter the address of your proxy server and its port in the designated fields and click "OK".
What else…