IP | Country | PORT | ADDED |
---|---|---|---|
111.177.48.18 | cn | 9501 | 21 minutes ago |
49.207.36.81 | in | 80 | 21 minutes ago |
201.148.32.162 | 80 | 21 minutes ago | |
46.47.197.210 | ru | 3128 | 21 minutes ago |
83.168.72.172 | pl | 8081 | 21 minutes ago |
123.30.154.171 | vn | 7777 | 21 minutes ago |
80.120.49.242 | at | 80 | 21 minutes ago |
213.143.113.82 | at | 80 | 21 minutes ago |
194.158.203.14 | by | 80 | 21 minutes ago |
161.35.70.249 | de | 80 | 21 minutes ago |
203.19.38.114 | cn | 1080 | 21 minutes ago |
119.3.113.152 | cn | 9094 | 21 minutes ago |
190.58.248.86 | tt | 80 | 21 minutes ago |
220.167.89.46 | cn | 1080 | 21 minutes ago |
158.255.77.168 | ae | 80 | 21 minutes ago |
221.231.13.198 | cn | 1080 | 21 minutes ago |
119.3.113.150 | cn | 9094 | 21 minutes ago |
168.196.214.187 | br | 80 | 21 minutes ago |
46.8.48.143 | ru | 1080 | 21 minutes ago |
59.53.80.122 | cn | 10024 | 21 minutes ago |
Our proxies work perfectly with all popular tools for web scraping, automation, and anti-detect browsers. Load your proxies into your favorite software or use them in your scripts in just seconds:
Connection formats you know and trust: IP:port or IP:port@login:password.
Any programming language: Python, JavaScript, PHP, Java, and more.
Top automation and scraping tools: Scrapy, Selenium, Puppeteer, ZennoPoster, BAS, and many others.
Anti-detect browsers: Multilogin, GoLogin, Dolphin, AdsPower, and other popular solutions.
Looking for full automation and proxy management?
Take advantage of our user-friendly PapaProxy API: purchase proxies, renew plans, update IP lists, manage IP bindings, and export ready-to-use lists — all in just a few clicks, no hassle.
PapaProxy offers the simplicity and flexibility that both beginners and experienced developers will appreciate.
And 500+ more tools and coding languages to explore
In JavaScript with Selenium, you can save and reuse cookies using the WebDriver's manage().getCookies() and manage().addCookie() methods. Here's a simple example:
const { Builder } = require('selenium-webdriver');
const firefox = require('selenium-webdriver/firefox');
// Create a new instance of the Firefox driver
const driver = new Builder()
.forBrowser('firefox')
.setFirefoxOptions(new firefox.Options().headless())
.build();
// Navigate to a webpage
async function navigateToPage() {
await driver.get('https://example.com');
}
// Save cookies
async function saveCookies() {
const cookies = await driver.manage().getCookies();
// Save the cookies to a file or some storage mechanism
// For simplicity, we'll just print them here
console.log('Cookies:', cookies);
}
// Reuse cookies
async function reuseCookies(savedCookies) {
// Delete existing cookies
await driver.manage().deleteAllCookies();
// Add the saved cookies to the browser session
for (const cookie of savedCookies) {
await driver.manage().addCookie(cookie);
}
// Navigate to a page to apply the cookies
await navigateToPage();
}
// Example usage
(async () => {
await navigateToPage(); // Navigate to the page and set some initial cookies
await saveCookies(); // Save the cookies
// Close and reopen the browser or navigate to a different page
// ...
// Reuse the saved cookies
await reuseCookies(savedCookies);
})();
The navigateToPage function navigates to a webpage and sets some initial cookies.
The saveCookies function retrieves the current cookies using manage().getCookies() and prints them. You would typically save them to a file or some storage mechanism.
The reuseCookies function deletes existing cookies, then adds the saved cookies back to the browser session using manage().addCookie(). It then navigates to a page to apply the cookies.
The example usage section demonstrates how to use these functions in a sequence.
If you can't proxy requests in Scrapy:
- Verify correct proxy configuration in Scrapy settings.
- Confirm proxy functionality with external tools.
- Check for typos or errors in your code and settings.
- Ensure proxy authentication details are correct.
- Test with a direct internet connection to isolate the issue.
- Check for IP blocking by the target website.
- Confirm proper configuration of the HttpProxyMiddleware.
- Use Scrapy logging to inspect requests and responses.
- Ensure your proxy supports HTTPS if needed.
- Test with a single, static proxy for simplicity.
- Keep Scrapy and dependencies up to date.
- Consider using middleware libraries like scrapy-rotating-proxies.
We recommend using SOCKS5 proxies for uTorrent. When using HTTP, HTTPS, and SOCKS4 protocols, users often encounter technical problems when downloading files. They may simply not be loaded on the device. It is also worth noting that SOCKS5 is the best anonymizer, which hides all the data of the computer.
Before choosing a proxy server provider, it is recommended to pay attention to the parameter "traffic limit". If there is one, money will be deducted from your account. To avoid loss of money, it is better to choose a vendor who has to pay not for traffic, but for the number of addresses.
The messenger has bots that allow you to get a free proxy - @socks5_bot. First, run the bot and select a location to connect to get the IP address, username, password and port. Now, in "Settings", find the "Data and Disk" section. There, under "Proxy Settings" enter the server, port, password and login. If the setting will be made in the "Desktop", in the menu will have to find "Connection method", select "TGP through Socks5" and enter similar data.
VPN allows you to hide your real IP address, as well as further encrypt your traffic. VPN is also actively used for address spoofing. For example, the user is in the Russian Federation, but by connecting through a VPN server, the site "thinks" that the user is from the United States.
What else…