Mobile Menu

Proxy for Sina Weibo

PapaProxy - premium datacenter proxies with the fastest speed. Fully unlimited traffic. Big Papa packages from 100 to 15,000 IP
  • Some of the lowest prices on the market, no hidden fees;
  • Guaranteed refund within 24 hours after payment.
  • All IPv4 proxies with HTTPS and SOCKS5 support;
  • Upgrade IP in a package without extra charges;
  • Fully unlimited traffic included in the price;
  • No KYC for all customers at any stage;
  • Several subnets in each package;
  • Impressive connection speed;
  • And many other benefits :)
Select your tariff
Price for 1 IP-address: 0$
We have over 100,000 addresses on the IPv4 network. All packets need to be bound to the IP address of the equipment you are going to work with. Proxy servers can be used with or without login/password authentication. Just elite and highly private proxies.
Types of proxies

Types of proxies

Datacenter proxies

Starting from
$19 / month
Unlimited Traffic
SOCKS5 Supported
Over 100,000 IPv4 proxies
Packages from 100 proxies
Good discount for wholesale

Private proxies

Starting from
$2,5 / month
Unlimited Traffic
SOCKS5 Supported
Proxies just for you
Speed up to 200 Mbps
For sale from 1 pc.

Rotating proxies

Starting from
$49 / month
Each request is a new IP
SOCKS5 Supported
Automatic rotation
Ideal for API work
All proxies available now

UDP proxies

Starting from
$19 / month
Unlimited traffic
SOCKS5 supported
PremiumFraud Shield
For games and broadcasts
Speed up to 200 Mbps

Try our proxies for free

Get test account for 60 minutes

Register an account and get a proxy for the test. You do not need to fill payment data. Support most of popular tasks: search engines, marketplaces, bulletin boards, online services, etc. tasks
Rectangle Rectangle Rectangle Rectangle
Available regions

Available regions

Accessing Sina Weibo through a proxy allows users outside of China or in restricted networks to connect with one of China’s largest social media platforms. This access is vital for individuals looking to keep up with trends, news, and discussions within Chinese communities or for businesses aiming to engage with the Chinese market. A proxy ensures that users can share thoughts, follow influencers, and participate in social discourse, bridging cultural and geographical gaps, and fostering a deeper understanding of Chinese social dynamics and public opinion.

  • IP updates in the package at no extra charge;

  • Unlimited traffic included in the price;

  • Automatic delivery of addresses after payment;

  • All proxies are IPv4 with HTTPS and SOCKS5 support;

  • Impressive connection speed;

  • Some of the cheapest cost on the market, with no hidden fees;

  • If the IP addresses don't suit you - money back within 24 hours;

  • And many more perks :)

You can buy proxies at cheap pricing and pay by any comfortable method:

  • VISA, MasterCard, UnionPay

  • Tether (TRC20, ERC20)

  • Bitcoin

  • Ethereum

  • AliPay

  • WebMoney WMZ

  • Perfect Money

You can use both HTTPS and SOCKS5 protocols at the same time. Proxies with and without authorization are available in the personal cabinet.

 

Port 8080 for HTTP and HTTPS proxies with authorization.

Port 1080 for SOCKS 4 and SOCKS 5 proxies with authorization.

Port 8085 for HTTP and HTTPS proxies without authorization.

Port 1085 for SOCKS4 and SOCKS5 proxy without authorization.

 

We also have a proxy list builder available - you can upload data in any convenient format. For professional users there is an extended API for your tasks.

Free proxy list

Free Sina Weibo proxy list

Note - these are NOT our test proxies. Publicly available free lists, collected from open sources, to test your software.
You can request a test of our proxies here.
IP
IP
192.252.214.20
IP
80.120.49.242
IP
50.172.75.123
IP
50.239.72.17
IP
50.144.166.226
IP
50.223.239.168
IP
194.26.229.46
IP
5.44.42.115
IP
50.217.226.46
IP
50.168.72.117
IP
50.168.72.122
IP
82.119.96.254
IP
184.178.172.5
IP
213.33.2.28
IP
213.33.2.27
IP
213.16.81.182
IP
72.10.164.178
IP
50.221.74.130
IP
109.232.106.150
IP
190.144.224.182
Country
Country
ca
Country
at
Country
us
Country
us
Country
us
Country
us
Country
ru
Country
ae
Country
us
Country
us
Country
us
Country
sk
Country
us
Country
at
Country
at
Country
hu
Country
ca
Country
us
Country
ru
Country
co
Port
Port
15864
Port
80
Port
80
Port
80
Port
80
Port
80
Port
20016
Port
58386
Port
80
Port
80
Port
80
Port
80
Port
15303
Port
80
Port
80
Port
35559
Port
1929
Port
80
Port
52435
Port
44550
Added

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago

36 minutes ago
Feedback

Feedback

This service impressed with its functionality: excellent speed, many locations, convenient for SEO-analysis of different regions. However, I would like more flexible rates.
Marcel

The support service is prompt and provides timely responses. I have purchased 5 proxies and am completely satisfied with their work. It's the third month I'm renewing my order. Simply the best!
cino

I was looking for a proxy with unlimited traffic for a long time, and this service turned out to be exactly what I needed. I would like to mention the support - if there are technical problems, they solve them quickly. Thank you for the high quality. I had some difficulties in my personal account, but we worked it out with support and I added an authorized IP. Great site, I will continue to use it.
JOHN

I am a developer and compatibility with my code is important to me. This service integrates with my Python projects without too much hassle.
Phoinix

I am satisfied with the service. Support service responds quickly and is competent in all matters. I buy proxies for FB here at a reasonable price without overpaying. No claims for speed, accounts are great on proxy. I recommend it!
Mohammed

I specialize in SEO promotion of various projects, including my own. The proxies sold at this service are perfect for finding "keywords" as well as for parsing. I am happy to buy proxies from different countries, as well as reasonable price. I am highly recommendable!
Evan Stewart

Hi everyone! I want to take the time to express my gratitude to this cool service! The proxies are great and affordable, the variety of countries and chat support is prompt. All this is available 24/7.
Sergio

Fast integration with API

Fast integration with API

Simple tool for complete proxy management - purchase, renewal, IP list update, binding change, upload lists. With easy integration into all popular programming languages, PapaProxy API is a great choice for developers looking to optimize their systems.

Quick and easy integration.
Full control and management of proxies via API.
Extensive documentation for a quick start.
Compatible with any programming language that supports HTTP requests.

Ready to improve your product? Explore our API and start integrating today!

Python
Golang
C++
NodeJS
Java
PHP
React
Delphi
Assembly
Rust
Ruby
Scratch

And 500+ more programming tools and languages

F.A.Q.

F.A.Q.

How to scrape protected sites? Close

Here are some general guidelines to approach scraping protected sites:

  1. Check Terms of Service:

    • Review the terms of service, robots.txt file, and any other usage policies provided by the website. Some websites explicitly prohibit scraping.
  2. Contact the Website Owner:

    • If you have a legitimate reason for scraping and the website provides contact information, consider reaching out to the website owner or administrator to request permission.
  3. Use Official APIs:

    • Many websites offer official APIs that allow controlled access to their data. Using an API is a sanctioned and structured way to obtain information.
  4. Simulate Human Behavior:

    • Mimic human behavior by setting appropriate headers, handling cookies, and introducing delays between requests. This helps avoid detection mechanisms that may identify automated bots.
  5. Handle CAPTCHAs:

    • Some sites may use CAPTCHAs to prevent automated access. Implement mechanisms to handle CAPTCHAs if they are encountered.
  6. Use Proxy Servers:

    • Rotate IP addresses using proxy servers to avoid IP-based blocking. However, be aware that some sites may block common proxy server IP ranges.
  7. Avoid Aggressive Scraping:

    • Limit the frequency and volume of your requests to avoid overloading the server. Implement rate limiting and throttling to mimic human browsing behavior.
  8. Stay Informed:

    • Monitor the website for changes in its structure or policies. Adjust your scraping strategy accordingly to adapt to any modifications.
Parsing JSON in TreeView Close

If you want to parse JSON data and display it in a TreeView in a Windows Forms application using C#, you can use the Newtonsoft.Json library for parsing JSON and the TreeView control for displaying the hierarchical structure. Below is an example demonstrating how to achieve this

Install Newtonsoft.Json

Use NuGet Package Manager Console to install the Newtonsoft.Json package:


Install-Package Newtonsoft.Json
  • Create a Windows Forms Application:

    • Open Visual Studio and create a new Windows Forms Application project.
  • Design the Form:

    • Drag and drop a TreeView control and a Button on the form.
  • Write Code to Parse JSON and Populate TreeView:


using System;
using System.Windows.Forms;
using Newtonsoft.Json.Linq;

namespace JsonTreeViewExample
{
    public partial class MainForm : Form
    {
        public MainForm()
        {
            InitializeComponent();
        }

        private void btnLoadJson_Click(object sender, EventArgs e)
        {
            // Replace with your JSON data or URL
            string jsonData = @"{
                ""name"": ""John"",
                ""age"": 30,
                ""address"": {
                    ""city"": ""New York"",
                    ""zip"": ""10001""
                },
                ""emails"": [
                    ""[email protected]"",
                    ""[email protected]""
                ]
            }";

            // Parse JSON data
            JObject jsonObject = JObject.Parse(jsonData);

            // Clear existing nodes in TreeView
            treeView.Nodes.Clear();

            // Populate TreeView
            PopulateTreeView(treeView.Nodes, jsonObject);
        }

        private void PopulateTreeView(TreeNodeCollection nodes, JToken token)
        {
            if (token is JValue)
            {
                // Display the value
                nodes.Add(token.ToString());
            }
            else if (token is JObject)
            {
                // Display object properties
                var obj = (JObject)token;
                foreach (var property in obj.Properties())
                {
                    TreeNode newNode = nodes.Add(property.Name);
                    PopulateTreeView(newNode.Nodes, property.Value);
                }
            }
            else if (token is JArray)
            {
                // Display array items
                var array = (JArray)token;
                for (int i = 0; i < array.Count; i++)
                {
                    TreeNode newNode = nodes.Add($"[{i}]");
                    PopulateTreeView(newNode.Nodes, array[i]);
                }
            }
        }
    }
}
    • In this example, the btnLoadJson_Click event handler simulates loading JSON data. You should replace it with your method of loading JSON data (e.g., from a file, a web service, etc.).
    • The PopulateTreeView method recursively populates the TreeView with nodes representing the JSON structure.
  1. Run the Application:

    • Build and run your application. Click the button to load the JSON data into the TreeView.

This example assumes a simple JSON structure. You may need to adjust the code based on the structure of your specific JSON data. The PopulateTreeView method handles objects, arrays, and values within the JSON data.

How to find out URL of new open windows in Selenium? Close

In Selenium, you can find out the URL of a newly opened window by switching to that window and retrieving its URL. Here's a step-by-step guide in Python:

1. Switch to the New Window

After opening a new window, you need to switch the focus of the WebDriver to that window.


from selenium import webdriver

driver = webdriver.Chrome()
driver.get("https://example.com")

# Open a new window (e.g., by clicking a link)
new_window_link = driver.find_element_by_link_text("Open New Window")
new_window_link.click()

# Switch to the new window
new_window_handle = driver.window_handles[-1]
driver.switch_to.window(new_window_handle)

In this example, replace "Open New Window" with the actual link text or locator that opens the new window.

2. Retrieve the URL of the New Window

Once you have switched to the new window, you can retrieve its URL using current_url.


new_window_url = driver.current_url
print("URL of the new window:", new_window_url)

This will print the URL of the new window. You can then store it in a variable or use it as needed in your script.

3. Switch Back to the Original Window (Optional)

If you need to switch back to the original window after retrieving the URL from the new window, you can do so using a similar process.


original_window_handle = driver.window_handles[0]
driver.switch_to.window(original_window_handle)

Replace 0 with the index of the original window's handle in the window_handles list.

Here's the complete example:


from selenium import webdriver

driver = webdriver.Chrome()
driver.get("https://example.com")

# Open a new window (replace with the actual link or action)
new_window_link = driver.find_element_by_link_text("Open New Window")
new_window_link.click()

# Switch to the new window
new_window_handle = driver.window_handles[-1]
driver.switch_to.window(new_window_handle)

# Retrieve the URL of the new window
new_window_url = driver.current_url
print("URL of the new window:", new_window_url)

# Switch back to the original window (optional)
original_window_handle = driver.window_handles[0]
driver.switch_to.window(original_window_handle)

# Continue with your script...

# Close the browser when done
driver.quit()

Make sure to adjust the code based on the actual actions and elements in your application that trigger the opening of a new window.

Scrapy: how to keep only unique external links? Close

To keep only unique external links while scraping with Scrapy, you can use a set to track the visited external links and filter out duplicates. Here's an example spider that demonstrates how to achieve this:


import scrapy
from urllib.parse import urlparse, urljoin

class UniqueLinksSpider(scrapy.Spider):
    name = 'unique_links'
    start_urls = ['http://example.com']  # Replace with the starting URL of your choice
    visited_external_links = set()

    def parse(self, response):
        # Extract all links from the current page
        all_links = response.css('a::attr(href)').extract()

        for link in all_links:
            full_url = urljoin(response.url, link)

            # Check if the link is external
            if urlparse(full_url).netloc != urlparse(response.url).netloc:
                # Check if it's a unique external link
                if full_url not in self.visited_external_links:
                    # Add the link to the set of visited external links
                    self.visited_external_links.add(full_url)

                    # Yield the link or process it further
                    yield {
                        'external_link': full_url
                    }

        # Follow links to other pages
        for next_page_url in response.css('a::attr(href)').extract():
            yield scrapy.Request(url=urljoin(response.url, next_page_url), callback=self.parse)

- visited_external_links is a class variable that keeps track of the unique external links across all instances of the spider.
- The parse method extracts all links from the current page.
- For each link, it checks if it is an external link by comparing the netloc (domain) of the current page and the link.
- If the link is external, it checks if it is unique by looking at the visited_external_links set.
- If the link is unique, it is added to the set, and the spider yields the link or processes it further.
- The spider then follows links to other pages, recursively calling the parse method.

Remember to replace the start_urls with the URL from which you want to start scraping.

What is a subnet in a proxy? Close

In simple terms, it is a logically separated part of the main local or public network. It is through it that many users can use a proxy through a single server at the same time. Each connection is allocated to a separate subnet.

Our statistics

>12 000

packages were sold in a few years

8 000 Tb

traffic spended by our clients per month.

6 out of 10

Number of clients that increase their tariff after the first month of usage

HTTP / HTTPS / Socks 4 / Socks 5

All popular proxy protocols that work with absolutely any software and device are available
With us you will receive

With us you will receive

  • Many payment methods: VISA, MasterCard, UnionPay, WMZ, Bitcoin, Ethereum, Litecoin, USDT TRC20, AliPay, etc;
  • No-questions-asked refunds within the first 24 hours of payment;
  • Personalized prices via customer support;
  • High proxy speed and no traffic restrictions;
  • Complete privacy on SOCKS protocols;
  • Automatic payment, issuance and renewal of proxies;
  • Only live support, no chatbots.
  • Personal manager for purchases of $500 or more.

What else…

  • Discounts for regular customers;
  • Discounts for large proxy volume;
  • Package of documents for legal entities;
  • Stability, speed, convenience;
  • Binding a proxy server for Sina Weibo unblock only to your IP address;
  • Comfortable control panel and downloading of proxy lists.
  • Advanced API.