Changing external IP address in Python - python

I have a python robot that accesses internet websites and performs some tasks (using selenium and google chrome). I would like make the python code to change my external IP address every, say, 10 interactions that the code performs in the websites. Is it possible? Would you have the adequate line of code for this task?
Note that the websites that I access recognize VPN, so when I try changing my IP address through VPN, I cannot access these websites.
Thank you

Related

How can I use Python to automate accessing a website through a VPN (providing the login details to said VPN in the code itself)

I am a beginner programmer. How would I approach this problem? I want to provide Python with certain webpages and certain actions to take on said webpages. The problem is, the webpages are region restricted, so I have to use a VPN constantly. Would there be any way to have Python automatically connect to a vpn service (Mullvad, NordVPN etc) to a specific country while running the code? Thanks.
Excluding VPNs you could use proxies. But if you need to use a VPN I suggest looking at the Google results for your specific provider, like this one for Nord.

is there any way to get web-app user IP address

In my python project I have to track user IP address and country and region, and other information, I used "https://ipapi.co/json/" for tracking the IP address, but thing is when i deploy my model(web-app) on the deployment website, it is giving me that server IP address instead who is using my web-app. when i try in my local machine it is giving my IP address. Actually, i am currently new and don't know that much about tracking IP another.
Please anyone can describe how i will achieve this thing. how to track the user of my web-app IP address instead of the deployment server IP address.
Ref- I am using streamlit sharing for the deployment it is giving me streamlit office address instead of my friend(test user) IP. I need my fried IP address when he is using the deployed web app.
Thank you sir for your consideration, a help will be great for me, am really stuck here. my project in Python.
It sounds like you're trying to get the IP from this 3rd party service from your server, which would of course return the server's IP address.
Based on a comment on the question:
I am using streamlit for creating the web app.
Based on a cursory Google search, it sounds like getting the client's IP is non-trivial in that system:
https://discuss.streamlit.io/t/i-run-a-streamlit-app-and-it-processes-some-records-based-on-client-user-input-i-want-to-log-ip-address-of-client-user-and-records-processed-by-that-user-can-you-help-please/2382
https://discuss.streamlit.io/t/how-to-get-all-ip-addresses-and-their-countries-connecting-to-a-live-streamlit-app-on-aws-ec2/2273
https://discuss.streamlit.io/t/how-to-output-client-or-remote-ip-s-to-console/832
(Based on this I would probably recommend using a different Framework/platform for building web applications. This sounds... extremely limited. But that's another matter entirely.)
The service you reference would need to be accessed from the client computer, which means accessing it from JavaScript. If that service doesn't allow CORS requests, there are other options available. Once you have the value(s) you need in JavaScript, you can send them to your server-side code via an AJAX request, a Form POST, etc.

How to hide my IP doing web scraping in python?

I am doing web scraping with python in some pages and I have been blocked from some of them. When I have tried to check it also through the TOR Browser I have seen that I cannot access to the pages neither, so I think that these pages have been able to track all my IP or I dont have well configurated TOR (and I think not cause I have checked my IP address with Chrome and TOR and are different), so, any one knows why?
Also, I am trying to do a function or method in my python code to change mi IP automatically. What I have seen is that the best is to do it through the TOR browser (using it as the search engine to get data from pages) but I am not able to make it work. Do you have any recommendation to create this function?
Thank you!
I would expect anti scrape protection to also block visits from known Tor exit nodes. I dont think they know it is you. Some websites hire/implement state of the art scrape protection services.
You could setup your own proxies at friends and family and use a very conservative crawl rate or maybe search for commercial residential proxy offerings.

Using Scrapy with VPN on Ubuntu

So i am using Scrapy to crawl some websites and i want to increase my privacy on the internet and also avoid getting banned so i read that i could achieve that by using premium proxy lists like http://www.ninjasproxy.com/ or http://hidemyass.com/ or VPN or Tor.
From what i understood a paid VPN would be a good option like the one http://hidemyass.com/ offers, but i can't seem to find any code that actually shows Scrapy integrating with a VPN like hidemyass.
I only saw an example like https://github.com/aivarsk/scrapy-proxies that shows how to use proxy lists.
How do i make Scrapy work with a VPN? If i can't are proxy lists good enough to maintain anonymity?
A VPN is something working system wide, not something that proxy selected traffic. All your internet traffic, browser, torrent, chat etc etc will be routed through the VPN, so just connect to the VPN and run the script.

Python Proxy Firefox

I am trying to make a "proxy" in Python that allows the user to route all of their web traffic through a host machine (this is mainly for me and a couple people I know to confuse/avoid hackers and/or spies, who would only see web pages and so on coming in through one IP). I have run into several difficulties. The first is that I would like to be able to use the final, compiled product with Firefox, which can be set to route all of its traffic through an installed proxy program. I don't know what kind of configuration my proxy needs to have to do this. Second, the way the proxy works is by using urllib.requests.urlretrieve (yes, going to die soon, but I like it) to download a webpage onto the host computer (it's inefficient and slow, but it's only going to be used for a max of 7-10 clients) and then sending the file to the client. However, this results in things like missing pictures or broken submission forms. What should I be using to get the webpages right (I want things like SSL and video streaming to work as well as pictures and whatnot).
(wince) this sounds like "security through obscurity", and a lot of unnecessary work.
Just set up an ssh tunnel and proxy your web traffic through that.
See http://www.linuxjournal.com/content/use-ssh-create-http-proxy

Categories

Resources