I need suggestions about how to capture the data from a webpage without normal HTML parsing. The data is rendered to the screen via behind the scenes scripts and server side methods that I can't seem to unravel. I use BeautifulSoup and Selenium regularly, but this output is different.
The web page is very simple and static. While I cannot get the BS and Selenium to work, a simple "copy all" works perfectly. (yes, the old manual way!).
Is there any advice about how to automate this? Basically - > Go to Website, press "Copy ALL", return data to python and/or save to file for archive.
I have tried many different methods of BS and Selenium and only get part of the data. I think the rendering is done as an "anti-piracy" method. I've fought with this website many times in the past, and the way they post the data seems to be deliberately inconsistent. Randomly, the website works perfectly if I do the manual "copy/paste" thing, except that's not automated.
The data you see on the page is loaded from different URL in
CSVform. To load it you can use this example:Prints: