51,746 questions
0
votes
0
answers
64
views
Downloading Barchart.com table using Excel VBA
I'm trying to download the barchart data table from https://www.barchart.com/investing-ideas/ai-stocks
using Excel VBA in similar manner as the python script in Automatic file downloading on Barchart....
-1
votes
1
answer
44
views
Selenium script marks all search results as “not found” because details load only after clicking a link
I’m using Python + Selenium + ChromeDriver to check a list of titles (from a CSV file) against an online library catalog.
My script searches each title and tries to determine if a specific library has ...
3
votes
2
answers
125
views
Importing a table from a webpage as a dataframe in Python
I am trying to read in a specific table from the US Customs and Border Protection's Dashboard on Southwest Land Border Encounters as a dataframe.
The url is: https://www.cbp.gov/newsroom/stats/...
-6
votes
0
answers
74
views
how implement in a Python algorithm to avoid exceeding the server's rate limits? [closed]
I'm working on a tourism data extractor focused on the Foursquare platform.
I currently don't have any proxies or techniques to avoid being blocked for exceeding the server's request limit (50 ...
-1
votes
1
answer
100
views
Webscrape links to download files based on word in page HTML
I am webscraping WHO pages using the following code:
pacman::p_load(rvest,
httr,
stringr,
purrr)
download_first_pdf_from_handle <- function(handle_id) {
...
1
vote
2
answers
137
views
Scraping booking.com with PHP Curl [closed]
I'm trying to setup a simple php script (below) to check when a certain property becomes available on booking.com.
The site seems to work with some pretty simple query parameters - you can just go ...
0
votes
0
answers
54
views
Unable to scrape product price from Shein ShareJump links in Laravel/Python
I’m working on a project in Laravel/Python where I want to fetch product information from Shein, but I’ve run into a major problem with ShareJump links.
Here’s an example link I’m working with:
http://...
1
vote
1
answer
96
views
Scraping archived content
I am a bit new to webscraping and trying to build a scraper to collect the title, text, and date from this archived page:
from selenium import webdriver
from selenium.webdriver.chrome.service import ...
1
vote
2
answers
71
views
Cannot click <a> element button with href="javascript:void(0)" with selenium
I'm using selenium in python and trying to click the "See all Properties" button to get to the next web page where all the properties will be listed and I can easily scrap the data.
Here's ...
0
votes
0
answers
144
views
Scraping Instagram Likes at Bulk
My goal is to find out if a given user has liked any post of another profile.
So the following question has to be answered: Has the user X liked any post on the profile Y in the past 24 months.
For ...
0
votes
2
answers
97
views
Selenium interaction with accordion list
I'm trying to scrape the data off this site.
The website shows a charging station, in this case you can click each to unravel the accordion and see the data per charger. I am trying to use this ...
3
votes
1
answer
130
views
How to clean inconsistent address strings in Python?
I'm working on a web scraping project in Python to collect data from a real estate website. I'm running into an issue with the addresses, as they are not always consistent.
I've already handled simple ...
1
vote
0
answers
61
views
Make.com Text parser: Attributes.href is empty — how to filter <a> links by href (relative + absolute) before aggregating?
ody:
I’m building a Make.com scenario like this:
HTTP (fetch website HTML)
→ Text parser (extract elements)
→ Filter "only good links"
→ Array aggregator
→ further processing
Goal
I want ...
-1
votes
3
answers
156
views
Unable to scrape 2nd table from Fbref.com for players table
I would like to scrape the 2nd table in the page seen below from the link - https://fbref.com/en/comps/9/2023-2024/stats/2023-2024-Premier-League-Stats on google collab. But pd.read_html only gives me ...
0
votes
1
answer
88
views
How Do I Use Proxies with Puppeteer and a Local Chrome Instance?
I'm using Puppeteer and JS to write a web scraper. The site I'm scraping is pretty intense, so I need to use a local chrome instance and a residential proxy service to get it working. Here's my basic ...