this post was submitted on 12 Nov 2023
1 points (100.0% liked)

Data Hoarder

0 readers
3 users here now

We are digital librarians. Among us are represented the various reasons to keep data -- legal requirements, competitive requirements, uncertainty of permanence of cloud services, distaste for transmitting your data externally (e.g. government or corporate espionage), cultural and familial archivists, internet collapse preppers, and people who do it themselves so they're sure it's done right. Everyone has their reasons for curating the data they have decided to keep (either forever or For A Damn Long Time (tm) ). Along the way we have sought out like-minded individuals to exchange strategies, war stories, and cautionary tales of failures.

founded 1 year ago
MODERATORS
 

https://www.resauctions.com/auctions/24572-two-day-richard-and-mary-lou-taylor-lifetime-collection-absolute-auction?page=2

I want to archive just this auction, mainly the photos but the closing prices would be nice too. So my goal is either a folder of photos, or that and a browsable offline copy.

Simply right clicking and saving works decently enough but surely there is a faster way. There are 16 pages and I'm actually doing 2 separate auctions.

My usual script for a webpage is crawling the whole site and pulling various vendor information instead of just this webpage. My script usually doesn't do that so I'm thinking it's something to do with how the site is structured.

wget -p --convert-links -e robots=off -U mozilla --no-parent https://www.resauctions.com/auctions/24572-two-day-richard-and-mary-lou-taylor-lifetime-collection-absolute-auction?page=2

Any advice would be appreciated.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 1 points 1 year ago

I suppose I could pull the entire site and delete the fluff afterwards, but that seems like a very long and resource intensive process for such a small grab.