zskid00 Posted April 11, 2017 Posted April 11, 2017 1 hour ago, xadrian said: I had to uninstall the app because I was getting email, desktop pop up, and an alert on my phone about it. Do you WUPHF? 3 1 Quote
Jeff Mack Posted April 11, 2017 Posted April 11, 2017 45 minutes ago, exciter1 said: We have a scripting thread, but I'm at Legoland and can't search it right now. Ok, i thought we did, merged it here now. 1 Quote
ingotheranchhand Posted June 15, 2020 Posted June 15, 2020 I searched this thread and didn't seem to find any discussion on this... I feel like there are computerized bot programs snatching up sets as soon as they become available. A few programs became very popular for snatching up Nintendo Switches online using bots - practically instantly selling them out as soon as they were restocked. Has any one experiences similar issues, hunches, or have any insight into how this can be combated? Or specifically what bots are being employed? Quote
Hardwired Posted June 15, 2020 Posted June 15, 2020 Is it against the rules anywhere to use bots? I, uh, ask for a friend. Quote
TheBrickClique Posted June 17, 2020 Posted June 17, 2020 I don’t think there are bots so much as there are other issues at play. Namely, Amazon has no limits on LEGO as far as I can tell. Folks are buying everything they can we something goes in stock. Target has relaxed limits as well. Many restocks are only one or two at a time. So you are racing against others for the sole item. And by the time you know about it, it could have been available to others for awhile. Years ago I had an auto-purchase script that I built myself to purchase — quite unfortunately — pet shops at RRP when they were retiring. It worked fine, but eventually Amazon found a way to break it. I haven’t bothered with it since. I’ve also dabbled with scripts that would alert you to web page changes. If set up to just poll a handful of pages fairly often (every minute for example) you can catch most items but you have to be ready when the alert goes off. The big in stock tracker sites are monitoring about once every ten minutes. When you get an alert from them, it’s a rat race! Lastly, I have been up working late more often the past few weeks. You can definitely get sets a lot easier at 3AM than 3PM. Which tells me the auto buy bots are not running since they don’t sleep. So, set up some page monitors, have your credit card info pre-programmed and amazon buy it now set up, Fire up your favorite game or queue up some movies, and see what you can score over night! 1 Quote
scurt1s Posted June 17, 2020 Posted June 17, 2020 1 minute ago, TheBrickClique said: I don’t think there are bots so much as there are other issues at play. Namely, Amazon has no limits on LEGO as far as I can tell. Folks are buying everything they can we something goes in stock. Target has relaxed limits as well. Many restocks are only one or two at a time. So you are racing against others for the sole item. And by the time you know about it, it could have been available to others for awhile. Years ago I had an auto-purchase script that I built myself to purchase — quite unfortunately — pet shops at RRP when they were retiring. It worked fine, but eventually Amazon found a way to break it. I haven’t bothered with it since. I’ve also dabbled with scripts that would alert you to web page changes. If set up to just poll a handful of pages fairly often (every minute for example) you can catch most items but you have to be ready when the alert goes off. The big in stock tracker sites are monitoring about once every ten minutes. When you get an alert from them, it’s a rat race! Lastly, I have been up working late more often the past few weeks. You can definitely get sets a lot easier at 3AM than 3PM. Which tells me the auto buy bots are not running since they don’t sleep. So, set up some page monitors, have your credit card info pre-programmed and amazon buy it now set up, Fire up your favorite game or queue up some movies, and see what you can score over night! Whats frustrates me is that my target red card always gives me problems during the checkout (even though I have enough $$$ for the debit card) and I've been cart jacked. No more messing around w/ it on time sensitive stuff. Quote
Hardwired Posted June 17, 2020 Posted June 17, 2020 I have a series of Selenium scripts that are running that will check for inventory, and they run constantly over a VPN connection that I shuffle the IP on pretty regularly. There is some other op sec that goes into it to avoid detection. I can tell you down to the minute last night when the Assembly Squares came up, because they would be gone almost instantly --- and so I know I can't be the only one. I've set some of them to auto-purchase and watch them the most for obvious reasons, but they are all set with a variable in the script that will not let them buy over a certain number so its a bit of a safeguard. 1 Quote
landphieran Posted June 18, 2020 Posted June 18, 2020 On 6/17/2020 at 2:34 AM, Hardwired said: I have a series of Selenium scripts that are running that will check for inventory, and they run constantly over a VPN connection that I shuffle the IP on pretty regularly. There is some other op sec that goes into it to avoid detection. I can tell you down to the minute last night when the Assembly Squares came up, because they would be gone almost instantly --- and so I know I can't be the only one. I've set some of them to auto-purchase and watch them the most for obvious reasons, but they are all set with a variable in the script that will not let them buy over a certain number so its a bit of a safeguard. I've been working on a strategy similar to this but found the amount of IP and VPN switching to be quite cumbersome. I went a different route and started putting together an application to track sales and pricing data on Amazon. It should be able to pull multiple data points on every Lego set a day. I plan on using it track set popularity over time and understand how frequently sets go out of stock, use it to monitor prices and available quantities (from other sellers) of sets i'm interested in selling. I don't have a lot of experience with web crawlers but once I finish up the Amazon application I was going to work on something like you mentioned above. I have a partially working email scrapper that grabs all my receipts and puts them in a database to make inventory intake much easier. I have a lot to finish up but I plan on automating some of the tedious portions of the job by the end of the year. 1 Quote
BrickLegacy Posted June 18, 2020 Posted June 18, 2020 I've been working on a strategy similar to this but found the amount of IP and VPN switching to be quite cumbersome. I went a different route and started putting together an application to track sales and pricing data on Amazon. It should be able to pull multiple data points on every Lego set a day. I plan on using it track set popularity over time and understand how frequently sets go out of stock, use it to monitor prices and available quantities (from other sellers) of sets i'm interested in selling. I don't have a lot of experience with web crawlers but once I finish up the Amazon application I was going to work on something like you mentioned above. I have a partially working email scrapper that grabs all my receipts and puts them in a database to make inventory intake much easier. I have a lot to finish up but I plan on automating some of the tedious portions of the job by the end of the year.A service to monitor the ups and downs of prices for amazon products? Reminds me of the humps in a camel. 2 Quote
landphieran Posted June 18, 2020 Posted June 18, 2020 20 minutes ago, BrickLegacy said: A service to monitor the ups and downs of prices for amazon products? Reminds me of the humps in a camel. Camel does a really poor job at monitoring 3rd party seller or non buy box offerings. Additionally it doesn't track sales rank (which is what I'm most interested in). Quote
BrickLegacy Posted June 18, 2020 Posted June 18, 2020 Camel does a really poor job at monitoring 3rd party seller or non buy box offerings. Additionally it doesn't track sales rank (which is what I'm most interested in).Just giving you a hard time... it’s definitely not a reseller-grade service. Quote
Hardwired Posted June 18, 2020 Posted June 18, 2020 3 hours ago, landphieran said: I've been working on a strategy similar to this but found the amount of IP and VPN switching to be quite cumbersome. I went a different route and started putting together an application to track sales and pricing data on Amazon. It should be able to pull multiple data points on every Lego set a day. I plan on using it track set popularity over time and understand how frequently sets go out of stock, use it to monitor prices and available quantities (from other sellers) of sets i'm interested in selling. I don't have a lot of experience with web crawlers but once I finish up the Amazon application I was going to work on something like you mentioned above. I have a partially working email scrapper that grabs all my receipts and puts them in a database to make inventory intake much easier. I have a lot to finish up but I plan on automating some of the tedious portions of the job by the end of the year. Why are you concerned about IP and VPNs if you are just tracking pricing and sales ? I don't want anyone to catch me as a bot in there performing transactions (hence the op sec) and truth be told, I do all this in headless mode so the longest delay I have anywhere is that damn VPN/IP switching. In and out of any listing in under 3 tenths of a second. Switching IPs... tens of seconds. Cursed technology. Quote
donbee Posted June 19, 2020 Posted June 19, 2020 Why are you concerned about IP and VPNs if you are just tracking pricing and sales ? I don't want anyone to catch me as a bot in there performing transactions (hence the op sec) and truth be told, I do all this in headless mode so the longest delay I have anywhere is that damn VPN/IP switching. In and out of any listing in under 3 tenths of a second. Switching IPs... tens of seconds. Cursed technology. Can I get one of those? Quote
Hardwired Posted June 19, 2020 Posted June 19, 2020 1 hour ago, donbee said: Can I get one of those? Selenium is open source, is that what you're talking about? Quote
donbee Posted June 20, 2020 Posted June 20, 2020 A bot, or an army of bots. Small party would suffice. Will trade for functioning scraper. Quote
Hardwired Posted June 20, 2020 Posted June 20, 2020 1 hour ago, donbee said: A bot, or an army of bots. Small party would suffice. Will trade for functioning scraper. What are you using a scraper for? Quote
Alpinemaps Posted June 20, 2020 Posted June 20, 2020 What are you using a scraper for?Not@donbee, but I’d like to create a scraper to look at LEGO Ideas daily to populate a spreadsheet. So I can do a little analysis whenever a new set breaks the 6,000 supporter mark, or achieves support, or is climbing the ladder quick. 1 Quote
Hardwired Posted June 20, 2020 Posted June 20, 2020 https://ideas.lego.com/search/global_search/ideas?support_value=6000&support_value=10000&idea_phase=idea_gathering_support&query=&sort=most_recent A URL that gives you all ideas with more than 6000 supporters, sort by newest first. I'm not sure what would count as rising quickly. What sort of analysis are you doing? 2 hours ago, Alpinemaps said: Not@donbee, but I’d like to create a scraper to look at LEGO Ideas daily to populate a spreadsheet. So I can do a little analysis whenever a new set breaks the 6,000 supporter mark, or achieves support, or is climbing the ladder quick. Quote
CsabaLazlo Posted June 20, 2020 Posted June 20, 2020 Can anyone recommend any decent tutorials for writing a Python script that monitors prices and emails me when they drop below a certain point? I've managed to write a script that monitors the price of one item using the Beautiful Soup module but, that obviously isn't that helpful when you consider that I'm interested in 20+ items across multiple sites. From what I can gather Selenium seems to be a better package for this application? All of my coding experience up to this point is using R to build models that predict species abundance and distribution so I'm very much a beginner with Python. 1 Quote
Hardwired Posted June 20, 2020 Posted June 20, 2020 You could conceivably do it in either language. The main appeal of Selenium for me was that it has everything in one package (with the possible exception of a database, maybe) and so it was faster to deploy for me. The problem I think you are going to run into is the reporting function you are looking for. If you are checking 20+ items across multiple sites your scope might be a little too broad. Quote
donbee Posted June 20, 2020 Posted June 20, 2020 Offer still stands. If you have ability, I have something to trade. Quote
Alpinemaps Posted June 20, 2020 Posted June 20, 2020 https://ideas.lego.com/search/global_search/ideas?support_value=6000&support_value=10000&idea_phase=idea_gathering_support&query=&sort=most_recent A URL that gives you all ideas with more than 6000 supporters, sort by newest first. I'm not sure what would count as rising quickly. What sort of analysis are you doing? That’s the easy part. What I want to scrape is a few things, daily, like the name, URL, author, and supporter count for an individual project that has at least 6,000 supporters. Anytime a project hits 6000 it needs to add it to the list. Compare the date a project was submitted to the number of votes that it has received. So If it hits 1000 supporters in say 5 days, add that to the list of “watch this one.”I’ll have to check my notes but I think the issue was scrapping the search page and not the individual product page. But it’s also trying to figure out exactly which element needed to be scraped causing me issues. I never got to the point of setting anything up in Python, as I was brute forcing it trying to figure out the elements first and never got past that stage. 1 Quote
Hardwired Posted June 20, 2020 Posted June 20, 2020 8 minutes ago, Alpinemaps said: That’s the easy part. What I want to scrape is a few things, daily, like the name, URL, author, and supporter count for an individual project that has at least 6,000 supporters. Anytime a project hits 6000 it needs to add it to the list. Compare the date a project was submitted to the number of votes that it has received. So If it hits 1000 supporters in say 5 days, add that to the list of “watch this one.” I’ll have to check my notes but I think the issue was scrapping the search page and not the individual product page. But it’s also trying to figure out exactly which element needed to be scraped causing me issues. I never got to the point of setting anything up in Python, as I was brute forcing it trying to figure out the elements first and never got past that stage. I'm just curious -- to what end? Are you looking for trends in ideas that get support? You could just as easily get a count of all projects over 6,000 votes on a daily basis and if that number changes then do a scrape. It would save you some cycles in theory since it should not happen too often. Unless you just want to track all projects with over 6000 votes, like you mentioned for velocity. Hmm, interesting concept but I still am unclear on why you'd do that? Quote
Alpinemaps Posted June 20, 2020 Posted June 20, 2020 1 hour ago, Hardwired said: I'm just curious -- to what end? Are you looking for trends in ideas that get support? You could just as easily get a count of all projects over 6,000 votes on a daily basis and if that number changes then do a scrape. It would save you some cycles in theory since it should not happen too often. Unless you just want to track all projects with over 6000 votes, like you mentioned for velocity. Hmm, interesting concept but I still am unclear on why you'd do that? I'm a nerd that likes that kind of stuff? lol This is a good example of what I like to do with the data: 1 Quote
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.