Guides on browser automation and reducing repetitive work
When setting up a browser automation, you must identify the elements to interact with. Here's how to use the Helper extension to select them easily.
This article will introduce you to some Chrome extensions for finding XPath, which help developers quickly locate and interact with specific elements on a web page while doing automated tests.
Scraped data must be stored in an easily accessible place so you can use it for other workflows. Here's how to automatically send data to Google Sheets without code.
Scraped data should be stored in an easily accessible place so you can use it for other workflows. Here's how to automatically send your data to Notion.
Learn how to take scheduled website screenshots in desktop, tablet, and mobile views using Browserbear in Node.js. You will also learn how to deploy AWS Lambda functions using the Serverless Framework to automate the screenshot task.
Scraped data should be stored in an easily accessible place so you can use it for other workflows. Here's how to automatically send your data to Airtable.
Web scraping has become a crucial part of data gathering in today's digital age. It involves extracting data from websites automatically using various tools. In this article, we will explore different web scraping tools including Cheerio, Puppeteer, Nightmare, Playwright, and Browserbear.
Automated screenshots are easy to set up, but tricky to optimize. Learn how to take better screenshots with Browserbear and store them efficiently for later use.
In this article, we will discuss the advanced techniques for web scraping with Browserbear. Building upon the basic understanding of web scraping introduced in Part 1 of the tutorial, we will show you how to scrape more information using the data that we got from the previous task.
Sign up for our once a fortnight newsletter update on new Roborabbit features and our business journey