@@codeRECODE Sir I'm using python and waiting for the next tutorial. Your teaching method is very easy to understand. Please upload the next video of this tutorial series.
As a user of both Selenium and Playwright at work, I must say Playwright is far superior, much faster, better documentation, and a lot less flaky. Once the Playwright community grows, it will become the industry standard without a doubt.
I am getting error: It looks like you are using Playwright Sync API inside the asyncio loop. Please use the Async API instead. I tried async calls too but still error. Any idea how to solve this? I am using jupyter notebook.
I've done scraping the contents on the first page and all the next pages but what to do when we need to store the scraped data. Do I need to use another library for that or playwright has inbuilt storing in different format commands??
Would’ve been helpful to mention why someone might use this over selenium, it’s advantages and disadvantages. Saying one is interesting doesn’t explain much. Thanks for the video.
It is complex. You need to know how websites work and then every website has it's own challenge :-) You need to spend time to learn, or hire someone else to do that for you!
Selenium miss match the time of selectors in headless it is so much easy with playwright and fast.. also we need to use explicit wait in each element find in selenium but in playwright we can simple pass the wait with so much ease... I will definitely go with playwright
My system does not show browser even if i have set timeout to page.wait_for_timeout(50000) simply terminal shows [Done] exited with code=0 in 0.175 seconds, headless=False
Run this and see if this works - github.com/coderecode-com/playwright_sync_api_intro/blob/main/quotes.py If this doesn't work, try another browser. For example, for firefox, change line 6 to the following ``` browser = p.firefox.launch(headless=False) ```
Absolutely great videos. I been following along for quite few videos now and am very grateful for your teachings... would definitely be interested on seeing more videos like this. (about scraping dynamics webpages)
Hi how can we do this with playwright or if is it even possible? in Scrapy response.replace(body=webdriver.page_source) using selenium, how can we do this with playwright, i tried response.replace(body=page.content()) but this doesn't work, kindly help!
Why replace the response? You can create a new selector object from scrapy.selector import Selector s = Selector(text=page.content()) s.xpath('').get()
Nice video bro. When installing pip install playwright browser in vscode, i'm getting failed error which says "failed install chromium v100 and unable to get local issuer certificate" in my company laptop. Do you have any suggestion?
I did a google search with your error and this is what I got: 1. Reported here github.com/microsoft/playwright/issues/5636 2. Solution here: playwright.dev/docs/browsers#install-behind-a-firewall-or-a-proxy
@@sheikhakbar2067 use any vpn. I recommend vyprvpn because it's cheap. VPN is not at browser level but machine level. Once you connect to VPN, all apps including selenium will use VPN
Very very very good explanation. You have convinced me to switch to playwright. Scrapy + playwright is a deadly combo and thank you for such a beautiful explanation
Web scraping with Go is a good idea. Not sure about Go with Playwright. There is a Playwright for Go - pkg.go.dev/github.com/mxschmitt/playwright-go It's not official though. Officially supported languages are JS, Python, Java and C#
I just love your content ❤️ you make it look so simple 😅 I've done no professional course of python yet I'm using scrapy to scrape sites and doing automation just because of your videos I've learnt alot ❤️ thank you and please keep up the good work.
Hello Sir, I have watched your Scrapy selenium tutorial and tried to implement the same procedure with Scrapy-Playwright. When I tried to run the script "raise Error( playwright._impl._api_types.Error: It looks like you are using Playwright Sync API inside the asyncio loop. Please use the Async API instead." this error is raised every time. Why this thing is happening? By the way, eagerly waiting for your scrapy playwright tutorial.
For writing to files, the usual CSV, JSON modules can be used. To extract title from all pages, visit all pages and use selectors. For this specific scenario, scrapy crawl spider is more suited though.