It’s a great tool for people without Python / BeautifulSoup skills. For most cases which are a bit more complex, the conventional approach is probably the best ;)
can you do a walkthrough to read a given url of news web page to come up with context of the news and come up with a summary through LLM, ideally using Ollama
Thanks for your comment. It’s impossible for me to make a video about every possible use case. Instead I try to make videos which can be easily applied to other scenarios. In this specific case, it’s just a matter of modifying the prompt and the url and you’re good to go 😉
can you do a walkthrough to read a given url of news web page to come up with context of the news and come up with a summary through LLM, ideally using Ollama
whenever I install scrapegraphai I am hit by AttributeError: module 'pkgutil' has no attribute 'ImpImporter'. Did you mean: 'zipimporter'? [end of output]
Hi, thanks for your comment. I have seen this error before, is it possible you're running Python 3.12? If so, try running it in a venv with 3.11. Good luck! If you need more help, you can reach me on my Discord (link in description).
I keep getting this error "PermissionError: [Errno 13] Permission denied: 'wired.xlsx'" and I've even typed the code shown in your video exactly and still I get that error, any help would be greatly appreciated
Hi Meta, it looks like you're trying to write a file to a directory where you don't have enough permissions. Assuming you're using Windows, can you try the following? Navigate to your Desktop, right click on any of your files there. Click properties. Go to the tab "general" and read location. This should be something like: "C:\Users\thoma\Desktop" (where thoma is your username). Copy this location to VS Code. Now add \wired.xlsx and change all slashes for double slashes (you should now have something like this: df.to_excel("C:\\Users\\thoma\\Desktop\\wired.xlsx",index=False) Run the code again. Let me know whether this works for you. If not, feel free to join my Discord if you need more help (link in video description).
@@TomsTechAcademy Hi, I managed to get it to produce two results and now I'm having issues again, It is now giving me this message "Traceback (most recent call last): File "c:\scrapperai\main.py", line 26, in df = pd.DataFrame(result['pages']) ~~~~~~^^^^^^^^^" any help would be greatly appreciated!
I found ScrapeGraphAI pretty useless, it doesn't except for the simplistic examples.
It’s a great tool for people without Python / BeautifulSoup skills. For most cases which are a bit more complex, the conventional approach is probably the best ;)
It extracted the titles, not the articles themselves.
Thanks for writing this
can you do a walkthrough to read a given url of news web page to come up with context of the news and come up with a summary through LLM, ideally using Ollama
Thanks for demo-ing the tool
Thank you for your comment, much apreciated!
Great video thx !
Thank you Bugra! 😀
Video could have been more helpful if it was about scraping helpful marketing information like emails and phone numbers
Thanks for your comment. It’s impossible for me to make a video about every possible use case. Instead I try to make videos which can be easily applied to other scenarios. In this specific case, it’s just a matter of modifying the prompt and the url and you’re good to go 😉
this video tutorial was suprising beginner friendly
Thank you for your comment Amanda, and glad my video was useful for you :) I always try to make my videos as beginner friendly as possible
can you do a walkthrough to read a given url of news web page to come up with context of the news and come up with a summary through LLM, ideally using Ollama
Can you scrape data from Google maps?
Thank you so much for showing the demo! I still have a question:does scrapegraphai support full stack crawling? or it only works for one page
i get api error
If you receive any error, feel free to join my discord (link in description) and post your question there 😀
whenever I install scrapegraphai I am hit by AttributeError: module 'pkgutil' has no attribute 'ImpImporter'. Did you mean: 'zipimporter'?
[end of output]
Hi, I think I had this exact same error - are you running Python in a virtual environment? If you run Python 3.12 you can try switching to 3.11
I just give chat gpt the html data as example. Give it what I want scraped and it’ll generate the code for me
Thanks for Sharing your expertise, that’s another way of doing it 😀
When i am trying this i get the error asyncio error cannot be called
Hi, thanks for your comment. I have seen this error before, is it possible you're running Python 3.12? If so, try running it in a venv with 3.11. Good luck! If you need more help, you can reach me on my Discord (link in description).
I keep getting this error "PermissionError: [Errno 13] Permission denied: 'wired.xlsx'" and I've even typed the code shown in your video exactly and still I get that error, any help would be greatly appreciated
Hi Meta, it looks like you're trying to write a file to a directory where you don't have enough permissions. Assuming you're using Windows, can you try the following?
Navigate to your Desktop, right click on any of your files there. Click properties. Go to the tab "general" and read location. This should be something like: "C:\Users\thoma\Desktop" (where thoma is your username). Copy this location to VS Code. Now add \wired.xlsx and change all slashes for double slashes (you should now have something like this:
df.to_excel("C:\\Users\\thoma\\Desktop\\wired.xlsx",index=False)
Run the code again. Let me know whether this works for you. If not, feel free to join my Discord if you need more help (link in video description).
@@TomsTechAcademy Hi, I managed to get it to produce two results and now I'm having issues again, It is now giving me this message "Traceback (most recent call last):
File "c:\scrapperai\main.py", line 26, in
df = pd.DataFrame(result['pages'])
~~~~~~^^^^^^^^^" any help would be greatly appreciated!