Blog
Nov 14, 2025 - 3 MIN READ
Exploring AI-Driven Web Crawling: A Beginner’s Journey

Exploring AI-Driven Web Crawling: A Beginner’s Journey

Follow my hands-on journey as I dive into the world of AI-powered web crawling. From understanding core concepts to testing real tools and building small experiments, this blog documents my discoveries, challenges, and insights while exploring how artificial intelligence is reshaping how we gather and analyze information online.

Niklas Grieger

Niklas Grieger

My journey

I'm building a plattform for the bitcoin community. On feature of this will price comparison app for bitcoin related products. To show all the products I need to crawl many different e-commerce websites.
I know traditional crawling and that you have to define the css schema of the website so the crawler can find the article information. A big disadvantage of this the structure of the website can change, you have to recognize the change and change the css schema. I started researching about AI web crawling and found ton's of tools to use. I can be overwhelming and it's not easy to find a solution that fits my needs.

Challenges

Problem 1: I cannot crawl all of these products and save them, that would be insane to save in a database
Problem 2: All of these websites are having different designs
Problem 3: These website are not all e-commerce pages, they can also be e-sim provider, domain provider etc.

Idea for Problem 1: I will scrape only the search page of each website e.g https://webshop.de/search?q=product1 so I will have a smaller playload to save.
Idea for Problem 2: I will try to find a good AI toll and define a prompt to find the correct data on each website. As alternative I will use the CSSShema extraction from Crawl4AI.
No Idea for Problem 3: This will be challenging to crawl with AI cause the parameter are totally different

Exploring AI web crawlers

Thunderbolt

https://thunderbit.com/

This was the first tool I found. I installed their Chrome Extention and tried it in one e-commerce shop search page.

It worked great! I defined the fields and one prompt per field. No css shema required. I could even define that the crawler does paging.

There is just one downside: Thunderbit has no API and it's just usable as a Chrome extention for manual crawling and export. In my case I need to dynamically crawl different websites in different structure. So this tool is not working for my case.
Anyway I can recommend this tool if you just have one or some websites to crawl and are fine with the manual effort to export the result to .csv or other formats.

BrowserAI

https://www.browse.ai/

I tried this tool by settings up a robot on their website. It extracted the data as I wanted but some information as the price were missing.

Additionally the pricing doesn't make sense for my case since I have 1000+ different websites.

Firecrawl

This tool is promising since it is developer friendly by offering a Rest API with crawling and scraping functionalities. On their website I could try the API's but turns out the data was often not correct or missing. The missing data comes probably from lazy loading of the website, but Firecrawl does not offer a good handling of lazy loading yet.

I tried their Rest API with postman but I had missing data, the task was running very long and occurs to timeouts, and all my credits were accidently used by one API call.

I think firecrawl is on a good way, is highly customizable, and offer llm extraction by API call. But the incorrect data and lazy loading handling is not good enough for my case.

Additionally I had the problem to scrape with an defined json array. When you define only one json object it works fine in the playground - see the video.
May I give it another try in the future

Crawl4AI - my solution for now

TBD...

Hello,

I have been working as a full-stack developer for 8 years, with a focus on frontend development. I have been working in frontend development for 7 years with VueJS versions 2 and 3 and have successfully migrated several projects to version 3, among other things. Furthermore, I also have extensive experience with NuxtJS 2/3 and nowadays Version 4. I have also used PrimeVue and Vuetify as component libraries.

In the backend, I use NodeJs or Java Spring Boot. For databases, I have experience with PostgreSQL or NoSQL databases such as MongoDB. I have been working with AWS as my cloud provider for 6 years (primarily with Lambda, Beanstalk, EC2, S3, and more). I have set up CI/CD pipelines in Github, GitLab, and Bitbucket and have maintained and optimized existing ones.

Furthermore, I have already supported many start-ups and have often been the lead developer, especially in the frontend. My biggest project was a B2B e-commerce shop for the veterinary industry, which I managed for over 4 years. This start-up began with a monolith project that used native JavaScript and NodeJs in the backend. I migrated the project to a microservice architecture and migrated existing frontends to Vue 3, as well as creating new microfrontends for various products. The Monolith Backend were extracted into multiple AWS Lamba functions or AWS Beankstalk services with the usage of Java Spring Boot. I developed NPM libraries in private scope to share Vue components and other JS functions across projects.

You can find other examples of my work on my LinkedIn profile or on my website, https://devnik.dev. If you have any further questions, we can discuss them in a phone call. Please also feel free to check out my CV or one of my public NPM libraries, which have ~9500 downloads per month. You can also contact me directly at niklas.grieger@devnik.dev.

I look forward to hearing from you. Best regards, Niklas Grieger

I like projects where I am able to make a difference, my expertise is welcomed, and I can contribute to better user and developer experiences

I'm working since 7 years with Vue, Nuxt, Vuetify and PrimeVue. I migrated 3 Vue2 projects to Vue3 incl. Nuxt and libraries like VueX -> Pinia and Vuetify. I am owner of the official certificates from certificates.dev. I used PrimeVue in my last 2 projects, and I'm working daily with TailwindCSS.

In the Backend, I'm building my Rest API's with NodeJs or Java Spring Boot. My experience with Java Spring Boot is 4 years.

https://github.com/devonikhttps://devnik.dev/

Created by Niklas Grieger • © 2025