Sort by: Newest, Oldest, Most Relevant
(#37xr3ra) As a quick experiment, I modified my code to remove the domain restrictions and low and behold: ``` All done! Crawled 516 feeds Found 52464 twts Found 736 feeds ``` The Twtxt network is larger than I thought. A significant no. of feeds no longer work obviously, but that's okay, we can prune dead feeds out.

matched #4vuxgtq score:11.2 Search by:
Search by 1 tags:
(#37xr3ra) @lyse @prologic very curious... i worked on a very similar track. i built a spider that will trace off any `follows = ` comments and mentions from other users and came up with: ``` twters: 744 total: 52073 ```

matched #6ltxv6q score:11.2 Search by:
Search by 2 mentions:
Search by 1 tags:
(#37xr3ra) Wait... So you actually wrote a more elaborate crawler without taking a shortcut like I did using colly (_not that it really helps much_) Hmmm? ๐Ÿค” Can we take it a bit further, make a daemon/server out of it, a web interface to search what it crawls using bleve and some tools (_API, Web UI_) to let people add more "feeds" to crawl? ๐Ÿค”

matched #acie3ca score:11.2 Search by:
Search by 1 tags:
(#37xr3ra) @prologic yeah it reads a seed file. I'm using mine. it scans for any mention links and then scans them recursively. it reads from http/s or gopher. i don't have much of a db yet.. it just writes to disk the feed and checks modified dates.. but I will add a db that has hashs/mentions/subjects and such.

matched #e5kep3q score:11.2 Search by:
Search by 1 mentions:
Search by 1 tags:
This is twtxt search engine and crawler. Please contact Support if you have any questions, concerns or feedback!