top | item 45991754

(no title)

zulko | 3 months ago

Total plug but this year I scraped 400,000 wikipedia pages with Gemini to create landnotes.org, an atlas where you can ask "what happened in Japan in 1923":

https://landnotes.org/?location=xnd284b0-6&date=1923&strictD...

https://github.com/Zulko/landnotes

My plan has been to overlay historical map borders on top of it, like the Geacron one from this post, but they all seem to be protected by copyright - and understandably so, given the amount of work involved.

discuss

order

milst|3 months ago

very cool. Made something with a similar idea, but using timelines instead of maps. I wonder if the two could be combined in some way https://timeline-of-everything.milst.dev/

zulko|3 months ago

Nice, how does your timeline work under the hood? Does it read from wikipedia? What could be interesting in your project is to be able to compare timelines. See for instance this website specifically for comparing composer works (with timelines pre-extracted from wikipedia):

https://zulko.github.io/composer-timelines/?selectedComposer...

llbbdd|3 months ago

This is very very cool! I went right to the month and year of my birth; kind of the same vibe as finding a newspaper published on the day you were born but all over the world. Thanks for sharing!

pu_pe|3 months ago

This looks pretty cool actually, nice job!

lippihom|3 months ago

Wikipedia doesn't have an API?

zulko|3 months ago

It does, why?

annodomini2019|3 months ago

Wow, this is actually so cool. Fantastic idea, I would LOVE something like this in Wikipedia. Nicely done!

zulko|3 months ago

Yeah it would be nice if Wikipedia would host it, but it would probably require some more serious ground work so the project fits in the wikipedia ecosystem. Could be a pipeline Wikipedia -> Wikidata -> Atlas.

There are many projects that could be done with with wikipedia and LLMs, for instance "equalizing" all languages by translating all pages into all other languages where they are missing. Or, more surgically, finding which facts are reported in some languages of a page but not others, and adding these facts to all languages.

For now, it seems that wikipedia doesn't want to use generative AI to produce wikipedia pages, and that's understandable, but there may be a point where model quality will be too good to ignore.

qq66|3 months ago

Cool project. Seems like your link to "wiki-dump-extractor" is broken.

zulko|3 months ago

Thank you for reporting this!