The goal is to parse, for example, gdata.youtube.com/feeds/api/users/FoxloidShiroko/uploads for links to put into the scroll area in utaite.wikia.com/wiki/Shiroko. As a starting point, we can take my Python code at tinyurl.com/aq2pe6a.
This task should be extremely simple because we only have to split out and format a few fields. There's no need to use the re library or the alternate references for this task.
1. It also be feasible for basically anyone for whom we don't yet have a list for, or for recently added singers.
2. There also is an API for playlists. See https://developers.google.com/youtube/2.0/developers_guide_protocol_playlists and the urls discussed there.
With the python code listed above, you have information about how to grab html from online pages. And to help you out with the string processing, you can use the guide at http://python-string-processing-tutorial.herobo.com/
Good luck getting this done. Don't be discouraged. This is a simple task and I'll help you out if you'd like.
Sorry for being dumb, but how will it work for utaite whose works are not compiled in any single playlist and may have covers that are not reprinted? Reprints for any one utaite are not guaranteed and have multiple sources.
You might have to run the program on multiple playlists, and merge them (while removing duplicates) by hand after getting the output. Since there aren't any easy automated ways to recognize duplicates, I wouldn't recommend attempting to automate this unless it gets extremely tedious.
To some extent. Small updates can be done my hand. But when you want to start from scratch, it would be less time-consuming and less tedious to use a program.
Edit: I see what you mean. You want to see what NND's API is like. Here are some examples: http://www.nicovideo.jp/user/9193089/video?rss=2.0 and http://www.nicovideo.jp/mylist/9258972?rss=2.0. Just add "?rss=2.0" and you get an rss feed that you can try to process.
By the way, this might be useful. "[%s %s] (%s)" % (url, description, date) gives you the string in quotes, with each %s replaced by the corresponding variable inside the parenthesis.
And text.split("<a href=\"").split("\"") ordinarily gives you the url of the first link on a page.
Oh so you can make a bot for NND links then? That might be super useful. I'll try to look more into this when I have time but even then, when I had tried researching on how to create/program a bot everything just went over my head
i wonder if this will work out well, or if this will only make more work, e.g. bc of some random code inserted or double/wrong linking
This will reduce some of the work, but you will still have to check the links by hand. It is probably impossible for programs to check for duplicates, so this cannot be a bot. The most this can be is a program that only writes part of the wikicode.
The benefit of having a program write part of the wikicode for your is pretty clear though. You don't have to focus on copying links or titles into the correct syntax. And you may not have to focus on getting the date.
i see... then i guess it's pretty heplful. esp. for youtaite, since i think completing youtaite playlists are more tedious than utaite playlists
however, i understood it this way:
if on a page the code is like this
# "World is mine" (XXXX.XX.XX) # "magnet" (XXXX.XX.XX) # "whatever" (XXXX.XX.XX)
we can use this to make it like this
# "[youtube.com/watch?v=XXXXXXX World is mine]" (2012.12.12) # "[youtube.com/watch?v=AAAAA magnet]" (2010.12.01) # "[youtube.com/watch?v=SSSSS whatever]" (2005.04.02)
but if the code varies like this
# "Suji Chigai" (Misdirected) (parody of [[Suki kirai]]) -Short ver.- feat. Faneru and [[kogeinu]] (2012.05.04)
(ok this is an extreme example) i wonder, if it can deal with it
and another thing is: can it convert the YT date format to the XXXX.XX.XX format, i wonder?
(sorry, i never dealt with bots/python codes)
What do you think?