Freshness of a YouTube API channel playlist feed? - rss

I am writing an application to read the updates happening in khan academy youtube playlists, by parsing their feeds, but the feed is not upto date
According to this feed url, the last update was on Aug,2012, how to get the latest feed?
http://gdata.youtube.com/feeds/api/users/khanacademy/playlists?v=2
Please tell me if there is anyway to get a new feed
Reg,
Karthi

The youtube method or organizing videos seems like something Khan academy has abandoned at some point in favor or organizing the lists separately. That is why newer videos don't appear in previous lists you would think they would. They now have a JSON returning RESTful API. For example to see the playlists visit this. There is a wiki entry explaining more on how to consume these lists and example client code is available on their github channel for a deeper direct use or code dive to emulate in your code.

http://gdata.youtube.com/feeds/api/users/khanacademy/playlists?v=2 returns a list of the most recent 25 playlists that are in the "khanacademy" channel, and that's what I see being returned. It matches what you see in the web interface at http://www.youtube.com/user/khanacademy/videos?shelf_index=5&sort=dd&tag_id=&view=1

Related

Google news topic url parameter list?

I'm currently trying to build a news reader that gets rss feeds from different google news topics and I found this website that breaks down the url parameters pretty well, but under the topic parameter, the link that it has for a list of topics brings you to this page which says nothing about topic keywords. The only topic key that I've been able to find is tc being the key for the tech topic.
Does anybody know where I could find a list of these topics or if the topic=[topic] parameter is deprecated what I could use instead?
It will depend of country.
Example (for US):
WORLD
NATION
BUSINESS
TECHNOLOGY
ENTERTAINMENT
SPORTS
SCIENCE
HEALTH
RSS business (US)
https://news.google.com/news/headlines/section/topic/BUSINESS?ned=us&hl=en
But, if you query other country, the topic will change.
Example (Colombia):
WORLD.es_co
NATION.es_co
BUSINESS.es_co
TECHNOLOGY.es_co
ENTERTAINMENT.es_co
SPORTS.es_co
SCIENCE.es_co
HEALTH.es_co
Here is the RSS about economy/business in Colombia:
https://news.google.com/news/rss/headlines/section/topic/BUSINESS.es_co/Econom%C3%ADa?ned=es_co&hl=es
Check too the ned and hls parameters added.

Google reviews counter

I want to know if there is any api that can allow me to get the number of reviews from an url.
I know that google offers the possibility to get this number by using the placeid, but the only information I have is the url of the website of a company.
Any ideas please?
Maybe, but probably not.
Places API Text Search seems to be able to find places by their URL:
https://maps.googleapis.com/maps/api/place/textsearch/json?key=YOURKEY&query=http://www.starbucks.com/store/1014527/us/303-congress-street/303-congress-street-boston-ma-02210
However, this is not a documented feature of the API and I do not think this can be relied upon, so I'd recommend filing a feature request, to make this a supported, reliable feature.
As for the amount of reviews, you may be interested in:
Issue 3484: Add # of reviews to the Place Details Results
I've written an API like this for Reviewsmaker, but I target specific business names not URLs. See this example (I activated a key for this purpose for now):
http://reviewsmaker.com/api/google/?business=life%20made%20a%20little%20easier&api_key=4a2819f3-2874-4eee-9c46-baa7fa17971c
Or, try yourself with any business name:
http://reviewsmaker.com/api/google/?business=Toys R Us&api_key=4a2819f3-2874-4eee-9c46-baa7fa17971c
The following call would return a JSON object which shows:
{
"results":{
"business_name":"Life Made A Little Easier",
"business_address":"1702 Sheepshead Bay Rd, Brooklyn, NY 11235, USA",
"place_id":"ChIJ_xjIR2REwokRH2qEigdFCvs",
"review_count":38
},
"api":{
"author":"Ilan Patao",
"home":"www.reviewsmaker.com"
}
}
Pinging this EP using a Chronjob for example once every hour or two and return the review_count can pretty much build your own review monitoring app;
You can probably do what you're looking for if you query the Places API Text Search or the CSE (Custom Search Engine) API to lookup the URL, return back the matching name of the business associated with this URL and calling an endpoint like this one to return back the associated review count.
You can probably code this in py or PHP. Not sure how familiar you are with data parsing, but I was able to build my API based on Google's CSE API. CSE provides metadata in its results which contain the total reviews, so if you create a CSE engine and use the CSE API looking for business schemas, review schemas, etc; you can return back items and within the PageMap node there are objects with data that you need very little tweaking to do (such as string replacing, trimming) which will return back the values you're looking for.
Hope my answer helped, at least to lead you in the right direction :)

how can I get a global rss feed on Gitlab?

In Gitlab there is an RSS news feed button for every project.
http://git.domain.name/userName/projectName/commits/master.atom?private_token=xxxxxxxxxxxxx
Is there a way to retrieve a global RSS feed for all projects related to one user?
Or even better, if there is a way to capture all user's feeds across all projects, that would be cool.
thanks
This is currently not supported.
The closest feature request there is would be "Team-/groupwide RSS feeds", to have RSS feeds for all events in a team or a group. (not enough vote for now).
The global "RSS" shown on demo.gitlab.com is actually a html result of the event-items representing the activity in response to http://demo.gitlab.com/?limit=20&offset=0:
That would be for all public projects, but I don't see it exposed as an RSS feed.
The OP mauro reports in the comments
the global "RSS" shown in demo.gitlab.com is also showing on my gitlab dashboard, the only thing is that mine is giving a 500 error instead.

RSS for Future Items

This may be a simple question, but for some reason I don't know this answer. Is it possible to create an RSS feed file that contains contents for an entire year but only publishes the current date and previous date information?
I have a client that wants to do a "this day in history" post. Currently, I am using IFTTT, and created around sixty dated posts for the next two months. Of course, this works -- but it is very labor intensive.
Is it possible to create an RSS feed that you could put all 365 days of data in to, but if someone pulls up the feed it only shows today's item and prior days in the feed?
Or is RSS not the proper technology to do this? The reason I am using RSS is for ease of use, and IFTTT will take those RSS feeds and pump it in to Facebook and Twitter for automatic status updates for my client.
There are various tools that let you define Facebook and Twitter posts in advance, to be published at a specified date and time in the future. Why not use one of those instead of writing your own?
A quick search for "scheduled twitter post" uncovered Later Bro, Twuffer and twAitter but there must be dozens to choose from.
If you're looking for just posting on Facebook and Twitter, and not an RSS feed as well, I'd follow Matthew's suggestion. If you want an RSS feed, there is a feed for each Twitter feed. But if you want actual RSS, you need to add something in between. An RSS feed is just an XML file. it's not a process. I suggest having a file of some type (maybe RSS, or other XML, or a database table, or even a csv file with all the posts and relevant information, including date. Then a small script that runs as a chron job (or IFTTT if it supports date as trigger and running a script as the "then" part) that pulls the day's feed and updates the actual RSS feed. Pretty simple.
Here is what I ended up doing
Using the Drupal backend of my website, I created a content type specifically for these posts.
I created individual articles for each day, and used the schedule module to schedule the publish date to the date I wanted.
I created an RSS feed of these posts through Drupal.
I linked the newly created RSS feed to IFTTT.
Created an IFTTT recipe to post the text from the RSS feed to Facebook/Twitter/etc.
It wasn't the best solution, but it worked. I was really trying to do this without having to rely on a third-party such as IFTTT, but never really figured out a good way to do it.

Retrieving RSS posts older than those included in feed

When creating an RSS reader, you download the XML formatted document pointed to by the RSS feed link, and you can parse it manually or using the functionality in the SyndicationFeed namespace.
So if we take Scott Guthrie's blog as an example, you download the RSS feed document here, and parse it. My problem is that this document only holds 15 items, yet he has been blogging for a number of years.
Is there a standard or established way of getting the older posts not included in the RSS feed document? Or do you have to find the base address for the blog posts and then parse the pages of the site from there to get them? How do you avoid missing posts on high volume blogs?
With RSS/Atom you can't query older articles.
I built a RSS archival service (https://app.pub.center). All of our data is free to use via REST. We charge money for push notifications.
PubCenter daily polls it's catalog of RSS feeds, and caches the articles. Then, you can get these articles back in a chronological order. For example:
Page 1 of The Atlantic https://pub.center/feed/02702624d8a4c825dde21af94e9169773454e0c3/articles?limit=10&page=1
Page 2 of The Atlantic https://pub.center/feed/02702624d8a4c825dde21af94e9169773454e0c3/articles?limit=10&page=2
As the replies to How Do I Fetch All Old Items on an RSS Feed? already mentioned, a feed may not provide archival data but historical items may be available from another source.
Archive.org’s Wayback Machine has an API to access historical content, including RSS feeds (if their bots have downloaded it). I’ve created the web tool Backfeed that uses this API to regenerate a feed containing concatenated historical items. If you'd like to discuss the implementation in detail please get in touch.

Resources