rss2json with Medium feed doesn't always work. "probably not a valid RSS feed" - rss

I use rss2json.com for my Medium posts.
https://api.rss2json.com/v1/api.json?rss_url=https://medium.com/feed/#username
works fine, but sometimes the response is
Feed could not be converted, probably not a valid RSS feed.

got same problem for my website. It used to work perfectly fine until 2 or 3 days, everything broke. I couldn't use the rss2 to fetch them anymore. I had to use a lambda on aws to get the feed there, parse it and send it as response. It seems aswell some property have changed, and the thumbnail/pubDate properties are gone as well. Im surprise not more people are reporting this issue.
The only thing to do is set up your own fetch/parsing from a server or a quick workaround that i found, but not spotless is to use this:
https://www.convertjson.com/cgi-bin/url-to-json.php?callback=loadDataAndRun&url=yourMediumFeedUrl
It's a UI online service that allow you to parse from an URL,not sure if they allow this usage, but i had to find a quick solution to fix it until i set up my own server to do this job.

Related

Linkedin SlideShare API "get_user_leads" 500 Internal Server Error and 410 Gone Error on apiexplorer.slideshare.net

Looking for help from a Linkedin SlideShare engineer on the SlideShare API here. Very frustrated that I was told to use StackExchange after being kicked all over the place by them and now I can't post enough detail (personal account info would be needed and StackExchange is limiting me to 2 links in this message).
Anyhow, I’m trying to install the SlideShare-Marketo Connector (http://launchpoint.marketo.com/assets/Uploads/SlideShare-MarketoConnector-Guide.pdf) on to an ancillary server. I’ve uploaded the PHP files just fine.
The expected output from my page should be “X leads synced from SlideShare.” (where “X” is a number), but instead I get a blank page. I added some echos to see if I could figure out the last spot the code was executing to. I found that it’s getting hung up in SSUtil.php at this line in the function “get_userLeads”:
$data=$this->LeadXMLtoArray($this->get_data("get_user_leads","&username=$user&password=$pass&begin=$begin"));
From what I can tell though the issue isn’t really with this line but when the get_data function tries to get the data at this line:
$res=file_get_contents($this->apiurl.$call."?api_key=$this->key&ts=$ts&hash=$hash".$params);
I echoed the URL to the browser see what it was looking for:
http://www.slideshare.net/api/2/get_user_leads?api_key=XXXXXXXX&ts=XXXXXXXX&hash=XXXXXXXX&username=XXXXXXXX&password=XXXXXXXX&begin=201603311600
Obviously I can't make this link clickable here without give away a bunch of account information but when I click on a real version of the link I get a 500 Internal Server Error.
I used apiexplorer.slideshare.net (but now it seems SlideShare has taken this down in the last day) and the URL it’s using looks the same as what I’ve got above but has a slightly different result: 410 Gone Error. Any idea what’s going wrong?

I get HTTP 500 and HTML content, what's wrong?

When I got on this page (same with lots of articles on this website) : http://thereasonmag.com/9231-2/
I get an error HTTP 500 (see in the Chrome Dev Tools) AND the article.
Well, I'm a bit lost with this. Do you know why it is designed like this ?
That's a problem for my crawler which is designed to avoid processing HTTP 5xx error responses.
I would say that this is hardly can be called "designed" and possible when somebody has an error in backend code/logic. Actually this is the first time I see anything like this, but I can think only of workaround for you in this case.
Because this response has 500 error AND correct HTTP body with html, you can avoid in your code processing 5xx error WITHOUT body with correct html.. How to determine if this html is correct? This is pretty risky to guess.. You can research their html and find some global variables or some comment tags/classes in html which won't be returned if real error page is returned.
Important: I understand (and sure you too) that my suggestion is absolutely crazy workaround just to make your code work. What I would do in your place, I would write those guys and ask them to fix their backend. Seems like this is the only place with some email at the bottom..
Try to write them, otherwise you will definitely face a case, where you will fail to meet the criteria of if (res.errorCode === 500 && res.body.anyPossiblePredicateYouMayThinkToCheckRightHTMLBody) {// show the post on your page }
1) Looks it is an expected behavior since PHP version 5.2.4.
2) The above url is using X-Powered-By: PHP/5.4.45 (wordpress app)
3) root cause could be,one of the wordpress plugin in the above site is having
wrong string thatt ph eval() could not parse it.
4) more info look at the link a) wordpress discussion
5) from ph forum
Finally, i don't think so you can do anything here.

Can anyone provide a good info on the various uses of hash(#) in urls?

I'm developing a software, which is going to provide in-deep information about url's.
While the get-params are simple, I'm having trouble with the hash.
At first it was used to mark places in the document to navigate to, but we're past that now. I've seen JS engines using it to store params similar to the get strings.
So, here's my question: is everything that comes after a hash free game, or are there any conventions about what it should look like?
Try these sites it could help. Fragment Identifier, Wikipedia or Pound Sign, Google
It's got a list of examples you could use.
It all depends on what you need. Hashes are used in modern web applications that make use of asynchronous calls to the server using ajax. This e.g. allows the user to copy the link and receive the same content after pasting (actions taken are put into hash which changes the url which otherwise would remain static).
You want to read http://www.jenitennison.com/blog/node/154

Bugzilla: How to get an rss feed for bug comments?

I can see where to get an rss feed for the BUG LIST, however I would like to get rss updates for modifications to current bugs if possible.
This is quite high up when searching via Google for it, so I'm adding a bit of advertisement here:
As Bugzilla still doesn't support this I wrote a small web service supporting exactly this. You can find its source code here and a running instance here.
What you're asking for is the subject of this enhancement bug:
https://bugzilla.mozilla.org/show_bug.cgi?id=256718
but no one seems to be working on it.
My first guess is that the way to do it is to add a template somewhere like template/en/default/bug/show.atom.tmpl with whatever you need. Put it in custom or an extension as needed.
If you're interested in working on it or helping someone with it, visit channel #mozwebtools on irc.mozilla.org.
Not a perfect solution, but with the resolution of bug #255606, Bugzilla now allows listing all bugs, by running a search with no criteria, and you can then get the results of the search in Atom format using the link in the bottom of the list.
From the release notes for 4.2:
Configuration: A new parameter search_allow_no_criteria has been added (default: on) which allows admins to forbid queries with no criteria. This is particularly useful for large installations with several tens of thousands bugs where returning all bugs doesn't make sense and would have a performance impact on the database.

Access to old, no longer available, feed entries

I am working on a project that requires reliable access to historic feed entries which are not necessarily available in the current feed of the website. I have found several ways to access such data, but none of them give me all the characteristics I need.
Look at this as a brainstorm. I will tell you how much I have found and you can contribute if you have any other ideas.
Google AJAX Feed API - will limit you to 250 items
Unofficial Google Reader API - Perfect but unofficial and therefore unreliable (and perhaps quasi-illegal?). Also, the authentication seems to be tricky.
Spinn3r - Costs a lot of money
Spidering the internet archive at the site of the feed - Lots of complexity, spotty coverage, only useful as a last resort
Yahoo! Feed API or Yahoo! Search BOSS - The first looks more like an aggregator, meaning I'd need a different registration for each feed and the second should give more access to Yahoo's data but I can find no mention of feeds.
(thanks to Lou Franco) Bloglines Sync API - Besides the problem of needing an account and being designed more as an aggregator, it does not have a way to add feeds to the account. So no retrieval of arbitrary feeds. You need to manually add them through the reader first.
Other search engines/blog search/whatever?
This is a really irritating problem as we are talking about semantic information that was once out there, is still (usually) valid, yet is difficult to access reliably, freely and without limits. Anybody know any alternative sources for feed entry goodness?
Bloglines has an API to sync accounts
http://www.bloglines.com/services/api/sync
You have to make an account, subscribe to the feed you want to download, but then then you can download based on Date, which can be way in the past. Not sure of the terms.
The best answer I've found so far, is this: Google reader's unofficial API turns out to have a public access point for their feeds, which means there is no authentication needed. Use is as follows:
http://www.google.com/reader/public/atom/feed/{your feed uri here}?n=1000
replace the text in the squigglies (including the squigglies themselves) with the feed URI you're interested in. More information about the precise arguments can be found here:
http://blog.martindoms.com/2009/10/16/using-the-google-reader-api-part-2/
but remember to use the /public/ url if you don't want to mess with authentication

Resources