Just curious (maybe this is an implementation detail so you may not know) but what are all these _1* chars are for in a LinkedIn link? When you start going to more profiles it gets longer and longer...
E.g.
&authType=name&authToken=FGH&goback=%2Efmp_*1_*1_*1_*1_*1_*1&trk=pbmap
The additional information in the URL is used to track what pages you've visited, in what order, so LinkedIn can better structure the site in a way that gets people to the information they're looking for.
Related
I had a look at LinkedIn's API and there are very few POST methods. I was wondering if there is a post method where I can actually edit the details, profile picture, edit intro, edit experience... sections, with my own application.
If not, I was wondering if this is possible through scraping or any other methods?
You can edit entire profiles, add and remove almost every element but you will need to be part of the LinkedIn Partner program.
https://developer.linkedin.com/docs/v1
One of the purposes of og:url -- I thought -- was that it was a way you could make sure sessions variables, or any other personal information that might find its way into a URL, would not be passed along by sharing in places like Facebook. According to the best practices on Facebook's developer pages: "URL
A URL with no session id or extraneous parameters. All shares on Facebook will use this as the identifying URL for this article."
(under good examples: developers.facebook.com/docs/sharing/best-practices)
This does NOT appear to be working, and I am puzzled as to either -- how I misunderstood, and/or what I have wrong in my code. Here's an example:
https://vault.sierraclub.org/fb/test.html?name=adrian
When I drop things into the debugger, it seems to be working fine...
https://developers.facebook.com/tools/debug/sharing/?q=https%3A%2F%2Fvault.sierraclub.org%2Ffb%2Ftest.html%3Fname%3Dadrian
og:url reads as expected (without name=adrian).
But if I share this on facebook -- and then click the link. The URL goes to the one with name=adrian in it, not the og:url.
Am I doing something incorrectly here, or have I misunderstood? If the latter, how does one keep things like sessions variables out of shares?
Thanks for any insight.
UPDATE
Facebook replied to a bug report on this, and I learned that I indeed was reading the documentation incorrectly
developers.facebook.com/bugs/178234669405574/
The question then remains -- is there any other method to keeping sessions variables/authentication tokens out of shares?
I'm not a really advanced Analytics user, so I've been trying to Google this, but haven't come up with a great answer. My analytics says 95% of my site visits to my blog today have come from site38.social-buttons.com and yesterday it was another subdomain of the same site. I visited social-buttons.com, but am unfamiliar with it, and have never deliberately put that code into my Wordpress site. I do have some plug-ins installed, which are "Subscribe / Connect / Follow Widget", which displays my social media links, and also "Really simple Facebook Twitter share buttons", which puts the like links on my posts.
My questions are, how are people finding my site through social-buttons.com? And are these quality hits?
Thanks, I appreciate any info!
This kind of visits are called Ghost Referrer Spam since they never reach your site. They use a GA weakness to make a fake visit and get a record in your data.
They do it to get traffic, people get curious to see who is visiting them and click on the link.
This specific Referrer Spam is nasty because it make multiple visits at the same time, is related to the number of the subdomain so if it says site38... it hits with 38 visits, I've also have many of these, here is a screenshot I took:
In my case is a different simple-share-buttons.com but is the same thing.
The easiest way to stop it is by making a filter for each spammer in your GA. Check this article to find more detailed information http://www.ohow.co/block-social-buttons-simple-share-buttons-referral/
As an alternative, you can make a more general filter to take care once and for all of all the Spammers by making a list of Valid Hostnames, this is more advanced and you have to be more careful. You can find more information about this solution here https://stackoverflow.com/a/28354319/3197362
It's actually referral spam. Take a look at this https://www.mooresoftwareservices.com/Web-Commerce/social-buttons-com-referrer-spam
So unfortunately they are not good quality hits.
Is it possible to track if someone links to data on my site? Specifically if my data is used in a site dynamically generated by a developer program? I would like to know if someone is blatantly passing off my site's data as their own. There are obviously ways around directly linking to content, such as content manipulation or even manual manipulation. But if someone where to link(or directly add word for word or manipulate) my content into their website, is there a way to track it?
Can I avoid someone being able to scrape my website at all, or is everything just up for grabs?
the best answer and the easy one is called GOOGLE - WEBMASTER TOOLS!
HERE
actually doing that is very hard and you would need to crawl the web to discover those links that address to your pages... dynamic content as well is linked so it would be find by google as well.
this tool will allow you to see outer links that address to your site.. and you can check them.
for extra - you can monitor requests and traffic to your site and find ip's that are using the same page over and over again. that can tell u that an outer page is dynamically loading content from your web page.
EDIT:
here is a good article in this subject: link - scroll down and you can see the use of google
webmaster tool with some other progrmas and method.
here is a good start guide to the google webmaster: link
ENJOY!
I'm doing custom-rolled view tracking on my website, and I just realize that I totally forgot about search bots hitting the pages. How do I filter out that traffic from my view tracking?
Look at the user-agents. It might seem logical to blacklist, that is filter out all the strings that contain "Googlebot" or other known search engine bots, but there are so many of them, it could well be easiest to just to whitelist: log visitors using a known browser.
Another approach would be to use some JavaScript to do the actual logging (like Google Analytics does). Bots won't load the JS and so won't count toward your statistics. You can also do a lot more detailed logging this way because you can see exactly (down to the pixel - if you want) which links were clicked.
You can check the user agent: here there is a nice list.
Or you could cross-check with the hits on robots.txt, since all the spiders should read that first and users usually don't.