Shrinking MP4 for Wordpress - wordpress

I am trying to compress a video for wordpress, as each time I open up my webpage the video barely loads and then freezes. How should I go about compressing the video (I have already zipped it and used a program, but at 324kb it still seems too large). I have heard something about changing the bitrate, is this helpful/how can I do that? I would like to keep it in an mp4 if possible.

The only way to change the bit-rate of a video file is to re-encode it. There are plenty of software that are capable of doing so, my favorite being avidemux which is free and reliable.
Open your file in the app, choose an encoding & a bit-rate, hit "save video" and you're good to go.
You might have to try a few different bit-rates until you get a file that will both load fast and look good on you website.
Be sure to always use the highest-quality source file available for the re-encoding operation, since re-encoding your video will always result in a decrease of your video's quality.

Related

Which format is more optimal in site speed?

I had a site design project and at the request of my employer I designed a gif in above the fold and I report from Gtmetrix said that this format should be MP4! How much do you think the gif format affects the speed of the site?
you better not use GIF or MP4, try to use something like Lottie json, I think it's more optimized for page speed.
the Gif file download with page load and Mp4 don't, that means more size and more request to the server
for 10 second video mp4 size is smaller than Gif,

Slow Loading Images on WordPress - How do I speed them up

I want to have a website page load images faster or where the images don't stop the user from seeing the text on the pages before the images finish loading.
1. if I have text on a website page. Does the text render first so the user can see text while the images are still loading?
2. If i have 3 mb of images and have my server that is hosting the website render 1.5 and then move the other 1.5mb to something like flickr and have the image source point to flickr, will that be better and load faster. Or say I just have all the images pull from flickr?
Please help me.
Its a WordPress site, and Clouldflare free cdn does something weird to my theme. So not an option. I also don't want to lazy load for other reason.
EDIT : o yea, if I host all the images on Flickr does that give away link juice (Domain Authority Rank)?
Images do increase overall load time, but they aren't render blocking. The whole page will render (barring render-blocking CSS or Scripts) and images will pop in when they are done. If you want to avoid that flash, you can lazy-load them, or otherwise put the final image's sizes on a container element, so the text is already "moved".
You can't really "Speed Up" image loads. The best thing to do is cut down on the number of images, if applicable, and lazy-load any that don't need to be requested initially. The next best thing (and arguably more important overall, I suppose) is to optimize your images. There are a handful of WordPress plugins that will do this for you, or you can do it in PhotoShop - and even some image CDN's will do it. This basically means, don't load a 1MB image if you don't have to, for "web display" purposes, you can serve a 200kb image that looks almost as good.
A side point, you can "speed them up" a little by having them served from a faster server, but that goes for anything web related. Throw it on better hardware to get better performance (for the most part).
If they are large images, you should first and foremost, optimize them. Make sure they are the smallest file size possible. Also, don't bother splitting "1.5mb" of them over to another host. Either upload them directly all through a CDN (some even tie into your WP Library, like Cloudinary), or keep them local to your site. You'll save yourself some headaches later.
If you have a photo heavy website (photography or other types of galleries), 3mb isn't really an awful lot - but again it's best to serve a more compressed image, you can even link the image or a button to the "full resolution" one.

Google Doodles archive scraping

I am trying to download all (about 2000) of the Google Doodles (not .gif or interactive ones) from their archive, however I have been having no luck in doing so.
I have tried many scraping commands for terminal, and a few packages from Github, such as ImageScraper. Through each of the different ways I have tried, I found that either:
the process just gets stopped after about 100 images using this chrome extension
it downloads only some sample images (most command-line tools did this)
just downloads a couple of images and did not keep going
The archive itself is not indexed from what I can see and each image has its own name, which also makes it harder. To keep viewing more images, the page also needs to keep being loaded as you scroll further down (causing the issue of only downloading a few images). I am also worried that the connection or something may be being cut-off so that these cannot be all downloaded at once, probably to not cause overload (not sure how to fix this though).
I would very much appreciate anyone's help, who has experience in retrieving/scraping such images, given the mentioned problems.

How is this site using a hi-res video as a background?

http://freelance.tv/
Seems like this video is too high-resolution to be downloaded that quickly — what is the trick being used here?
The video does not seem to be higher than 720p, once I opened the page the video stopped a few times to buffer but with high-speed internet, it will load fairly quickly.

How does YouTube prevent video content from being saved/redistributed?

Sure, you can embed a YouTube video on any site, but the content ultimately must come from their server. What technology(ies) do they have that prevents us from saving/redistributing content?
From a protocol standpoint, you would think that anything that comes over the wire could be saved. I hope I am not the only guy on Earth who does not know how to "save" a YouTube video...
There are a couple of plugins for Firefox out there that let you save the content. Basically it parses the sourcecode and looks for the videofile (either .flv or .mp4) and downloads that directly. The flash player on the page just plays the supplied file. They could of course obfuscate the path to the video file, but that can be reverse engineered as well. They can't really do anything about it, because the video file has to be on the user's computer at some point, or if not, the stream could be intercepted as well.
eg. https://addons.mozilla.org/de/firefox/addon/6584/?src=api
Mostly it's a legal deterrent rather than technical. There are a plethora of programs out there that will allow you to download their video. But there are two things they do that help reduce unauthorized downloads:
Use is flash to control the download and playback.
Hosting video yourself is not cheap, and thus it's much easier to simply leave the video on youtube.
They don't do anything about it. Very likely your Flash viewer downloads a copy and puts in somewhere on your harddrive (under my Linux system with Firefox and Adobe Flash in /tmp). After you are done viewing the file is removed to save disk space, but since it is on your harddrive nothing prevents you from making a copy elsewhere.
You might want to look at the 'analogue hole', in the end, data still has to be displayed on your screen, or get through your speakers and what not. It's always theoretically possible to intercept it at that point, or even just record your audio-out into another machine.
So as far as the analogue hole goes, the only solution is to skip that, in this form:
(source: thisdomainisirrelevant.net)
Which is not that marketable.

Resources