I would love it if the pages I went to got stored somewhere on my computer locally, there was a set amount of space that was allocated for this function, and as soon as it was full older stuff would be deleted. Ideally if I go to a site like https://golang.org/ and I click around the documentation, I should be able to access all this content when I'm offline again. Why not?
Ideally I love the idea that I could just go straight to that url and act like I have internet. I've gotten close to doing this with sitesucker and editing my host file to create a kind of "intranet". But I never can remember the pages I visit, or capture all of them.
I can understand from a security perspective it might not be best to have the server out of sync with the domain. It seams like an easy way to get peoples Facebook passwords if you had access to the router. If that's the problem we could prepend the url like view-source:google.com
works in chrome and make it something like offline-cache:golang.org
, so golang.org might redirect to that if we're offline.
Shouldn't browsers be able to handle this? Does this exist already? If not is there a reason it doesn't?
Just found this amazing post by Addy Osmani titled Offline Browsing Secrets
Chrome has a flag to
Enable Offline Load Stale Button
All you need to do is go to
about:flags
in the address bar.