You can use this program and download parts of the site to your computer, it'll run fine.
You will have to copy the URLs for a lot of pages, and put them in the list to download.
There's a program called HTTrack. It crawls a website and downloads everything from it. I used it on a2b2.org and there's literally nothing there that we haven't seen.
https://www.httrack.com/ if you don't believe me and want to try it. Pretty much what I got from it today was a couple of image files and the html. Not much more than that from what I saw.
If you want a decent local backup of your stat pages
Use this site download the program and put as many URLs on your stat page as you can, in the list it provides you.
You can also right click any page and save it as a HTML, it'll be a local back up.
You can also use wayback archive to save everything. But you'll have to go one page at a time.
Just use Httptracker.
I have the whole website in my computer (and my phone because yes there is an app too) in case I don't have the internet (no videos though)
It's just a download-and-click type of software/app so it's pretty straightforward to use.
Can someone save an offline version of the sub to their harddrive as a back up using HTTrack Website Copier? Then in case the sub got banned we wouldn't loose all the data. I'd do it but my hard drive is almost full and I'm not quite sure which settings to chose in the program.
If you don’t have server access, you can still download it but you’ll lose any CMS functionality (database, admin ability and frontend features like search). You need server access if you want to keep the CMS functioning.
there is httrack
, I have not tried it in ages however.
the fully download any site
requirement may be impossible, given how complex some sites can get.
good luck.
seems it has a gui front end also.
+1 for https://www.httrack.com.
--
It really isn't too complicated. Show him how to backup a few websites first. It is always the same repetitive process; I suspect he would get after a few tries.
There is a few tools out there which work with various success as it depends on the web site and the code it uses. The common name for the programs is a web spider
I have used https://www.httrack.com/ in the past, but just google "web spider download" and you should find a heap.
HTTTRACK should work fine in this use-case if the files are directly linked to the webpage directly and you have set the filters correctly. But if you have something like YouTube, where they use swf files then I think you can only do them one by one. If there is a better way to to about it? No idea
Some useful links
https://www.httrack.com/ to download a webpage, all the sub-pages, etc.
You have to go into "Set options" below where you enter the URL, go "Scan Rules" then "Includes links" and add "File names with extension" "pdf" so it downloads the PDFs as well. Same goes for ppt and pptx files.
you can save the website for yourselves with https://www.httrack.com/
when i downloaded the site it is 550MB is size for me. its pretty slow to go from page to page but hey great for archiving and you can browse the website while offline. First time using it and it was really easy.
I play Dungeons and Dragons, unfortunately Wizards has been sending cease and desists to websites with D&D 5e content on them (usually spell compendiums that are easier to navigate than the book).
So i use httrack to make a perfectly functional offline copy of these resources before they get taken down.
"HTTrack is a free (GPL, libre/free software) and easy-to-use offline browser utility.
It allows you to download a World Wide Web site from the Internet to a local directory, building recursively all directories, getting HTML, images, and other files from the server to your computer."
The rest of the thread explains that WB doesn't want epubs circulating... but there are a few generic ways to make them:
If you just want all the content offline (so you don't have to have data on), you could use https://www.httrack.com/, which is a generic solution for any website. It's a bit complicated to use, but it'll give you basically any website you want, frozen in time at the moment you download it. It works nicely on a computer, but I haven't tried moving a site it downloads to a phone and opening it from a phone browser. The phone browser wouldn't need to have data turned on, but it'd still be in the browser, not in an e-reader format.
I've not used this, but google suggests http://www.bloxp.com/ can give you an epub from any blog, and then you could definitely transfer that to your ereader.
HTTrack can be used to crawl/make static copies of websites — thereby creating a current snapshot of the accessible contents.
PhantomJS is a headless WebKit and can be used to render screen captures if that's your kind of (visual) snapshot.
from a professional point of view I would tell you to start again, but as a possible solution, you could download it with this https://www.httrack.com
but bear in mind, from that point onwards once you’ve removed the old site on wix, you will have to manually edit the html if you want to change anything.
I had no idea what that was, so I found this utility and used it to copy her website to my hard drive.
Any way to do the same to all her posts? She posted so much good info.
Powershell has wget, but it is VERY janky, so I just installed the Windows Subsystem for Linux (guide) and then use the Linux wget, which works well. I've also tried installing a third party wget for Windows, accessible via the command line, but that straight up didn't work.
If you want a GUI option, take a look at HTTrack (guide on how to use here). I've used that to download all of BBC Goodfood just before they shut it down, along with all the North Korean websites when they were temporarily online.
I'm guessing that includes the official wiki as well??
HTTrack can download entire sites, and should grab all the files. It'd probably require a fair chunk of drive space, but could be a simpler way of grabbing the content (I've been playing around with HTTrack and the forum)
When people web scrape they are typically looking to extract data. In which case they would use something like BeautifulSoup for python.
But you're talking about downloading and keeping entire pages. Something like HTTrack would be more appropriate. No scripting is necessary here.
https://www.httrack.com/ will scan & pull static copies of everything. It may take a little getting familiar with but would do the trick.
Right now I'm using a tool called HTTrack to try and download most of the site data into a locally browsable version. Once it's done, I plan to make a torrent out of it that can be shared around for members of the community to download and seed. The only ways that I can think of to help with that would be to make sure that your hosting service doesn't take exception to the entire site being traversed by an automated script, give a rough estimate of the total size of the site on the host server's storage, or spread information about the archive and where it can be gotten from. The threads on /r/DHExchange (here) and /r/Archiveteam (here) may be able to provide more info on how to help out.
^^^^^^^^^.
The easiest way to make sure the full site (PM logs and all) could be preserved would be to make it possible for at least one person attempting to archive the site to have FTP access with downloading permissions or convert the site into something like an Open Directory (see /r/opendirectories) close to the end of the uptime.
I used to use this but it's been awhile...
It gets you the contents of the website by following all the links but you won't get the code that actually makes the site run.
Kind of...
The whole homepage is made entirely in static HTML, CSS, JS files. Most of the sub pages are too. So you could just use something like https://www.httrack.com or https://github.com/skallwar/suckit to download the site. A lot of the JS and CSS files are minimized, but also upload the non minimized files to webserver too. So if you need the uncompressed version of /js/script.min.js it would be publicly available at /js/script.js - Hope that makes sense. Good luck.
If you're still at it - or maybe for the next one - see if HTTrack does what you need. Failing that, Bulk Image Downloader is really effective for downloading entire galleries quickly.
Certainly complicates things, other than web.archive.org and archive.is there aren't really many sources to verify citation.
You can still use programs such as HTTrack or Web-Capture to keep a local backup. I think webcapture + re-uploading to imgur may be the work around for your use case. Unlikely Canada will block all of Imgur.
There’s actually a pretty detailed user guide.
It’s a little early here so I may be completely wrong but these syntax peak of interest to me:
Links options:
%P *extended parsing, attempt to parse all links, even in unknown tags or Javascript (%P0 don’t use) (—extended-parsing[=N])
n get non-html files ‘near’ an html file (ex: an image located outside) (—near)
t test all URLs (even forbidden ones) (—test) %L )``
——
``a *stay on the same address
This indicates that only the web site(s) where the search started are to be collected. Other sites they point to are not to be imaged. `
After the introduction it goes further in-depth with explanations and examples.`
Coffee time.
I did the same. Because I'm impatient, I went crazy in CS6 to make some custom Samus Returns themes for my 3DS. Great minds think alike and what not. I am mildly curios, what software did you use? I used WinHTTTrack myself.
You could use a software like httrack: https://www.httrack.com/
I actually downloaded and modded the game, so you can save your progress again :D - so if you would like a working save function I would recommend downloading it from my website: http://candybox.duckdns.org
Have fun!
wget would be my goto solution, but maybe this chrome addon may be handy.
Also there's a tool for windows called HTTrack wich should be able to get all the images (quite powerful tool).
Maybe httrack will be good? But this will not generate a single file. I haven't found which methods is archive.is using. Maybe you could talk to the guy behind that project?
I've used httrack before to copy sites... If you have a set list of sites you could just mirror them occasionally... or somehow sniff your traffic and grab site as they show up.
Maybe you should be using HTTrack:
>[HTTrack] allows you to download a World Wide Web site from the Internet to a local directory, building recursively all directories, getting HTML, images, and other files from the server to your computer. HTTrack arranges the original site's relative link-structure. Simply open a page of the "mirrored" website in your browser, and you can browse the site from link to link, as if you were viewing it online.
I used this a lot back in school when creating dummy sites for projects without just turning in gimpy and hastily made HTML crap which most other students were doing.
I used it on a linux CLI so I haven't tried that Windows GUI variant, but it did well in as far as recursively going down and getting many links deep of pages. Not sure how it does with PHP/forum stuff, but it's worth looking at
> Don't know how you do it on Windows
HTTrack Website Copier - Free Software Offline Browser
https://www.httrack.com/
Each Google Apps account receives at least 10GB of storage in Google Sites.
https://www.google.com/sites/overview.html
There may be a better tool—<code>httrack</code>:
>HTTrack is a free (GPL, libre/free software) and easy-to-use offline browser utility.
>It allows you to download a World Wide Web site from the Internet to a local directory, building recursively all directories, getting HTML, images, and other files from the server to your computer. HTTrack arranges the original site's relative link-structure. Simply open a page of the "mirrored" website in your browser, and you can browse the site from link to link, as if you were viewing it online. HTTrack can also update an existing mirrored site, and resume interrupted downloads. HTTrack is fully configurable, and has an integrated help system.
There may be a better tool: HTTrack
From Httrack Users Guide (3.10):
> This option allows the mirror to go only into subdirectories of the initial directory on the remote host. You might want to combine it with the -n option to get all non-html files linked from the pages you find.
> httrack http://www.shoesizes.com/bob/ -O /tmp/shoesizes -D -n
Or something along those lines.
Since you're likely going to want to download the DA wiki, you should know that a good resource to download entire websites to check offline is HTTrack: https://www.httrack.com/
You should use it responsibly, particularly in smaller websites which can get overloaded, but it's a great tool.
OK I just had a go and I can't get it to work either. The error you're talking about is just a warning to say that httrack didn't download any data. I suspect that this wiki is set up in such a way to prevent downloading, and I'm not sure how to get around it. You can read about such shenanigans here
https://www.httrack.com/html/abuse.html
Maybe wget will be better but I have not used that.
Quel format as-tu besoin au final ?
Juste pour relire les articles -> impression en PDF
Les articles au format HTML, méthode sale -> wget --random-wait -r -p -e robots=off -U mozilla http://www.example.com
sous Linux/WSL
Les articles au format HTML, méthode propre -> utiliser un aspirateur de site web comme HTTrack
This may be handy: https://www.httrack.com/
The other way I can think of is programmatically with web spider tools like cheerio. If you code I would much rather go this way. Because you can download the page, inspect it with jquery like selects and download exactly what you want. I downloaded an entire corporate site like this.
You mean something like <strong>this</strong> ?
NOTE: Keep in mind, if the pages have a ton of javascript (and whatnot) then the "offline" copy may not render properly. And many links may actually point to the real website thus making the "copy" rather useless..
Oh, interesting... I was not aware of those libraries - https://www.httrack.com/
Thanks for the suggestion!
Someone else similarly suggested I just use wget or curl to grab a copy of the site.
I could do that, and just run the command every time I want to grab a copy of the updated site.
What I liked about the WP2Static plugin was that I could just push a button and it would export all the relevant files into a directory ... but that's not conceptually different from running a command in a terminal window. 🙂
so i found some old software that is still VERY good but a little slowish for downloading
https://www.httrack.com/ im using it right now. as i couldn't get Jdownloader to download the whole open directory like i wanted it to
it is saving an index.html for each folder, saving folder paths and original naming. the currently OD i am using is only letting me get about 2mb/s which its only anime ost's and its good enough, i haven't tested other OD's to see if i can get faster speeds.
so far its flawlessly working
I use https://www.httrack.com and if you know python.... the below script.
import requests
from bs4 import BeautifulSoup
# Make a request
page = requests.get(
"https://example.com")
soup = BeautifulSoup(page.content, 'html.parser')
# Extract title of page
page_title = soup.title.text
# Extract body of page
page_body = soup.body
# Extract head of page
page_head = soup.head
# print the result
print(page_body, page_head)
You can try httptrack (it also included in many distros). curl
isn't really a tool to download websites. If you want to use cli program, then wget
is more suitable for such tasks. There also is a solutions that used headless browsers and those probably the best to do it, because nowadays internet barely works without javascripts
Not 100% sure this is what you meant, I think the function went out of fashion due to modern sites using different ways of pulling in dynamic content.
Having had a deeper dive it seems my memory was wrong and it doesn’t automatically track as it goes, sorry.
A long time ago I used a website copier. I would download websites that looked really nice, then I would look at the html coding to see how it all worked. I borrowed a few ideas from doing that and made a couple of cool web pages. This was about 20 years ago so I'm sure they've changed but that software is probably still available.
It's called HTTtrack Website Copier. It will copy an entire website to a separate directory on your computer with html files and anything that's downloadable like images, sounds, etc.
I recently tried to copy a website that required authorization (a login and password) because I wanted to have that feature on one of my sites. If there's an off site link to more files on another website , it's supposed to download those files as well but because of the login authorization needed I can't download the actual login procedure they used which is a bummer.
But any websites that don't require a login should work with this software.
Have you looked into HTTrack? You can log in from your browser and then copy over the cookie to HTTrack so you can scrape everything and the website still thinks you're logged in.
Have you tired HTTrack? It works well accessing public pages on the webserver and saving them as flat HTML files, but is a bit of a pain for any pages that need login access.
If you want to copy a website for exploration purposes, you can use HTTrack.
As for tools, the website looks to have a lot of internal content. You could likely use JQuery. For bigger projects I prefer to use some newer libraries like React or Vue.js for all that beautiful data handling.
Not sure what you mean with "convert to regular php", but I find HTTrack (https://www.httrack.com/) to be the most reliable tool for dumping a WP site (or any site really) to HTML. Please note that you may have to rework dynamic/interactive parts of your website (like contact forms) to work in a serverless environment.
Funny I just went through a bunch of this trying to save as much of my halo.bungie.net profile.
A lot of us used httrack to get the job done, it does exactly what it sounds like you’re trying to do!
BTW, there's no telling how long the website will be up. I'm sure if Mazda were to catch wind, they'd probably send a C&D order. I'd recommend downloading the entire site using a tool like this. It takes a bit of tweaking, but you can get everything to download properly, including the flash-based wiring diagrams.
I've never had much luck getting wget
to do anything fancy. So, to mirror a website I use something like <code>httrack</code>. To scrape a website I use a combination of curl
and bash-fu to generate a list of file URLs from the site's HTML, then feed those URLs to wget
individually.
To simply snarf images off a single web page, I use a web browser extension (e.g., Image Downloader). Luckily for you, you're dealing with a wiki, and many wikis have special pages that do things like aggregate all the uploaded files into one place. So, a browser extension for downloading images should do what you want, when used on: thekingoffightersallstar.fandom.com/wiki/Special:ListFiles?limit=10000
I've done some work with MIA in the past, I've used httrack to copy the parts of the website I was using to produce ebooks. Imagine it would work to copy the entire site as well. Cool tool.
E sad se setih, lakša alternativa ti je da skineš ceo sajt, a preko ovoga možda možeš i samo sve fajlove, imaš neke filtere, čuda. Ali nije teško skapirati kako da ga koristiš, ja sam pre skidao neke sajtove sa njim.
I don't have any further suggestions but I do find this interesting, and would love to hear more about anything you find out about how best to do this. Most of the options look like they haven't been updated in a while, which doesn't bode well.
django-bakery looks like the most recently updated of the lot, although it always felt like a bit of a drawback to me that you have to integrate it closely with your views. In my head, this kind of thing should be possible by somehow using the Sitemap framework and grabbing all the pages included... but maybe that would be too slow, or have other problems, I've no idea.
Another option would be to use httrack to crawl your Django site (whether it's run remotely or using the development server) to generate flat HTML files. I've successfully done this for clients' sites in the past to create archivable copies. I ended up using this command:
httrack -%k -#p -%c4 -v -%v -i -s0 -R5 -n -a -%F '' -F 'Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:47.0) Gecko/20100101 Firefox/47.0' 'http://www.example.org/'
And then, within the generated www.example.org
directory doing this to remove index.html
from all of the links (which aren't needed if you're serving the page on a webserver, but would be if you were using it locally from your fileserver):
find . -name '*.html' -exec sed -i '' 's%index.html%%g' {} \;
I did find that some sites required some manual fixing of paths and other things in their HTML, but for a simple, modern site that might not be necessary.
That might or might not be easier than using a Django-based generator!
Tools like https://www.httrack.com/ seem to scratch that itch. No idea what it's worth though, so proceed with caution. FWIW, httrack is available as a community package for archlinux (so it shouldn't be a scam)
So you built a website using wagtail, but now you want to have the website as a static site?
I used this to convert a Wordpress website into a static site: HTTrack
The resulting HTML was a real mess, but it looked the same once rendered by a web browser.
Ich bin mit HTTracker (https://www.httrack.com/page/2/) ganz gut gefahren. 4,5h Laufzeit 5 Fehler, die mich nicht beeinträchtigen (4 Icons und ein Bild). Die internen Links klappen alle. Der Ordner liegt jetzt für meine Gruppe im OneDrive, klappt ganz gut.
As far as I know, downloading a website means visiting every link on that domain. This can be done easily with a recursive wget call on linux, but please be aware that it is really stressing on the targets servers and can be considered misuse. I don't know how HTTrack works but you should absolutely follow https://www.httrack.com/html/abuse.html .
By googling it seems like HTTrack is just a simple GUI where you enter URL-addresses and click go. What is the specific problem you are facing with using it?
You can use Httrack to rip the sales page but I think any option you do will require a bit of cleaning up (removing JavaScript, tracking pixels etc..) and also to update the links. Just takes some basic HTML knowledge.
It sounds like you're trying to make a static local site (on your pc) of a live dynamic online site.
Including your wget string would be a good start here.
You're probably better off using httrack to download the site and then playing with the content using that. My experience is that httrack handles downloading and creating working local links better than wget.
Bear in mind as well that they appear to be a business so altho their site is in the public domain the content may be owned by them.
With HTTrack, you can download entire website or selected web pages for offline viewing at later time. It can also update existing mirrored site and resume interrupted downloads.
Anything you want to do, you should learn the basics how the technology works first, then you will have a foundation of knowledge to build upon instead of flying blind. I have used httrack for years, it's still great https://www.httrack.com/
Not sure if it would work in this day and age, but I remember using this in the past:
​
Just spits out a folder full of HTML files you can do whatever you want with.
There are website backup programs. Depending on how you achieve that backup will drive some archivist insane (if the links are rewritten to refer to the offline backup rather than the online original, etc) but if the content is there then it's there.
I've used this guy in the past. https://www.httrack.com/
Careful with your settings- you dont want your computer trying to back up other sites after following an external link.
Also be aware that this eats away at bandwidth, and I expect quite a few people are doing this right now on this server.
Have you considered trying https://www.httrack.com/ ?
You can download an entire website with it, then run the site from your phone or PC.
A longer way of doing it would be to print every section as a pdf and then use PDFTK to combine them into 1 pdf.
>you might be interested in https://www.httrack.com/
I did come across HTTrack, I am just looking at per page basis not the whole website. Does it have option for just a page too? I would love if I can have an app like pocket let me export the pages that it has copied in good format. Be it with web clipping for Onenote, Evernote & like or service like Pocket, I cannot access the file that is stored.
This is the where the website was stolen from.
https://www.drivenworldwide.com/
This is the tool they used to steal the website
It takes half a brain to do this kinda of thing so they could be making money being productive members of society, but unfortunately they did stupid stuff like this.
Heck of a shame, but glad to hear you're not closing the Youtube channel. Seems like the site is flat HTML though, couldn't you just scrape it and re-upload it as flat pages? I've used HTTrack in the past with good results, free as in free (https://www.httrack.com/). Best of luck, look forward to still following you on Youtube.
Download full website using HTTrack.
I have downloaded few guides in PDF format for my favorite games.
Remember the golden rule: everything not saved on your hard drive will eventually disappear from the Internet. Download important things and keep them on detached external hard drives.
This subreddit is not /r/recommendmeaprogram (wish that it really existed).
Your question has absolutely nothing to do with learning to program and thus is off topic here as per Rule #3.
What you want is a website downloader. There are several, like the already mentioned wget and skipfish, or Httrack.
Removed as per Rule #3
You can use Chrome or FF inspector to see the loaded js files. If you have a large website you could also scrape with ht track ( https://www.httrack.com/) and see the js it pulls. If you are using a framework, there are possibly minified files added dynamically, these are comprised of multiple js snippets appended and minified.
What I tend to look for when optimizing such scenarios is widgets or js dependent features that are, for example used on a single page yet imementation is done in a way that js is added to base template that is loaded on every page.
Also, if you want bonus optimization add minification to your tool chain and if that is not enough, host the js css img ressources on a CDN. 👍 That is actually a cool task you have in hand, have fun!
The old Scrapbook addon used to be able to do this, but I don't think it's possible currently within Firefox since they killed all the old addons. I'd probably use HTTrack (a standalone program, available for Windows or Linux) for this sort of thing.
Maybe https://www.httrack.com/? I initially thought archive.org, but I'm actually not sure if it would recurse through all the pages. Then I googled recursive website crawler and archiver and httrack was the first option. Hope it works out for you!
Hi there! Are you wanting to clone the entire website layout and all? If so you might find it'll be super hard to maintain since a lot of these sites are auto generated, so editing it by hand or trying to get it to work on a new platform probably just isn't worth the effort! I gave it a spin trying to clone it with <code>httrack</code> but the only working page it can scrape is the homepage. While you could definitely go page by page and save the HTML like that, it really probably isn't worth it!
There are loads of open source technologies out there that you might be interested, namely Static Site Generators which let you build sites for free ( doesn't include hosting ), but these might be a little technical to setup.
I hope this answers your question, and best of luck with the writing!
Okay, so a bit more explanation:
`wget` is a command-line program. There are other programs that have a graphical interface that may be more useful to you. I know HTTrack is one that gets recommended a lot, but it's designed for mirroring entire websites so I'd strongly recommend changing its settings to only grab what you need (sorry I don't know details for this, not what I normally use).
A user-agent is something that tells a website which browser is being used. You probably won't need to worry about it if you use pre-made tools. And if your deviant art pages are public, you don't need to provide a login to the downloader.
Any sort of text file, or just a list you can copy-paste, would be good if you already have one. If not, there's probably a way to scrape a list, but it may take some custom scripting.
You can use wget to download a website/blog. Here's a quick guide. I just used it to download a blog on blogspot and it worked just fine, it got all the postings and saved them as an HTML file that you can open up and browse through, you can also use something like HTTrack if you're not comfortable with command line and want a GUI
I don't think there's an easy way, but using HTTrack I copied the docs as they are today, 15 February 2019. They can be found in this Dropbox folder. I hope they are complete, my internet cut out in the middle of the upload, but I think I got everything.
The nice little drop-down menus don't work, unfortunately, but you can just use your file explorer to open the relevant pages. If anybody has a better way than this to access the documentation offline please chime in!
There certainly is Httrack for OSX fifth from the top on their download page.
https://www.httrack.com/page/2/en/index.html
There's a manual on their website here
https://www.httrack.com/page/2/en/index.html
Take care to check out the "what not to do" page so you don't get in trouble.
In that case, I'm of no use. I don't really use that app other then downloading albums one at a time. It seems to work well for that but Idk about downloading a whole site. There is HTTRACK which I have little to no experience with but it might be what you are looking for.
Take a look at https://www.httrack.com/
It's doesn't currently respond for me at work, but https://downforeveryoneorjustme.com/httrack.com says it's up...
If you can't access it, google "website offline viewer" as I think that's what they might be referred to as. I'm sure there are many options, but I think httrack is one of the more popular ones.
If you can get FTP, SFTP and database access or CPanel access you can copy down most if not all of the site and move it to a new host.
Alternatively, if it is a site that doesn't have any dynamic functionality and you just need the layout you can use something like HTTrack (https://www.httrack.com/) to copy down the HTML files for the site and move those to a new host.
There is a way to view web pages on a computer while offline. The two ways I know of is using a firefox plugin named unmht:
-https://addons.mozilla.org/en-US/firefox/addon/unmht/
Another way is to use a software called Httrack:
You will be able to download webpages from the internet (using an internet connection), then view those webpages for offline use. Though, do keep in mind that you can't watch videos offline using these tools. Both of these options are strictly for viewing plain web pages and visuals. Good luck.
If it's not a very complex site, you might create a static copy (pages and assets) using HTTrack. Forms won't work but it usually makes a useful site that can be browsed normally.
One of my professors puts up his notes on a school server (he basically wrote an online textbook but had it in website form so it was easy to navigate), and due to some weirdness with the school, it was only accessible if you were connected to the school's internet while on campus. So he told us to download the website ourselves and have a copy of it to read offline. (This was a computer science course, so that was actually part of the class, he wasn't just being a dick by not offering the download himself)
Point of this randomass anecdote is, I did this with Ayumilove cause of all the ads, and you could too if you wanted. I used HTTrack link, but there are simpler ones out there too, I just like this one cause I can tell it not to download comments and shit.
Just start studying now and get ahead of it. Sure MACM 101 will get you used to proofs, but if you can avoid taking an extra course, always do it. If you want to gain the experience, download http://www.cs.sfu.ca/CourseCentral/101.MACM/abulatov/ before the prof takes it down (use https://www.httrack.com) and go through it with the textbook.
University is all supposed to be self-taught. You go through the text on your own and teach yourself everything, then the professor lightly touches on the main points and goes through through some examples during lecture. Lectures are just for reinforcement. You should be spending 2-3 hours on your own learning for every hour in class.
A client of mine used Vistaprint and we basically just left the domain as is and setup a second domain under a different TLD for their email.
https://www.httrack.com is what I've used in the past to pull a website down.
The City of Heroes Wikia is still there 4.5 years after the game closed, despite the objections of the people they stole their content from. I expect the MAA Wikia will be around for a good while.
That said, if you really need it on a permanent basis, you could download a tool like (Win)HTTrack to download the entire site.
You can use this program to download whole webpages including correct links. Not all websites work. https://www.httrack.com/page/2/
Firefox also allows off-line viewing of web pages that you have already been to.
In past I have just downloaded the site (with all of it's resources as files). Not sure if that is easy for you.
There is likely a tool out there that can do that for you, something like this: https://www.httrack.com
Your browser will probably have a "Save" option somewhere, which should save the main part of the page as a .html file, and then also save a folder of associated files (images, formatting code, etc). Try pressing Ctrl+S and see what happens.
If you need a deeper copy (e.g. of a whole site rather than just a single page) then you'll want a more sophisticated tool, like HTTrack, which will recursively search for all the pages on a site and download them all to a folder, along with all the associated extra files. Might have some broken images/links if they go to a separate domain (and that domain then goes down), but it'll come close to replicating the full experience of a site.
There's also an app I've been making use of lately, "Offline Browser" which will save a page to your phone's memory. The use-case they mostly talk about is to save pages to read when you're out of signal, which is what I've been using it for, but it would work just the same if the site went down.
And other web crawlers will allow you to do just that.
It will probably need some adjustments to the default install settings, with some trial and error you can get it to work.
Here is the first half to your question.
https://www.httrack.com/html/fcguide.html
I've not tried this conversion tool however, you can check it out.