Save an entire webpage, with all images and css, into just one folder and one file, with wget?











up vote
1
down vote

favorite












Is there a way to mimic the "Save - as" function in a browser with wget?



When I save a webpage in a browser, I get a folder with the assets(images, js, css) and the index file, which has the page name:



Nov 28 reddit: the front page of the internet_files
Nov 28 reddit: the front page of the internet.html


But no matter what, when I use wget, I get something like this:



Nov 28  a.thumbs.redditmedia.com
Nov 28 b.thumbs.redditmedia.com
Nov 28 m.reddit.com
Nov 28 out.reddit.com
Nov 28 reddit.com
Nov 28 www.reddit.com
Nov 28 www.redditstatic.com


I tried using these:



wget -E -H -k -K -p https://reddit.com
wget -r -x -mirror https://reddit.com



and came up with this:
wget -E -H -k -p -e robots=off https://www.reddit.com



but all of them either made several folders or didn't download everything needed to view the page offline.



How would I set this up?










share|improve this question


























    up vote
    1
    down vote

    favorite












    Is there a way to mimic the "Save - as" function in a browser with wget?



    When I save a webpage in a browser, I get a folder with the assets(images, js, css) and the index file, which has the page name:



    Nov 28 reddit: the front page of the internet_files
    Nov 28 reddit: the front page of the internet.html


    But no matter what, when I use wget, I get something like this:



    Nov 28  a.thumbs.redditmedia.com
    Nov 28 b.thumbs.redditmedia.com
    Nov 28 m.reddit.com
    Nov 28 out.reddit.com
    Nov 28 reddit.com
    Nov 28 www.reddit.com
    Nov 28 www.redditstatic.com


    I tried using these:



    wget -E -H -k -K -p https://reddit.com
    wget -r -x -mirror https://reddit.com



    and came up with this:
    wget -E -H -k -p -e robots=off https://www.reddit.com



    but all of them either made several folders or didn't download everything needed to view the page offline.



    How would I set this up?










    share|improve this question
























      up vote
      1
      down vote

      favorite









      up vote
      1
      down vote

      favorite











      Is there a way to mimic the "Save - as" function in a browser with wget?



      When I save a webpage in a browser, I get a folder with the assets(images, js, css) and the index file, which has the page name:



      Nov 28 reddit: the front page of the internet_files
      Nov 28 reddit: the front page of the internet.html


      But no matter what, when I use wget, I get something like this:



      Nov 28  a.thumbs.redditmedia.com
      Nov 28 b.thumbs.redditmedia.com
      Nov 28 m.reddit.com
      Nov 28 out.reddit.com
      Nov 28 reddit.com
      Nov 28 www.reddit.com
      Nov 28 www.redditstatic.com


      I tried using these:



      wget -E -H -k -K -p https://reddit.com
      wget -r -x -mirror https://reddit.com



      and came up with this:
      wget -E -H -k -p -e robots=off https://www.reddit.com



      but all of them either made several folders or didn't download everything needed to view the page offline.



      How would I set this up?










      share|improve this question













      Is there a way to mimic the "Save - as" function in a browser with wget?



      When I save a webpage in a browser, I get a folder with the assets(images, js, css) and the index file, which has the page name:



      Nov 28 reddit: the front page of the internet_files
      Nov 28 reddit: the front page of the internet.html


      But no matter what, when I use wget, I get something like this:



      Nov 28  a.thumbs.redditmedia.com
      Nov 28 b.thumbs.redditmedia.com
      Nov 28 m.reddit.com
      Nov 28 out.reddit.com
      Nov 28 reddit.com
      Nov 28 www.reddit.com
      Nov 28 www.redditstatic.com


      I tried using these:



      wget -E -H -k -K -p https://reddit.com
      wget -r -x -mirror https://reddit.com



      and came up with this:
      wget -E -H -k -p -e robots=off https://www.reddit.com



      but all of them either made several folders or didn't download everything needed to view the page offline.



      How would I set this up?







      browser wget css webpage save-as






      share|improve this question













      share|improve this question











      share|improve this question




      share|improve this question










      asked Nov 28 '16 at 8:01









      Rich

      1062




      1062






















          1 Answer
          1






          active

          oldest

          votes

















          up vote
          0
          down vote













          You/WGET can't. It can download all linked resources in one given download, but that would result in multiple folders due to its nature of crawling, not interpreting (and not being bound to HTTP either).



          Also your impression is too narrow: there are web browsers which can save pages into MHT files/archives, which is even a standard - see https://en.wikipedia.org/wiki/MHTML






          share|improve this answer





















            Your Answer








            StackExchange.ready(function() {
            var channelOptions = {
            tags: "".split(" "),
            id: "3"
            };
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function() {
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled) {
            StackExchange.using("snippets", function() {
            createEditor();
            });
            }
            else {
            createEditor();
            }
            });

            function createEditor() {
            StackExchange.prepareEditor({
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: true,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            imageUploader: {
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            },
            onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            });


            }
            });














            draft saved

            draft discarded


















            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fsuperuser.com%2fquestions%2f1150495%2fsave-an-entire-webpage-with-all-images-and-css-into-just-one-folder-and-one-fi%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown

























            1 Answer
            1






            active

            oldest

            votes








            1 Answer
            1






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes








            up vote
            0
            down vote













            You/WGET can't. It can download all linked resources in one given download, but that would result in multiple folders due to its nature of crawling, not interpreting (and not being bound to HTTP either).



            Also your impression is too narrow: there are web browsers which can save pages into MHT files/archives, which is even a standard - see https://en.wikipedia.org/wiki/MHTML






            share|improve this answer

























              up vote
              0
              down vote













              You/WGET can't. It can download all linked resources in one given download, but that would result in multiple folders due to its nature of crawling, not interpreting (and not being bound to HTTP either).



              Also your impression is too narrow: there are web browsers which can save pages into MHT files/archives, which is even a standard - see https://en.wikipedia.org/wiki/MHTML






              share|improve this answer























                up vote
                0
                down vote










                up vote
                0
                down vote









                You/WGET can't. It can download all linked resources in one given download, but that would result in multiple folders due to its nature of crawling, not interpreting (and not being bound to HTTP either).



                Also your impression is too narrow: there are web browsers which can save pages into MHT files/archives, which is even a standard - see https://en.wikipedia.org/wiki/MHTML






                share|improve this answer












                You/WGET can't. It can download all linked resources in one given download, but that would result in multiple folders due to its nature of crawling, not interpreting (and not being bound to HTTP either).



                Also your impression is too narrow: there are web browsers which can save pages into MHT files/archives, which is even a standard - see https://en.wikipedia.org/wiki/MHTML







                share|improve this answer












                share|improve this answer



                share|improve this answer










                answered Dec 7 at 23:11









                AmigoJack

                101




                101






























                    draft saved

                    draft discarded




















































                    Thanks for contributing an answer to Super User!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid



                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.


                    To learn more, see our tips on writing great answers.





                    Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


                    Please pay close attention to the following guidance:


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid



                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.


                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function () {
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fsuperuser.com%2fquestions%2f1150495%2fsave-an-entire-webpage-with-all-images-and-css-into-just-one-folder-and-one-fi%23new-answer', 'question_page');
                    }
                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    Probability when a professor distributes a quiz and homework assignment to a class of n students.

                    Aardman Animations

                    Are they similar matrix