Slashdot is powered by your submissions, so send in your scoop

 



Forgot your password?
typodupeerror
×
Image

Even Faster Web Sites 171

Michael J. Ross writes "Slow Web page loading can discourage visitors to a site more than any other problem, regardless of how attractive or feature-rich the given site might be. Consequently, many Web developers hope to achieve faster response times using AJAX (Asynchronous JavaScript and XML), since only portion(s) of an AJAX page need to be reloaded. But for many rich Internet applications (RIAs), such potential performance gains can be lost as a result of non-optimized JavaScript, graphics, and CSS files. Steve Souders — a Web performance expert previously at Yahoo and now with Google — addresses these topics in his second book, Even Faster Web Sites: Performance Best Practices for Web Developers." Read on for the rest of Michael's review.
Even Faster Web Sites
author Steve Souders
pages 254 pages
publisher O'Reilly Media
rating 8/10
reviewer Michael J. Ross
ISBN 978-0596522308
summary Advanced techniques for improving website performance.
The book was published by O'Reilly Media on 18 June 2009, under the ISBN 978-0596522308. The publisher makes available a Web page, where visitors can purchase the print and electronic versions of the book (as well as a bundle of the two), read the book online as part of the Safari library service, and check the reported errata — comprising those confirmed by the author (of which there are currently two) and any unconfirmed errors (all six of which are valid, though the fifth one may be a coincidence). In a break with traditional practice among technical publishers nowadays, there is no sample chapter available, as of this writing.

In many ways, this second book is similar to Steve's previous one, High Performance Web Sites: It presents methods of enhancing the performance of websites, with a focus on client-side factors. It is fairly slender (this one is 254 pages), relative to most programming books nowadays, and the material is organized into 14 chapters. However, unlike its predecessor, Even Faster Web Sites emphasizes generally more advanced topics, such as script splitting, coupling, blocking, and chunking (which to non-developers may sound like a list of the more nefarious techniques in professional hockey). This second book also has employed a team approach to authorship, such that six of the chapters are written by contributing authors. In his preface, Steve notes that the 14 chapters are grouped into three broad areas: JavaScript performance (Chapters 1-7), network performance (Chapters 8-12), and browser performance (Chapters 13-14). The book concludes with an appendix in which he presents his favorite tools for performance analysis, organized into four types of applications: packet sniffers, Web development tools, performance analyzers, and some miscellaneous applications.

In the first chapter, "Understanding Ajax Performance," guest author Douglas Crockford briefly describe some of the key trade-offs and principles of optimizing applications, and how JavaScript now plays a pivotal role in that equation — as websites nowadays are designed to operate increasingly like desktop programs. On pages 2 and 3, he uses some figures to illustrate fixed versus variable overhead, and the dangers of attempting to optimize the wrong portions of one's code. By the way, the so-called "axes" are not axes, or even Cartesian grid lines, but simply levels. Aside from its choppy narrative style and a pointless religious reference in the first paragraph, the material serves as a thought-provoking springboard for what follows. Chapter 2, titled "Creating Responsive Web Applications," was written by Ben Galbraith and Dion Almaer, who discuss response times, user perception of them, techniques for measuring latency, browser threads, Web Workers, Google Gears, timers, and memory issues. The material is neatly explained, although Figure 2-2 is quite confusing; moreover, both of the figures on that page should not have been made Mac- and Firefox-specific.

In the subsequent four chapters, Steve dives into the critical topic of how to optimize the performance of JavaScript-heavy pages through better script content and organization — specifically, how and when to split up large scripts into smaller ones, how to load scripts without blocking one another or breaking dependencies within the code, and how to best in-line scripts, when called for. Each of the four chapters follows an effective methodology: The first author delineates a particular performance mistake made by even some of the most popular websites, with the statistics to back it up. He presents one or more solutions, including any relevant tools, again with waterfall charts illustrating how well the solutions work. Lastly, he explains any browser-specific issues, oftentimes with a handy chart showing which possible method would likely be optimal for the reader's given situation, such as expected browser choices in the site's target audience. When there are potential pitfalls, Steve points them out, with helpful workarounds. He generally provides enough example source code to allow any experienced developer to implement the proposed solutions. Unfortunately, the example code does not appear to be available for download from O'Reilly's website.

The discussion of JavaScript optimization is capped off by the seventh chapter, written by Nicholas C. Zakas, who explains variable scope within JavaScript code, the advantages of choosing local variables as much as possible, scope chain augmentation, the performance ramifications of the four major data types (literal values, variables, arrays, and objects), optimizing flow control statements, and string concatenation. He outlines what sorts of problems can cause the user's Web browser to freeze up, and the differing responses she would see depending upon her chosen browser. Nicholas concludes his chapter by explaining how to utilize timer code to force long-running scripts to yield, in order to avoid these problems. By the way, in Figures 7-2 and 7-3, the data point symbols need to be enlarged so as to be distinguishable; as it is, they are quite difficult to read. More importantly, on page 93, the sentence beginning "This makes array lookup ideal..." is either misworded or mistaken, since array lookup cannot be used for testing inclusion in ranges.

With the eighth chapter, the book shifts gears to focus on network considerations — namely, how to improve the site visitor's experience by optimizing the number of bytes that must be pushed down the wire. In "Scaling with Comet," Dylan Schiemann introduces an emerging set of techniques that Steve Souders describes as "an architecture that goes beyond Ajax to provide high-volume, low-latency communication for real-time applications such as chat and document collaboration" — specifically, by reducing the server-side resources per connection. In Chapter 9, Tony Gentilcore discusses a rather involved problem with using gzip compression — one that negatively impacts at least 15% of Internet users. Even though videos, podcasts, and other audiovisual files consume a lot of the Internet's bandwidth, images are still far more common on websites, and this alone is reason enough for Chapter 10, in which Stoyan Stefanov and Nicole Sullivan explain how to reduce the size of image files without degrading visible quality. They compare the most popular image formats, and also explain alpha transparency and the use of sprites. The only clear improvement that could be made to their presentation is on page 157, where the phrase "named /favicon.ico that sits in the web root" should instead read something like "usually named favicon.ico," since a favicon can have any filename, and can be located anywhere in a site's directory structure.

The lead author returns in Chapter 11, in which he explains how to best divide resources among multiple domains (termed "sharding"). In the subsequent chapter, "Flushing the Document Early," Steve explores the approach of utilizing chunked encoding in order to begin rendering the Web page before its full contents have been downloaded to the browser. The third and final section of the book, devoted to Web browser performance, consists of two chapters, both of whose titles neatly summarize their contents: "Using Iframes Sparingly" and "Simplifying CSS Selectors." That last chapter contains some performance tips that even some of the most experienced CSS wizards may have never heard of before. As with most of the earlier chapters, the narrative tends to be stronger than the illustrations. For instance, Figure 14-5, a multiline chart, is quite misleading, because it appears to depict three values varying over time, when actually each of the ten x-axis coordinates represents a separate major website. A bar chart would obviously have been a much better choice.

Like any first edition of a technical book, this one contains a number of errata (aside from those mentioned earlier): In Figure 1-1, "iteration" is misspelled. On page 23, in the sentence beginning "Thus, if...," the term "was" should instead read "were." In Figures 7-1 and 7-4, the "Global object" box should not contain "num2." On page 95, in the phrase "the terminal condition evaluates to true," that instead should read "false." On page 147, in the sentence beginning "However, the same icon...," the "was" should instead read "were." On page 214, "Web-Pagetest. AOL" should instead read "Web-Pagetest, then AOL," because the first sentence is one long absolute phrase (i.e., lacking a finite noun and verb).

All of these defects can be easily corrected in future printings. What will probably need to wait for a second edition, are improvements to the figures that are in need of replacement or clarification. What the publisher can rectify immediately — should the author and O'Reilly choose to do so — would be to make all of the example source code available for download.

Even though this book is decidedly longer than High Performance Web Sites, and has many more contributing authors, it does not appear to contain as much actionable information as his predecessor — at least for small- to medium-sized websites, which probably make up the majority of all sites on the Web. Even though such methodologies as Comet, Doloto, and Web Workers appear impressive, one has to wonder just how many real-world websites can justify the development and maintenance costs of implementing them, and whether their overhead could easily outweigh any possible benefits. Naturally, these are the sorts of questions that are best answered through equally hard-nosed experimentation — as exemplified by Steve Souders's admirable emphasis upon proving what techniques really work.

Fortunately, none of this detracts from the application development and optimization knowledge presented in the book. With its no-nonsense analysis of Internet performance hurdles, and balanced recommendations of the most promising solutions, Even Faster Web Sites truly delivers on its title's promise to help Web developers wring even more speed out of their websites.

Michael J. Ross is a freelance Web developer and writer.

You can purchase Even Faster Web Sites from amazon.com. Slashdot welcomes readers' book reviews — to see your own review here, read the book review guidelines, then visit the submission page.

*

This discussion has been archived. No new comments can be posted.

Even Faster Web Sites

Comments Filter:
  • by Anonymous Coward on Wednesday July 22, 2009 @01:48PM (#28784583)

    Slashdot is the SLOWEST web site on the net.

    • Re: (Score:2, Flamebait)

      by pha7boy ( 1242512 )
      maybe it's just you.

      the parent is about as useful a comment as saying "first."
    • Re: (Score:3, Interesting)

      by Phroggy ( 441 )

      It's fast enough for me.

      The CSS is horribly broken, but I have no complaints about the speed. Posting is certainly faster for me than it used to be before they switched to AJAX. It just looks like crap.

      • by dyefade ( 735994 )

        Posting is certainly faster for me than it used to be before they switched to AJAX. It just looks like crap.

        Really? It takes AGES to preview a comment, I'm thinking 10-20 seconds, during which time you can't scroll down or do anything else (well you can, you'd just have to remember to scroll back up). At least with the vanilla html version you could open a new tab, and I don't remember having any issues with speed back then.

        The rest of the site is great though - keyboard navigation for comments is something I really miss when on other sites now.

      • Re: (Score:2, Insightful)

        by MogNuts ( 97512 )

        "Classic" /. was always fast posting that I recall. I just wish I could have the features of the new /. (read: "the +/-" to meta-mod stories), with the old that was lost (read: ability to only display stories from certain sections, rather than only by author it seems now).

        And maybe it's just me, but where is the option to not timeout your login? I don't want auto-login or to stay logged in after I close the browser (for security), but I hate that /. logs me out after only like 2 minutes of inactivity. I can

        • Have your browser delete all cookies on exit. Firefox can do that. That will log you out of Slashdot, and everything else too.
      • by godrik ( 1287354 )
        I was going to reply to say it is usually slow for me and just when I clicked the 'reply to this' button, I waited more than 15 seconds for the reply box to appear. Sometimes the preview button takes more than a minute to process the preview. There is probably a lot of parameter that impact the times. But it is usually slow for me.
    • by eln ( 21727 ) on Wednesday July 22, 2009 @02:02PM (#28784815)
      Slashdot isn't slow, it's just buggy as hell.
      • by sootman ( 158191 )

        Agreed. Also, it's slow.

      • I mean, how can you overlook something as glaring as a missing space? "Anonymous Cowardon" indeed!
        What I miss most is some kind of feedback corner - ok, sometimes the admins invite comments on new stuff, but those threads vanish too rapidly from the front page. Why not include a simple new slashbox named "Feedback" where people could notify the editors and administrators of things that bug them? This might also be an easy way to improve the site by pointing out common problems (spelling errors in submissio
    • "Preview Post" lag (Score:5, Interesting)

      by electrosoccertux ( 874415 ) on Wednesday July 22, 2009 @02:03PM (#28784827)

      The only part that bothers me is the "Preview Post" lag lasting for 20-35 seconds. I love everything else about the navigation on the site, though.

      • Re: (Score:3, Informative)

        by Ant P. ( 974313 )

        That's the part where it port scans your computer to see if you're running an open proxy. Apparently everyone's considered untrustworthy regardless of current karma or account age.

      • I don't know how you can love it. Slashdot is one of the most awful web sites around. I come back for the community and the comments but really, especially for a community of highly opinionated geeks it's just pathetic.

        For example, this very post page I am writing in fails to render correctly in Chrome, one of the most standards compliant browsers you can get. A simple thing, the "Reply to This" button with padding that overflows so it looks crap.

        Apparently there is some whiz bang ajax interface for expa

    • by Hatta ( 162192 ) *

      Turn off javascript, it's actually usable then. Nothing worthwhile on the site actually uses javascript.

    • by 1u3hr ( 530656 )
      As I post, there are only 31 comments. Yet it still took THIRTY FUCKING SECONDS before the hourglass turned into a cursor and I could scroll the fucking page. During which time I notice the CPU usage going to 100%. WHAT THE FUCK IS SLASHDOT DOING????

      A couple of years ago Slashdot was bearable, especially as I could turn it to the "low tech" preference, which has now disappeared, and that even let me see it in my default font and not the sans font I do now. Now I have a later browser, twice as much RAM, and

    • by HTH NE1 ( 675604 )

      It actually gets slower when you have mod points on some platforms (I can't say all as I don't use all). It can take a long time just to open the combobox choose how to mod the message, close the box, see it change, and get focus away from the box so when you try to scroll using the keyboard it doesn't change how you're modding the post. I think it is a problem with form efficiency in the browser: it just doesn't expect that many comboboxes to exist in a single form.

      Then again, it may be just as slow to hav

    • You're obviously not a Gawker Media fan, or haven't visited them since their last commenting system update. I still can't get comments to loand in Firefox, and switching to an IE tab means the comments load in a minute or two. They JUST fixed comment viewing in IE 8.
  • No... (Score:4, Informative)

    by Last_Available_Usern ( 756093 ) on Wednesday July 22, 2009 @01:49PM (#28784597)
    The greatest barrier to website use is information overload, kinda like this review.

    TLDR

    WALL OF TEXT CRITS YOU FOR +5 TROLL
    • Re: (Score:3, Interesting)

      by mcgrew ( 92797 )

      I'd say the greatest barrier to website use is advertising, which isn't information; it's useless data. As to the book, AJAXing your site makes it slow. If you want a fast site, don't put anything on any page that doesn't absolutely need to be there, and have fast pipes and fast and enough servers.

      If I see a really fancy layout on a web site, I always wonder why its developer thought the content was so bad that they needed to disguise it with slickness?

      bing comes to mind. I've tried it quite a few times, an

      • by Chabo ( 880571 )

        BTW, shouldn't that be TLTR instead of TLDR? Oh, too long DIDN'T read. Gotcha. You wouldn't like most of my journals, or anybody's books.

        "tl;dr" usually implies more than just the length of the text involved (although in some cases people are just that impatient). Usually it means that the writing either drags, or is just unnecessarily long for the message conveyed.

      • Re:No... (Score:4, Insightful)

        by Bogtha ( 906264 ) on Wednesday July 22, 2009 @03:58PM (#28786611)

        AJAXing your site makes it slow.

        Nonsense. Using Ajax can slow down a site, or it can speed it up immensely. It depends on how you use it. When GMail was first launched, everybody raved about how fast it was. That was the Ajax.

        If you want a fast site, don't put anything on any page that doesn't absolutely need to be there, and have fast pipes and fast and enough servers.

        Assume you did that for Slashdot. You could then make further speed improvements by using Ajax in various places, for example so that it didn't reload a page full of hundreds of comments whenever you moderated.

        • by mcgrew ( 92797 )

          Yes, if your site is like slashdot ajax could speed it up if done properly, but unfortunately most aren't like that. Any tool can be misused, and the webmastering community is really bad about jumping on the bandwagon using tools that just aren't appropriate for the task at hand.

          I'm especially annoyed with sites that use javascript for a simple link. There are a few cases where that would be appropriate, but very few of the ones I've seen that use it are appropriate. If you right click one of these to open

        • by pongo000 ( 97357 )

          When GMail was first launched, everybody raved about how fast it was. That was the Ajax.

          So what happened? GMail definitely can't be classified as "fast" by any stretch of the imagination...

    • Yeah the more text in a web site or Slashdot story, the longer it takes to load.

      Sometimes less is more. Learn to say more by using less text.

      Turn off "Verbose" mode.

  • AJAX (Score:5, Insightful)

    by metamechanical ( 545566 ) on Wednesday July 22, 2009 @01:50PM (#28784611)
    Has anybody else noticed that of all the websites visited, some of the SLOWEST make heavy use of AJAX? Or is that just me?
    • Re:AJAX (Score:4, Insightful)

      by Archangel Michael ( 180766 ) on Wednesday July 22, 2009 @02:07PM (#28784907) Journal

      Conversely, some of the fast websites use basic TEXT and skimp on the graphics.

      I do websites for non-profits and I one of my rules in doing a website is NO UNNECESSARY graphics. Graphics that add pretties but don't add anything else are not allowed. I also run all the graphics through a size filter to limit unneeded pixel count and quality. You don't need 25MB photo on the front page; a scaled and reduced quality is a better choice.

      My way of saying this is, "I don't need to see the pimples on the porn stars butt". Sometimes too much detail is too much.

      • >>>Conversely, some of the fast websites use basic TEXT and skimp on the graphics.

        I concur. On my 50kbit/s dialup connection the sites that load fastest are the ones that resemble old 90s-era sites - just plain text and images. The slow ones are those sites that have to preload some javascript, flash, or other executable before they can display anything. Staring at a blank screen for 2 minutes is annoying.

        >>>NO UNNECESSARY graphics. Graphics that add pretties but don't add anything else

        • Check this out ... http://web.forret.com/tools/megapixel.asp?width=3464&height=3464 [forret.com]

          okay, so it isn't 25, it is 18MB (RAW). I would consider 7MB PNG to be overkill as well. Scale down, crop, and you can easily get it under 250k or even smaller.

        • Second, images aren't really a big deal on fast connections because they zip right through,

          That's not really true. Every image requires a separate http connection - which takes time to set up - and it has to be rendered and flowed, which takes time to do (depending on browser and processor speed etc).

          and on slow connections virtually all dialup ISPs provide compression. So a 100 kilobyte photo would be squeezed to about 20 kilobyte before being sent over the phoneline... a mere 3 seconds.

          Not even close. Ph

          • Generally if you compress a file that's already compressed, it gets bigger.

            Then you're doing it wrong. To compress a JPEG to a lower size you would decrease the quality setting in whatever editing program you are using and reencode the picture. It's really not that hard.

            • That's not what i'm talking about. Nor what the GP was referring to. I'm talking about, for example, gzipping a jpeg file.

              • But that wasn't what commodore64_love was talking about. By recompressing the image he does mean that they recompress the image to a lower quality before sending it to the user.

                • virtually all dialup ISPs provide compression. So a 100 kilobyte photo would be squeezed to about 20 kilobyte before being sent over the phoneline

                  • Yes, I read what he's talking about and I've seen the compression he refers to in action. The ISP will re-render the images to a lower quality they aren't gzipping them.

                    • Maybe. But it's not very common (not anywhere i've ever used dialup, anyway) and it produces ugly results.

                      V.42bis on-the-fly compression, which is handled by the modem, is considerably more common - and that recompresses image files, making the files bigger.

                    • >>>[image compression is] not very common and it produces ugly results.

                      ???. AOL has it. Netzero has it. Juno has it. Erols had it. Netscape has it. And you're right it produces ugly results, but if you're on dialup you're more-interested in speed than pretty pictures. Plus there's always the option to load the original image if, for example, you're viewing playboy.com. You don't have to use image compression.

                      >>>V.42bis on-the-fly compression, which is handled by the modem, is consi

                    • ???. AOL has it. Netzero has it. Juno has it. Erols had it. Netscape has it.

                      None of those operate in any of the countries i've used dialup in.

                      And you're right it produces ugly results, but if you're on dialup you're more-interested in speed than pretty pictures.

                      I sincerely hope i never have to use dialup again (last time was the year before last in a remote part of Australia), but if i do, i'll stick to just turning off images in my browser. That's even faster.

                • So compression means two things now? Damn kids. I could compress my family photos by taking them again with a crap old camera.
                  • >>>So compression means two things now? Damn kids

                    Yes. There's lossless compression where you take a text or executable file and squash it with ARC or ZIP. And then there's lossy compression where you take a photo and store it as a JPEG, or music and store it as MP2 or MP3. It's been that way for about twenty years now.... grandpa. ;-)

              • >>>>>To compress a JPEG to a lower size you would decrease the quality setting in whatever editing program you are using and reencode the picture.

                >>That's not what i'm talking about. Nor what [commodore64_love] was referring to

                Yes Will I was referring to that (adjusting the quality setting). You made the mistake of thinking the word "compression" meant lossless compression, but when my ISP squeezes a 100KB image to a 20KB image, they are using *lossy* compression.

          • >>>>>on slow connections virtually all dialup ISPs provide compression. So a 100 kilobyte photo would be squeezed to about 20 kilobyte before being sent over the phoneline... a mere 3 seconds.

            >>Not even close. Photos are usually jpegs and therefore they're already compressed.

            Bzzz. They are compressed, but they can still be compressed further. The ISP opens the JPEG or GIF image, resets the quality from 90 to 10* and then sends it over the phoneline. That's how an image can be compres

        • So a 100 kilobyte photo would be squeezed to about 20 kilobyte before being sent over the phoneline... a mere 3 seconds

          Come again on that? You want to compress a compressed image by a factor of five?

      • Re: (Score:3, Interesting)

        by Chabo ( 880571 )

        Conversely, some of the fast websites use basic TEXT and skimp on the graphics.

        As I said in another thread the other day: Whether or not you like his writing, I think Maddox hit the peak of usable web design: dark background, with large-font bright text. It's the easiest webpage on the internet to read, and despite having some graphics, it loads very quickly because he uses the graphics as actual content, not just filler.

        • No way man, websites designed like that have a persistence of vision effect that makes it extremely difficult to read any of the other windows on my computer for like 30 seconds. Poorest choice in web design. Now, I am all for using bigger fonts. Too many people are in denial about their need for glasses, so small fonts == lost customers.

          • Maddox says black text on a white background is like "Staring into a lightbulb". I never thought of it like that, but he makes a good point. Whe was the last time you saw a TV show display pure white background with black text on it?

            Virtually never. The television world has learned the "friendliest" screen is a dark background (black, blue) with light-colored white test superimposed over it. Well computers use the same CRT or LCD technology as televisions. It makes sense to me they should follow the sa

      • My way of saying this is, "I don't need to see the pimples on the porn stars butt".

        You might not, and I certainly don't, but I have a friend who's interests are exceptionally "backal" who'd probably appreciate that type of detail. Ugol's Law implies that there are at least a few people out there looking for that in their pr0n. Have you considered an alternate page that caters to them?

      • by drwho ( 4190 )

        I agree.. For many years I had a very fast web site because everything was done in text. Now there's some graphics, but it still loads pretty fast. The problem is, there's all these new inventions, loading content that is not static, but computed, from a wide number of sites, each with several DNS lookups (multiple dns lookups slow down site view immensely). Too many web designers are paid by the hour, and tell their clients that they need to look 'modern' with whatever gadgets and eye candy is out there. M

    • Re:AJAX (Score:4, Insightful)

      by iknowcss ( 937215 ) on Wednesday July 22, 2009 @02:22PM (#28785149) Homepage
      You're right, but it's not the AJAX that is the cause. It's the incompetent "web developers" who think AJAX is going to solve all of their problems by getting them a better job and making them worth more. These are the same people who firmly believe that Ruby on Rails is the only way to create an AJAX application and not-so-coincidentally couldn't code their way out of a paper bag. I should know. I went to high school with them :)
      • You're right, but it's not the AJAX that is the cause. It's the incompetent "web developers" who think AJAX is going to solve all of their problems....

        Almost, but I would also lay the blame on the "when all you have is a hammer, all the world's a nail" developers who insist upon putting EVERYTHING into a database, because "that's how you make websites DUH LOL KTHXBYE". So every image, every bit of text, every CSS file, all require accessing a database, rather than a simple pull of a file.

        Sure, you can have a database table that maps images to URLs, so that the code that IS accessing the database to build the page can find what URL to embed, and you can ch

        • Sure, you can have a database table that maps images to URLs, so that the code that IS accessing the database to build the page can find what URL to embed, and you can change which image to use by altering the database table, but the ACTUAL IMAGE FETCH should be nothing more than pulling a file from the web server's file system. Nice, simple, and it allows the web server to correctly handle IF-MODIFIED-SINCE and Expires.

          What is a file system, other than a database that maps paths to blobs?

    • by Ant P. ( 974313 )

      The way most sites implement it, all AJAX does is offload the server's page-rendering load onto the user's CPU, because the developers are too lazy to write efficient server-side code.

      There was a comment on that story a while back about the MJ death news slowing the entire web down, pointing out that Twitter crashed because some abysmally low number of requests per second overloaded it.

      Cars won't end the world through an energy crisis, Web 2.0 developers will.

  • by Anonymous Coward on Wednesday July 22, 2009 @01:51PM (#28784639)

    See slashdot, slow loading pages == bad.

    I hope this book helps you.

  • Slow sites lose me (Score:5, Interesting)

    by WiiVault ( 1039946 ) on Wednesday July 22, 2009 @01:54PM (#28784675)
    I used to go to VGChartz all the time until a few months ago they "updated" their site with seeming dozens of constant flash ads, talking, moving, popping up on the forums, ect. It got to the point where my Core2Duo desktop was litterally pausing for 5 seconds everytime I navigated to a new page. Pretty quickly I realized that the content on the site wasn't worth my time. I see this happening more and more. Sure I could have used adblock but frankly they were asking for a price (getting attacked by ads) and I chose not to pay it (leaving the site). I was glad to see a few ads, but abuse is a sure sign you have little respect for your readers.
    • by Inda ( 580031 )

      This site below lost me way back. Click the link, wait 15 seconds. Yes, 15 seconds on a 50mbit connection. I wouldn't have waited that long on 56k.

      One of my cores is running at 100%. 100% on a frigging webpage. I play 3D games on this 2-year-old laptop without maxing the CPU.

      Scroll up and down and watch the page tear. Yes, tearing on a 2D webpage. What are they thinking?

      The funny thing is though, the RSS feed has nice pictures, and full article text. They fail on so many levels.

      http://www.tottenham [tottenhamhotspur.com]

  • by tcopeland ( 32225 ) <tom AT thomasleecopeland DOT com> on Wednesday July 22, 2009 @01:59PM (#28784765) Homepage

    ...when your site is slow is to fire up YSlow [yahoo.com] and see what it says. Sometimes all you have to do is enable mod_expires in Apache to get a huge performance increase and a much lighter server load.

    If YSlow doesn't flag anything, then you've got to start digging. But at least you've eliminated some of the easy fixes.

    • You probably already know this, but since it wasn't mentioned in the review: YSlow was also written by Steve Souders (the author of Even Faster Web Sites).

  • by intx13 ( 808988 ) on Wednesday July 22, 2009 @02:00PM (#28784777) Homepage
    In my opinion, tuning a Javascript laden website for speed is an exercise in futility. The Web's great difference is user-extensibility; practically anyone can throw together some HTML and make a website. It may be a MySpace page, but hey, it's something, and it contributes to the culture of the Internet. Speed, however, is not a feature of the HTML/Javascript/CSS world.

    Were I to be developing a new AJAX-driven Web application I would focus first on simplicity. I feel that if you have so much AJAX stuff going on that you need to resort to crazy tricks you have already lost. Take the following, quoted from the review:

    In the subsequent chapter, "Flushing the Document Early," Steve explores the approach of utilizing chunked encoding in order to begin rendering the Web page before its full contents have been downloaded to the browser.

    While good advice, something like this should be implemented as a natural part of the specification, or not at all. This rings to me as an attempt to manhandle HTML/Javascript/CSS into a use case for which it is not intended.

    I want to see a real protocol for webpages - something between Postscript (except less document-oriented; and yes I know about NeXT's work) and a windowing environment (except more constrained). Then, to preserve the ease of user-input, a simple HTML/XML-like layer. For 99% of the sites that are constructed directly by the user, original HTML with italics, colors, fonts, etc. is sufficient. For projects beyond the scope of Joe Facebook, a true system is needed that allows seperation of design and content. But all attempts thus far to do both, frankly, suck.

    • by godrik ( 1287354 )

      I completely agree with you, those technologies were not thought to be used like that. Thus, some effects are not directly possible and need serious hacking to be reached. HTML/Javascript/CSS are definitively not the right tool for what we are doing with it. The problem is that there is no tool/technology/language to replace it; and I can't see one coming.

      I am taking here the risk to be modded troll. I do not believe in so called "web applications". If you need an application, just right a classical applica

      • I do not believe in so called "web applications". If you need an application, just right a classical application that connects to remote servers.

        Assuming "right" means "write": Your audience includes people who use a computer that they do not own and people who use lockout-chipped platforms such as Wii and iPhone. These people generally do not have the privilege to install such "a classical application".

        We do not lack portable tool today. python is present on most (all?) operating systems.

        Python does not come with Microsoft Windows, and non-administrators are not permitted to install a Python interpreter.

        You can still even use java or .net

        A Java program that does not require installation is called an applet, and web apps prior to widespread use of AJAX were often writt

    • by commodore64_love ( 1445365 ) on Wednesday July 22, 2009 @02:42PM (#28785419) Journal

      >>>>>begin rendering the Web page before its full contents have been downloaded to the browser.

      >>This rings to me as an attempt to manhandle HTML/Javascript/CSS into a use case for which it is not intended.

      I disagree. Today's websites don't do it, but in the simpler 1990s era of pure HTML, the website DID render before completing download. The browser was expected to grab the HTML first, render the page with "X" placeholders, and then download the images last. That way the user could read the website even with the image only partially present.

      So yes prerendering the webpage before download was completed *was* the original intent of the web. It is only lately that webpages have shifted away from that, and I for one would like to see them restore it.

      • Wait, no... reducing responsiveness is a feature. People go through a lot of trouble to prevent the infamous Flash of Unstyled Content.

    • While good advice, something like this should be implemented as a natural part of the specification, or not at all. This rings to me as an attempt to manhandle HTML/Javascript/CSS into a use case for which it is not intended.

      HTML already does chunked loading. This is one of the big reasons you shouldn't use tables for layout. Since most tables are set to use auto layout, the browser has to wait for the table content before it can start rendering.

      • This is one of the big reasons you shouldn't use tables for layout. Since most tables are set to use auto layout, the browser has to wait for the table content before it can start rendering.

        But what's the widely compatible alternative? Some document layouts fit better into a grid model than a flow model. CSS supports grid layout (display: table, table-row, and table-cell), but the subset of CSS implemented in widely deployed versions of Windows Internet Explorer doesn't.

        • But what's the widely compatible alternative?

          There isn't.

          Well, that's not entirely true. You can use a handful of nested DIVs and another handful of CSS hacks to play nice with IE and get something very close to what used to work out-of-the-box with tables.

          But wait, it gets worse! If you make the mistake of adding a valid DOCTYPE to your HTML, you can't even use tables for layouts (at least, not 100% height/width layouts).

          HTML is a giant, festering boil of suck. You can't even just blame Microsoft; Mozil

          • When I use images, I make sure to provide the size parameters so the browser can display the page even if the images are still downloading.

            Can't you do something similar with tables, such that the table can be displayed even if the content hasn't downloaded yet?

  • Wait a second... (Score:3, Insightful)

    by damn_registrars ( 1103043 ) <damn.registrars@gmail.com> on Wednesday July 22, 2009 @02:08PM (#28784925) Homepage Journal
    At first I thought this would help the slashdot coders, until I read more closely:

    Slow Web page loading can discourage visitors to a site more than any other problem, regardless of how attractive or feature-rich the given site might be

    Attractive and feature-rich are not adjectives that are appropriate to apply to slashdot it its current form. Hence I don't think this book is the cure to what ails them.

  • Or how about... (Score:5, Insightful)

    by castironpigeon ( 1056188 ) on Wednesday July 22, 2009 @02:09PM (#28784953)
    Most websites don't need to be ridiculously complicated to be effective. Go ahead and call me a luddite, but does everyone need comments on their website? Fancy sliding menus? Pop-up image galleries? Flash or other web programming up the wazoo?

    When did simple, efficient web design that presents content well and doesn't get in the way of it get tossed out the window?
    • by godrik ( 1287354 )

      I completely agree with you. I am often browsing website on low powered device and sometimes with low bandwidth and it is so slow...

      And there are website that can not be browsed if you do display images which you really don't want to do when you have a low bandwidth

      webmasters, just strip the shit out of your pages !

    • by thedonger ( 1317951 ) on Wednesday July 22, 2009 @02:35PM (#28785339)

      Most web sites do not need to be as ridiculously complicated to look so complicated. The web is relatively young, and AJAXification even younger. Give it time and common sense will prevail.

      The semantic web, HTML5, and CSS3 will eventually usher in an era of peace and tranquility. Music will be free. Passwords will no longer be necessary. Web sites will design themselves to look different to everyone (like the alien in Contact looking like Jodie Foster's dad, or the talking taco pooping ice cream in South Park), appearing in a form with which the user feels most comfortable. Accessibility and affordance [wikipedia.org][wikipedia] will be implicit. Fluid design will exist in four dimensions, and tables will only be used for tabular data. JavaScript will be so unobtrusive it may never load, but that will be OK because all web sites will be semantic and thus 100% accessible and functional without it. And Noscript will no longer be needed on the browser because it will be installed at the ISP level...

      Whoops. Sorry. I dozed off there but I kept typing while I was dreaming...

  • by Anonymous Coward on Wednesday July 22, 2009 @02:31PM (#28785273)

    I read this guy's other book, "Clearer thought using vodka". Great read. Can't remember much of it.

  • K.I.S.S. (Score:4, Insightful)

    by Tablizer ( 95088 ) on Wednesday July 22, 2009 @02:32PM (#28785281) Journal

    The fastest websites I see generally use plain-jane HTML and pre-sized graphics (via size attributes on IMG tag). AJAX, CSS, and too much JavaScript seem to cause confusing UI behavior, unexpected pausing while something fancy is doing GC, or jerky scroll motion that can be perceived as sluggishness, especially on older hardware.

    • I also wonder how much sluggishness is caused by putting everything in a database. I tried implementing a simple shopping-cart system for my website a while ago, where I have a very small handful of products I sell using Paypal buttons currently, and tried using some typical open-source packages like OScommerce. All of them make very heavy use of MySQL for every little thing: configuration settings, the actual content on each page, even images! This really seemed like overkill to me, and wasn't exactly f

  • by Anonymous Coward on Wednesday July 22, 2009 @02:39PM (#28785387)

    I've seen a lot of good sites "re-designed" into oblivion. I think it's mostly a social problem.

    1. The "we're artistes" mentality. Unless your site is explicitly an art site (which it usually isn't), people will just want information from your site. You could do most of that with static HTML and a few simple images.

    2. The "we're losing eyeballs, we need to do something" problem. The site design is an easy target; but it may not be your site design. Maybe your content is just boring. Competition moved in. You were participating in a fad. The honeymoon is over. The public tastes change. You can't control those things, but you can control your site design so that's what you do. It's like pushing the walk button at an intersection when somebody else has already pushed it.

    3. Now, this is the one I really hate to say; but here goes... "You can't figure out how to let people go". The project is done. You need a few maintenance coders, the most experienced people. OK, maybe there's some justification for keeping staff around for new projects; but I bet often there isn't. If you're directly involved with the developers, you don't want to do this. The risk is that you'll spark brain-drain and cause the people you really need to leave. It's far less risky for you to keep everybody, even the grunts. Better yet, you can go up to the next level of management and tell them that site redesign is needed. That makes you the hero instead of the villain. The next level up from you is probably not going to figure out that you are just trying to keep your department busy. Perhaps this problem could be solved by making it clear right from the start that there will be a permanent staff and some contract workers. Instead, I bet there are a lot of companies where the web developers figured out how to justify their existance--to the detrement of the site, and the company.

    4. Cool new technology. 'nuff said.

    5. People just aren't that bright. They dynamicly generate javascript via a scripting language with a framework on top of it. Just to put "Hello World" on a page. Look at me! I'm a web developer.

  • Google Mail is, hands down, *the* slowest webapp that I deal with. I'm not talking about the copout "basic HTML" option. I run any number of browsers (FF, Opera, Safari), and *all* take tens of seconds to load and interact with Google Mail. I dream about sharp needles in the eye as I wait for Google Mail to do its...magic.

    So the author of this book is a "web performance expert" who is now employed by Google. Instead of writing books, maybe said author should figure out what's wrong with the Google Mail

  • This page takes six seconds to load. "data.coremetrics.com" is slow today. "c.fsdn" is slow, too. "doubleclick.com" only wasted about a second.

    Maybe someone at Slashdot should read the book.

  • by greenreaper ( 205818 ) on Wednesday July 22, 2009 @04:28PM (#28787079) Homepage Journal
    The average webmaster doesn't need this book; they need tools like Page Speed [google.com] that highlight very simple issues that webmasters can fix, usually without any code changes. And they need clear instructions on how to do it, and why it will benefit their users. I recently ran a project to cut the fluff [wikifur.com] on large furry websites. Most failed to even gzip their HTML, CSS or JS, or were resizing images into thumbnails. It's not hard, people just don't think to do it, or don't know how.
  • Efficient code? (Score:3, Insightful)

    by dontmakemethink ( 1186169 ) on Wednesday July 22, 2009 @04:57PM (#28787599)
    Someone advocating more efficient coding of websites should surely take less than 12 paragraphs to do so!
  • by roc97007 ( 608802 ) on Wednesday July 22, 2009 @05:23PM (#28788055) Journal

    In my experience (having 20/5 Mbit fibre to the house) is that it's almost never the website's local content that loads slow, it's the damned ads. Let me say that again so that we're clear -- I'm sitting there waiting not for the content I was seeking to load, but for the ads that I don't want to load. I know there's work-arounds for this -- I use some of them and they do help. But Fred and Ethyl Nongeek is not going to know about ad suppression. To them the 'net is "just slow" and they don't know why.

    Making your local content lightning fast isn't going to help if your customers are waiting for unwanted content from some other website. And (grrrr....) this includes CAPTCHA! Sometimes it takes almost a half minute for the captcha image to come up, long after the rest of the login page has loaded.

    The worst example, and a sign of things to come, was when Google Ads went down awhile back and took a bunch of websites down with it, including (as I remember) Slashdot. If your page included google ads, it just wouldn't load. I don't think AJAX is going to help with that. Feel free to disagree, but be specific.

    • The worst example, and a sign of things to come, was when Google Ads went down awhile back and took a bunch of websites down with it, including (as I remember) Slashdot. If your page included google ads, it just wouldn't load. I don't think AJAX is going to help with that. Feel free to disagree, but be specific.

      Well, it's kind of obvious: Ads should support loading through AJAX.

      "We'll take your site down with us" is not a great selling point.

  • Slow websites ... (Score:2, Insightful)

    by SlashDev ( 627697 )
    ... are slow mostly due to advertising. That being said, slow websites can also be due to: - Limited bandwidth where the site is hosted - Slow disk subsystem - Bandwidth throttling due to a large number of users
  • My personal website has been dying a slow death, not because of the fancy graphics I use on it but because I'm stuck on shared hosting with 100+ other websites (1and1). That alone can make the difference. To download 100K to 300K isn't a big deal now days.

    Also, for instance try tallying up all of Facebook's JS, CS + HTML files, and you'll discover the site clocks in over 1MB. Now consider how fast the pages on the site load (super fast) See the difference how your site is hosted can make?

He has not acquired a fortune; the fortune has acquired him. -- Bion

Working...