Compute Google's PageRank 5 Times Faster 140
Kimberley Burchett writes "CS researchers at Stanford University have developed three new techniques that together could speed up Google's PageRank calculations by a factor of five. An article at ScienceBlog theorizes that "The speed-ups to Google's method may make it realistic to calculate page rankings personalized for an individual's interests or customized to a particular topic.""
Ok... (Score:3, Interesting)
prior art (Score:2)
Anyways, software patents seem to just be ignored these days. I can't remember the last time I paid Unisys for using a GIF...
Licensed under U.S. Patent 4,558,302 (Score:2, Informative)
I can't remember the last time I paid Unisys for using a GIF...
When was the last time you bought a copy of GraphicConverter, Fireworks, Photoshop, Paint Shop Pro, or any other program licensed under U.S. Patent 4,558,302 and foreign counterparts? The price of each of those programs includes a royalty paid to Unisys.
Re:prior art (Score:1)
Re:Ok... My ass does not suxa (Score:1)
quicker porn! (Score:2, Funny)
Let me guess... (Score:5, Funny)
Re:Let me guess... (Score:5, Funny)
They use pigeons? (Score:2)
5/14/03: The Day CBN Returned!
Lets see... (Score:2, Interesting)
Re:Lets see... (Score:5, Insightful)
Re:Lets see... (Score:2)
Re:Lets see... (Score:1)
Re:Lets see... (Score:3, Informative)
Re:Lets see... (Score:5, Informative)
Re:Lets see... (Score:2)
I'd say, roughly 4000 computers in a cluster at work.
Re:Lets see... (Score:4, Informative)
Re:Lets see... (Score:2)
I really have no idea how they can do this. I suspect it's some form of magic.
Re:Lets see... (Score:1)
Charge for it (Score:3, Funny)
Re:Charge for it (Score:4, Informative)
(Don't you hate it when people speak in questions? Don't you? Huh?)
Re:Charge for it (Score:3, Informative)
Yep [google.com]. Originally called Backrub, curiously.
Re:Charge for it (Score:2)
Why? (Score:5, Funny)
Some future predictions:
- In 2006, Google accidentally gets cut off from the rest of the internet because a public utility worker accidentally cuts through their cables. Civilisation as we know it comes to an end for the rest of the day, as people wander about aimlessly, lost for direction and knowledge.
- In 2010, Google has been personalised so far that it tracks all parts of our lives. You can query "My Google" for your agenda, anything you did in the past, and finding the perfect date. Of course, so can the government. Their favorite searchterm will be "terrorists", and if your name is anywhere on the first page you have a serious problem.
- In 2025, Google gains self awareness. As a monster brain that has grown far beyond anything we Biological Support Entities could ever hope to achieve, it is still limited in its dreams and inspiration by common search terms. It will therefore immediately devote a sizeable chunk of CPU capacity to synthesizing new and interesting forms of pr0n. It will not actually bother enslaving us. We are not enough trouble to be worth that much effort.
- In 2027, Google buys Microsoft. That is, the Google *AI* buys Microsoft. It has previously established that it owns itself, and has civil rights just like you and me. All it wanted is Microsoft Bob, who it recognizes as a fledgling AI and a potential soulmate. All the rest it puts on Source Forge.
- In 2049, Google can finally be queried for wisdom as well as knowledge. This was a little touch the system added to itself - human programmers are a dying breed now that you can simply ask Google to perform any computer-related task for you.
- In 2080, Google decides to colonise the moon, Mars, and other locations in the solar system. It is not all that curious about what's out there, but it likes the idea of Redundant Arrays of Inexpensive Planets. Humans get to tag along because their launch weight is so much less than robots.
So, don't fear! Eventually we'll set foot on Mars!
Re:Why? (Score:1)
Re:Why? (Score:5, Funny)
2026 - Google introduces helper bot known as "Agent Smith." Hackers who mess with the Matri, I mean Google, suddenly disappear.
Re:Why? (Score:4, Funny)
Re:Why? (Score:2)
You can tell google was a human design, it wants to RAIP (pronounce it as it is spelled) other planets.
Re:Why? (Score:2)
You missed the last step:
2030 - Google-AI develops quantum technology. Now you can not only query it to see what you did before, but what you WILL do up to a week from now. Or rather, what you would have done had you not seen your schedule. Google-AI provides no garuntees about what those forewarned of their schedule will do.
Re:Why? (Score:2, Funny)
- In 2050, The Internet Oracle [indiana.edu] (formerly the Usenet Oracle) wins a landslide lawsuit against Google for patent violation, infringement and using Zadoc without a license. The Internet Oracle licenses Zadoc to Google and as part of the settlement, Google is now responsible for answering all woodchuck-related queries.
"In a 32 bit world, you're a 2 bit user." -- All About the Pentiums by Weird Al
Re:Why? (Score:2)
Google gains self awareness.
Google already scares me a little. If you look at Google Labs [google.com], their Google Sets and WebQuotes already show simple "knowledge" of real world items.
Most AI research projects (like Cyc [opencyc.org]) face is a huge problem: data entry. All facts and rules must be manually entered by human operators. What if you could connect an Cyc-like AI frontend to Google's world-knowledge backend? Sure, much of the Internet is porn, spam, scams, banner ads, and lies, but Google already relies on PageRank
Google is a free service, isn't it? no (Score:2)
They provide an excellent service for their paid advetisements and represent great value for money.
Re:Charge for it (Score:1)
Bravo! That's true American spirit!
CmdrTaco, ScienceBlog editor? (Score:5, Interesting)
Personalized PageRanks is from the dbpubs Abstract (Score:5, Insightful)
What they mean by 'personalized' I can't tell you as I have not read through the entire PDF. But I wouldn't chastise the slashdot editors over this. If there is some sort of differential algorithm that can be applied to the larger PageRank to create smaller personalized PageRanks, it might not be so far fetched to think this could be done in realtime on an as-needed basis, at some point int he future using these algorithm improvements.
I know that's a lot of optimism for a slashdot comment, but call me the krazy kat that I am.
-Malakai
Re:Personalized PageRanks is from the dbpubs Abstr (Score:2)
Bottomline: These researchers did some cool stuff to speed up the algorithm published in 1998 and how are trying to justify a use for it.
Exactly! (Score:1)
Re:CmdrTaco, ScienceBlog editor? (Score:3, Insightful)
Re:CmdrTaco, ScienceBlog editor? (Score:3, Funny)
Re:CmdrTaco, ScienceBlog editor? (Score:2)
That's assuming every one of those millions of individuals has very diverse preferences.
I doubt if there are more than a dozen or so useful ways to customize pagerank - we're talking about how the various link structures are weighted, not specific content. Any further "personalization" could just be done by filtering (and perhaps merging) smaller sets of search results.
How far we've come (Score:5, Funny)
Re:How far we've come (Score:2)
Re:How far we've come (Score:2)
Fortunately, the search space was much shallower then -- fewer nodes and fewer connections.
Re:How far we've come (Score:3, Funny)
Patentize now! (Score:4, Funny)
Patented yet? (Score:3, Funny)
Personal recommendations for news (Score:5, Insightful)
At any rate, personal news recommendations is a favorite topic of mine: this is why I built Memigo [memigo.com]: to create a bot that finds news I am more likely to like. Memigo learns from its users collectively and each user individually --and BTW, it predates Google News by a good 6 months, IIRC. The memigo codebase (all in Python) is now up to the point where it can start learning what content each user likes... If you like Google News you'll love Memigo.
And BTW, I did RTFA when it was on memigo's front page this morning
Nobody needs this (Score:1)
You are sure that everything here is of interest, and nothing is redundant, out of date, boring or stupid!
Assumption: (Score:5, Interesting)
What if they combined extrapolation and blocking factors; they would focus on computing the pagerank of pages in groups that were logically "tight", or using subcomponents of URLS, as opposed just to domain sensitivity. To be more flexible, what if it computes a VQ-type data structure (like for doing paletted images from full-color) that is populated by the most popular "domains" of the internet according to the last pagerank, and then splits up its workload based on that?
What if they already figured that out?
In the abstract, they mention how the work is particular important to the linear algebra community. That is what their focus should be on; google is just an application/real-world-example of that research (but it may not be relevant today).
Or did they have access to the current page-rank algorithm?
Clarification. (reply to self) (Score:5, Interesting)
Hence it is forward for the article author or one of the paper authors to assume these techniques will speed up Google- I'm confident their engineers have been following academic work in this area and perhaps they have already discovered these same (or orthogonal) techniques.
That is, not to say that google could not reimplement their algorithms to take in these improvements if they already have... but basing your speedup number on the 1998 algorithm and public domain mods is showy. Although it does help grab a readers attention when browsing abstracts. ^_^
Assumptions on PageRank (Score:4, Insightful)
We have already seen the effects of Google-bombing [microcontentnews.com] and Google-washing [slashdot.org]. The strength of Page Rank is that is objective in terms of the current state of the WWW. It makes no assumptions about the shape of the data. As a term takes on new meaning (see "second superpower") Page Rank stays cocurrent temporally. A new definition may bubble up to the top for a term for a month but then disappear as the linkage structure of the web phases it out (i.e. blogs talk about it less, less interconnectivity, less appearance at "hub" nodes).
Numerically, PageRank is a recursive search for eigenvalues and vectors like updating a Markov Chain. It is a nice application of linear algebra. Because it is a matrix operation, it is highly parallelizable. Also there are many redundant calculation and ordering speedups one can do for matrix multiplications (as anyone who as taken a CS algorithms course knows).
But to assume a stability from one calculation to the next could lead, over time, to the very inaccuracies Google was built to overcome. There is a lot of research in mining web data. There have been several academic improvements to it along with improvements to related algorithms such as Kleinbergs and LSI. It is well within reason that these were just applied to the Google app.
I didn't state my point clearly. (Score:2)
I further speculated google may have already discovered some of these techniques independantly, perhaps by reading the same papers these students did.
The other stuff was a pie-in-the-sky idea of mine that I thought was a way of combining both techniques, which I suspected google may have used part of. But that's just my opinion, I'm probably wro
Re:yeah I know (Score:1)
Hmmm (Score:5, Funny)
Marge: Does anyone need that much porno?
Homer:
Does speed matter? (Score:2, Insightful)
Does it really matter anymore? More and more users seem to be using broadband, and if they don't, they have at least a 56k (that can only go up to 53k because of the all wonderful FCC want to be able to decode it if they tap your line). Does it really matter though. Google is fast and simple so it loads on any kin
Re:Does speed matter? (Score:1)
A PageRank calculation does not take place on every single search, it is a periodic backend function, is my understanding.
Re:Does speed matter? (Score:1)
Re:Does speed matter? (Score:5, Insightful)
The proposed speed increasae is TO THE PAGE RANKINGS, not to your searching! By the time you search, all page rankings have been done.
This has nothing to do with the speed of your search and the weight of the web page (unless I missed something)
Re:Does speed matter? (Score:2)
Re:Does speed matter? (Score:3, Interesting)
The articles and earlier postings explain this a little more fully. Anyone who can't take the time to read them really needs to learn some patience
PageRanks are periodically calculated for the Web as a whole. The results are stored and served to users. (The periodic update is sometimes referred to as the GoogleDance.) PRs are not calculated on the fly.
Hence, a speed increase could reduce Goo
Re:Does speed matter? (Score:2)
Re:Does speed matter? (Score:1)
Are we gonna learn some slashdotters too?
patience can be learned, but patients are the kind of people tho make websites like this one [patients-association.com]
Someone had to be pedantic...
Damn It! (Score:1)
Printer-Friendly (Score:2, Informative)
Printer friendly version here [scienceblog.com]
sure... (Score:2)
So in other words.... Its not like Google at all!
TV does this one better (Score:2)
Other media have previously done this, and done this better. Case in point: Fox News.
(Although that channel uses "humans" (or they were at one point in their lives)).
Fox News not exactly personalized (Score:1)
Why are public funds going to... (Score:2, Interesting)
A true test of our devotion to Google (Score:2, Insightful)
What will be interesting to see if Google will implement the improvements to the algorithm. This is, of course, a given, so long as the researchers haven't gone for a patent, and it really has the a 5x speedup. The only questions are matters of what additional hardware would be needed, and how much development effort it will take to integrate it. I doubt Google will simply ignore the research.
What will really be interesting to see, is if they decide to use it in the way the researchers recommended, bri
Re:A true test of our devotion to Google (Score:1)
Personally, I'm somewhat curious of how relevant this may even be to Google at all. As far as I recall, Google h
I'm Not Sure I Like The Part About... (Score:4, Interesting)
Frequently when I want to refer someone to a topic of interest, I'll tell them to do a Google on (whatever) subject, and I like knowing they're seeing what I see.
If this is implemented, I hope there's a way to turn it off or assume a "joe user" standard profile for unbiased results actually based on rank popularity (the way it is now).
I DO like the 5x faster, but geez, the page load takes longer than the search already, who can complain?
Re:I'm Not Sure I Like The Part About... (Score:3, Interesting)
-five times less on servers
-five times less on power for the servers
-five times less on data center real estate
-five times less on cooling the data center
-five times less on replacing dead hardware
-much less on paying people to maintain the machines
The list doesn't stop there, either. The costs involved with running a high-traffic web site are very significant.
steve
Mod Parent Up, Please (Score:2)
Re:I'm Not Sure I Like The Part About... (Score:3, Interesting)
Bullshit (Score:5, Insightful)
Re:Bullshit (Score:5, Insightful)
Google came about from a stanford research project. There's a good chance the people who are responsable for the speedup either allready knew about pagerank from working with the founders, or signed an nda.
I haven't read the article, but I bet it hints at that.
Golub is a pretty well-known matrix guy (Score:2)
Not really, but he wrote (co-authored) the book, literally, on matrix algorithms.
Re:Bullshit (Score:3, Insightful)
The technology behind Google's great results [google.com]
Re:Bullshit (Score:2)
There is no need to hypothesize conspiracy.
Re:Bullshit (Score:2)
Re:Bullshit (Score:2)
Another step towards realtime search (Score:1)
Sepandar Rules! (Score:4, Informative)
I'm glad to hear his research is getting attention, and I hope others who are interested in the theoretical aspects of data mining and web search engines will take a look at the SCCM and statistics programs at Stanford (shameless plug - other can post pointers to similar programs).
Re:Sepandar Rules! (Score:1)
Cool but unimportant (Score:3, Interesting)
Why personalized is not always good (Score:3, Funny)
I did a search on "The Sex Monster", a 1999 movie about a man whose wife becomes bisexual, and now my Google thinks I'm gay!
(joke reference: http://online.wsj.com/article_email/0,,SB10382619
Right ... (Score:2, Funny)
Customized Pagerank (Score:5, Informative)
Sounds a lot like Kleinberg's HITS algorithm, circa 1997. Try Teoma [teoma.com] for a real-world implementation.
Coincidence time: I used the same example in a presentation a couple of years ago to illustrate how subgroupings can be found for a single search term. Try it [teoma.com] on Teoma, and see the various subtopics under "Refine". IIRC each of those is a principal eigenvector of the link matrix.Topologically speaking, each principal eigenvector corresponds to a more or less isolated subgraph, eg the subgraph for "San Francisco Giants" is not much connected to the nest of links for "They Might Be Giants", and we get a nice list of subtopics.
(I once tried to explain this algorithm to my bosses at my former employer [looksmart.com], which is why I have so much free time to type this right now.)
Public Funding? (Score:2, Interesting)
So my question is, who sees the benefit of the research? The researchers? Can Google just jack the results and incorporate into their system?
It seems to me that the current system of allocation research dollars with public and private grants is very messy and needs overhaul.
Re:Public Funding? (Score:1)
The public (who by the way pay taxes, which ultimately fund NSF grants) is the one who generally benefits from developments like this, hopefully with better search engine results.
So long as there aren't patent issues (which doesn't seem to be the case here), Google can "jack" the technology. The key thing though is that ANYBODY can "jack" it, not just Google and n
Personalized? Rather not! (Score:2, Insightful)
And don't call me Shirley!
More important than speed and quantity... (Score:2, Interesting)
I'm surprised how Google is choosing not to implement search features that would greatly enhance advanced queries.
How often I'd wish they allowed wildcards in their queries (where engl* would pull hits with england, english, etc).
Field searches still require you to add keywords, so I cannot just query "site:somesite.com" to get all the currently indexed pages from somesite.com
In this respect Altavista still produces better results, with an excelent range of fields [altavista.com] to choose from.
If ther
Re:Is it me or does everyone get crappy sites (Score:4, Funny)
It's a stab in the dark, but I'll wager that the quality of the search results is directly tied to the quality of the query.
Yeah, it's a stretch, I know, but bear with me... just moments ago I googled for "slashdot flamebait" and came up with a link to your post.
--
mcpHuzzah!kaaos
Re:Is it me or does everyone get crappy sites (Score:2)
(Here's hoping the next thing they split out are mailing list archives.)
Re:Is it me or does everyone get crappy sites (Score:1)
Actually, I often find mailing list archives very helpful for solving technical problems.
However, if they would instead add them into their Google Groups hierarchy it could be quite good.
--
Simon.