Science Project Quadruples Surfing Speed - Reportedly 646
johnp. writes "A computer browser that is said to least quadruple surfing speeds on the Internet has won the top prize at an Irish exhibition for young scientists, it was announced on Saturday. Adnan Osmani, 16, a student at Saint Finian's College in Mullingar, central Ireland spent 18 months writing 780,000 lines of computer code to develop the browser. Known as "XWEBS", the system works with an ordinary Internet connection using a 56K modem on a normal telephone line.
" A number of people had submitted this over the weekend - there's absolutely no hard data that I can find to go along with this, so if you find anything more on it, plz. post below - somehow 1500 lines of code per day, "every media player" built in doesn't ring true for me.
ooh, sign me up! (Score:2)
Basic maths. (Score:4, Interesting)
Re:Basic maths. (Score:5, Funny)
you see, it's not that hard to make 1500 lines of code per day!
Re:Basic maths. (Score:3, Funny)
Re:Basic maths. (Score:2, Interesting)
Re:Basic maths. (Score:4, Interesting)
Most likely he has taken an open source browser and added in his own extensions. This is the type of innovation that making the browser open source is meant to support.
As for speeding up browsing by a factor 100% that is pretty easy. We did a lot of work on HTTP-NG and you can speed up downloads a lot just by compressing the headers so that they fit into a single packet. HTML is also very compressible. The biggest mistake we made in the original Web code was not putting a lightweight compression scheme into the code, although it did make it into the specs.
Of course the reason this did not happen was the LZ patent GIF fiasco and the then state of the GNU compression libraries. Even so Microsoft has supported compression in IE for some time.
I am rather more skeptical about the 500% claim. I don't think that there is that much redundancy unless you have completely artificial examples.
Re:Basic maths. (Score:2, Funny)
Do curly brackets on a single line count?
}
Re:Basic maths. (Score:5, Funny)
Re:Basic maths. (Score:5, Funny)
I wonder what ever happened to reading webpages while surfing... Ooohh, right, so that's what you call surfing
Re:Basic maths. (Score:3, Funny)
Re:Basic maths. (Score:5, Interesting)
Indeed. I remember reading that IBM reckon that, including design, coding, testing, debugging and documentation, a programmer's doing well to get 10 lines of code per day, averaged over the life of the project.
Also depends how he's counting lines. In C, because that can vary so much depending on individual formatting style, a good rule of thumb is to count semicolons. And even then it won't tell you if programmer A is writing fast but hard to read code and programmer B is checking the return value of every system call (as you're supposed to but few ever do), adding lines and robustness with no extra actual functionality.
Re:Basic maths. (Score:3, Interesting)
From my software engineering course way back in college I think I remember the number being 4 or 5. But that is more like an industry average. One thing about software is that the best programmers are something like two to three orders of magnitude more productive than the average. Between that and the communication costs growing exponentially in a group you find that a few very talented programmers are vastly more productive than a mass of average programmers.
Still, sustaining 1,500 LOC per day for a year and a half ... that's beyond the productivity level of anyone I've ever seen. I personally have managed 4,500 per day for a period of about a week on occasion ... but I wasn't sleeping much during that period.
I am not sure I'd take that number at face value though. If this were real he would almost certainly be using a lot of prewritten code for codecs and the like and that would balloon the LOC for little effort on his part. It's more than a little unlikely that he'd be able to write all his own codecs in the first place.
So, while the LOC sounds specious, it's potentially believable given the probability of code reuse.
The thing that makes this entirely unbelievable is the performance claim. 4x performance of existing browsers over a 56k line? That's simply not possible since the bulk of time spent waiting is data transmission time. That could be improved but only with a server side component and it's doubtful it could be improved substantially without a large loss in quality.
I'm not going to dismiss the claim of a new web browser, but I'd be surprised if any of the size and performance claims hold water.
Re:Basic maths. (Score:5, Insightful)
I broke 1,000 LOC per day for about a week while working for an unnamed gigantic CPU monopolist. I was behind schedule, over budget, and had a hard deadline, and the code itself was fairly repetitive and not terribly efficient. Ordinarily, I'd figure I produce closer to 250 LOC per day during a normal coding period.
Provided this story isn't complete hogwash, my guess is that the reporter asked the boy about the writing the program and he answered that it consisted of 780,000 LOC and took him a year and a half to build. He probably neglected to mention that 90% of those lines were in libraries written by other people. He may not have even intended to be deceptive in any way, figuring that any fool would know that was the case, but not realizing that the reporter was a fool.
Re:Basic maths. (Score:5, Informative)
when I had "software engineering" in my computer science courses, we got this figures for LOC per say:
Application programs: 25 - 100
Service programs: 5 - 25
System programs: 1
Application programs are things like an editor (albeit some editors are rather complex), service programs are things like cc and ld or asm (albeit some of them are not "that" complex) system programs are stuff like the kernal itself or, dynamic link loaders, device drivers etc.
Well,
we all know that LOC is not a defined "value" but people working a lot with that "measure" just define it
E.g. if you work with COCOMO or with PSP(personal software process) the typical LOC is defined as a single definition, a single expression(some even say every part of an expression), an argument to a function call, every include, every define and so on:
fprintf(stderr, "this is an error number: %ld", errnum);
That would be 4 LOC, one LOC for the "statement" and 3 for the 3 arguments. Consider you can make an error/bug in every argument or 'misstype' fprintf for fscanf
LOCs do not realy get interesting in comparing hero programmers (10 to 20 times more effective) with standard programmers, but by comparing programming languages!!!
The VERY INTERESTING point about LOCs is that the noted rules of thumb above are independend from programming languages!!!
A programmer writing lets say 12 LOC per day C also writes ~ 12 LOC per day in assembler, in LISP in PERL or what ever language is appointed for the project.
So: the more expressive and the more abstract a language is the more "algorithm" or "computation" is defined in the lines of code.
In other words: 10 lines of C are far more calculation than 10 lines of assembler, while 10 lines of LISP, SQL or Prolog are even more than C.
Bottom line: the number of statements the average programmer can write depends far more on the problem domain than on the language choosen!
Well, the productivity of the so called hero programmer is in general not in lines of code, but in "abstractions" he implemetns. Or in number of features he implements. And that is often acomplished by choosing the right language constructs(not by writing more lines)
angel'o'sphere
Re:Basic maths. (Score:4, Insightful)
I am not surprised you don't believe me. It's hard to believe it myself, and I'm the one that has done it. I do not suggest that that level of productivity is typical of either myself or others nor sustainable over a long period of time. My long-term average is less than a tenth that (and dropping as I get older).
The state I'm in when coding like that is best described as a fugue state. My mind is racing and everything else just gets ignored. Meals. Sleeping. I call it "going under" because that's what it feels like when I come out of it. The productivity that I see during such periods is prodigious to say the least.
But it's absolutely brutal on the body. You used a 16 hour day in your calculation, but that's understating it by nearly 50%. Because, in that state, I'm not sleeping at all. I'm incapable of sleep. Nor am I taking regular meal breaks. This allows coding for about 23 hours per day.
Typically when I get into that state it only lasts for about two days (40-50 hours), but there have been a handful of times when it has lasted longer ... as many as five days straight.
As for whether or not the code produced was trivial, the last such time I did this I wrote a Java debugger from scratch. Mostly that was UI code (Swing didn't exist at the time so I had to write a lot of rendering and layout components) but the class disassembly and debug engines were fairly complicated (but nowhere near the complexity of a JPEG or MPEG decoder!). It took 96 hours to write almost 14,000 LOC.
Now, there are two other interesting productivity data points. When coding in C or C++ (doesn't really matter which) my productivity maxes out at around 1,500 LOC in a day. Java triples it! And the bug rate falls by ~90% in Java. I love Java.
Anyway, that's my story, believe it or not. And, as such, it's not entirely unbelievable to me that someone could do 1,500 LOC/day for at least a few days at a shot. Doing it for 18 months straight though? Doesn't seem likely.
Re:Basic maths. (Score:5, Interesting)
have you ever been 16 (Score:2)
I've written a C++ windowing wrapper for windows, inluding string and varient datatypes (which are quite big in thenselfs) in less than a 5 day week, it contained few bugs and had 10000+ lines of code.
Working against RFC's and using other peoples designs makes coding easy, there's not much to think about and you can just sit there are curn away line after line, about as quickly as you can cut and paste.
I don't believe the speed imporvement though, TCP/IP over a 56k modem is TCP/IP over a 56K modem.
Re:have you ever been 16 (Score:5, Insightful)
I don't deny that it's possible to write 10,000+ lines of code in 5 days but, unless you're some sort of prodigy, I would have serious reservations about the quality of that code.
All of us who chose development as a career because we love to write code, rather than just because it's a well-paid and relatively easy-going job, have at some time cranked out amazing amounts of code in a short time. My doubts are caused by the duration. I don't believe that it's possible to sustain that sort of output for that period of time.
Re:have you ever been 16 (Score:2, Interesting)
I've always been proud that when looking back on my own projects I had something like 20-30 lines a day. On really good days you can write hundreds of lines but sometimes you have to throw everything out again because it's crap.
I hope this guy isn't for real, he'll be burnt out by the time he's 30.
Re:have you ever been 16 (Score:5, Funny)
10-20?! (Score:4, Insightful)
A good software engineer should be able to write at least a few hundred lines of code in one full day
Re:10-20?! (Score:4, Insightful)
Of course, if you only spew out code and do nothing else, then yes, 10-20 LOC is not very much.
Besides, how many programmers does Microsoft employ? How long are they working on Windows now? Let's assume the Windows source code contains about 10 million LOC - that's 500.000 days if one programmer writes 20 lines a day. Let us further assume, that 100 programmers are working on Windows. That's around 1.400 man-years or 14 years in our case. That's not unreasonable, is it?
Re:Basic maths. (Score:3, Interesting)
I must be getting old. In my younger days, that was called "libraries", and you only counted each line once, no matter how many times they were reused.
Re:Basic math[s]. (Score:3, Insightful)
Yes, but for a year-and-a-half? EVERY day? AND while, presumably, taking other classes and studying for tests in other courses, having friends, etc?
Think not.
Re:Basic math[s]. (Score:3, Funny)
Whoa...slow down on those assumptions there hoss...
--trb
Re:Basic maths. (Score:3, Funny)
Hmmmm he even admits it crashes (Score:4, Interesting)
This isn't a microprocessor - the speed it runs should be completely unrelated to it's caching.
I'm very very skeptical that this is anything more than a browser with precaching.
It also makes other ludicrous statements about how blind people can now access the web. I'm not sure how they do it presently, but i know that they do.
Warning! Your internet connection is not optomized (Score:5, Funny)
Re:Hmmmm he even admits it crashes (Score:3, Interesting)
There is also BLinux [leb.net], a project to make Linux accessible for blind users.
Jaws is pretty cool, I use it at work sometimes to test sites.
Re:Hmmmm he even admits it crashes (Score:5, Funny)
The downside on this would be that you age four times as fast while surfing, but hey, there's always something ...
Re:Hmmmm he even admits it crashes (Score:4, Funny)
Neat! I never thought that a web page would have a terminal velocity. Guess it's between 6 and 7 times the normal speed of web pages.
suspicious (Score:4, Informative)
He wants to study computer engineering in Harvard University and eventually set up his own Internet or computer company.
(For people who don't get it, Harvard's CS department, while reasonably good, is not exactly the obvious top pick among CS hotshots.)
Re:suspicious (Score:2, Funny)
Good God, he could make a fortune if he's on the internet! Genius.
Re:suspicious (Score:4, Informative)
I'd be suspicious about the alleged speed of writing code. (That's thousands of lines a day!) It seems to be like this is just a browser which loads up links ahead of displaying them. Which, amazingly enough, is what all those "Your Internet Connection Is Not Optimized!!!" programs do.
How doing this faster can make the computer crash is a bit of a mystery to me. (I can't think of a single program with a speed dial, and above a certain speed, the computer crashes...
Re:suspicious (Score:4, Insightful)
No this is either total bullshit, or a huge exearation. Remember, with real science, computer or otherwise, the MOST important part is subjecting work to peer review. Anyhting which can only be demonstrated in one lab in a hands-off, no-details demonstration isn't science and the person is hiding something.
Re:suspicious (Score:4, Informative)
He is in Ireland, but Dublin's no tech backwater. Trinity College Dublin is world-renowned for science and maths, and a short flight away are Imperial College and UCL in London, not to mention Oxford and Cambridge. A little further than that is the Sorbonne. There's no reason he shouldn't be as familiar with the rankings as anyone else.
And thanks to the Irish government's very sensible tax policy (i.e. less is better), the country has a sizeable presence of US high-tech firms, like Oracle and Sun.
As others have said, tho', anyone who claims to be able to sustain 1500 LOC/day for 18 months, is probably not to be taken seriously.
Re:MIT is better (Score:3, Interesting)
Speeding up browsing? (Score:2, Informative)
I've heard of tools in the past that claim to speed up browsing by cacheing ahead. These tools follow links on a page before you request them so that they are already in the browser's cache when you come to click on a link.
The other possibility is some heavy compression server side, but this would require a server module (e.g. mod_gzip) and this rules out any kind of built in compression in ppp, so the sppeedup would, I guess, not be as noticable as 5x.
Needless to say, I'm fairly sceptical that this is an actual speedup of browsing. If you can only fit 56Kbps down a line then you can only fit 56Kbps down a line...
Re:Speeding up browsing? (Score:4, Funny)
Yes, unless you have one of these! [slashdot.org]
I get it... (Score:4, Funny)
It then makes use of network magic. You mean no-one ever told you about the magic ?
Hmm. (Score:5, Interesting)
But what really got me where the two most important features someone could ever want in a Web Browser - it can play dvd's [it incorporates ever media player!], and also has a handy animated assisant called Pheobe.
Now, I am most probably wrong, and will happily eat my hat, but I cant help but feel that this isn't an entirely accurate article.
ps. Does anyone know if it is standard compliant ?
Re:Hmm. (Score:5, Interesting)
I have a feeling this project is nothing but hot air.
Re:Hmm. (Score:4, Interesting)
The claim that it's 100 to 500% faster is probably accurate in some sense, but compared to what? An old version of Netscape or Explorer? And on what kind of set-up? You can probably see that kind of variation in a single browser installation just by changing to different options and settings or by closing other windows or background applications. Personally, I often find myself switching between browsers depending on what seems to be working better on a particular day or on a particular network or machine.
On the other hand, he does sound like he's a bright kid with a good future, but probably one that just took Mozilla and turned it into snazzy looking bloatware with a bunch of extra features. Or, perhaps an even brighter kid who did the same thing from "scratch" with a lot of cutting and pasting (of his own work and from existing programs and libraries) to end up "writing" so many lines of code.
Accurate and not... (Score:2)
1) Use COM to incorporate every Active Document Web Browser there is
2) Use IE as a basis for the rendering
3) Use those annoying little characters that MS calls agents
4) Develop a compression utility that works on the server as a proxy.
My guess is that his compression is partially lossless, meaning some data gets lost. I am guessing that is why when he has 7x compression the system crashes. Below that the system "ignores" the lost data.
So what I think is unique with this browser is that it is an all in one solution that probably is pretty user friendly. And remember what amazes people is not the tech, but the presentation of the tech....
Re:Accurate and not... (Score:3, Interesting)
You are right on the presentation bit
Re:Accurate and not... (Score:3, Interesting)
Lets say that you want to increase compression of some data. EG HTML. Could there not be a technique to speed things up? Sure there is, get rid of the spaces, remove some tags, etc.
Well lets say that with each compression technique there are levels of what can be thrown away. And maybe when he tweaks to level 7 he throws away too much. At that point the app does crash since he may be throwing away something interesting.
That was my point of partially lossless....
Extraordinary claims require extraordinary proof (Score:5, Insightful)
That's massive work _and_ a revolutionary breakthrough. If he's that good - and in a way that others hadn't thought of despite the efforts of several of the world's largest companies going into browser and network research - then this is remarkable. But without hard evidence (or even a mention on the competition's admittedly poor website) this just sounds way too much like a scam.
Giant VB Applicaiton? (Score:4, Insightful)
Prediciton: It turns out to be some Visual Basic application which uses built-in windows components such as media player... thus allowing "All media formats, and DVD playing capabilities"
Quadrupling "Surfing Speed" is so bizzare a claim that I have no idea what it could mean. Maybe he's blocking banner ads... at 56k it could make a difference.
As for the "lines of code" I strongly doubt that a kid is using the same criteria for lines of code that everyone else is using... it probably includes his html test suite, and all his test code, abandoned code and documentation added together. Or maybe he didn't know how to write a function, so it is a big cut-and-paste one-function VB program with Goto's.
It's not that I doubt that a kid can pull this sort of thing off, it is that I doubt the school teachers nor the media have enough knowledge to judge it or report it accurately.
Re:Giant VB Applicaiton? (Score:3, Informative)
That would be the Microsoft Agent [microsoft.com].
Agent is simple to use. There are several dozen "agents" you can easily download that are ready to use, or you can make your own fairly easily. Here is a module I created to use the Agent in Visual Basic almost 4 years ago. Notice how easy it is to animate the Agent, and make it interactive. Once the character is loaded, you can make it do almost anything with a single line of code.
The code for the agent module can be found here [chrisnaimee.com].
Strong sense of deja vu (Score:3, Interesting)
Why am I thinking this is just another one of those snake-oil web speedups that does lots of caching and pre-emptive downloading of pages on the off chance you are going to view it? I'll be taking this story with a large pinch of salt for now I think.
Re:Strong sense of deja vu (Score:5, Informative)
Actually, something did come of that (Score:3, Informative)
The Cayley-Purser algorithm [wolfram.com] she developed was subsequently shown to have security flaws; I don't recall if this was before or after the EU prize, but thats immaterial, the work was original and interesting, and worth a prize for a 16 year old!
She has subsequently written a book [amazon.com] , which is a pop science introduction to crypto, and I understand from the blurb she's now studying maths at Cambridge.
-Baz
Re:Strong sense of deja vu (Score:5, Informative)
Bullshit. Get your facts straight before you malign someone. Sarah Flannery
She used Mathematica, so the Wolfram website has review [wolfram.com] of the book.
Here's a quote from Bruce Schneier in his 15 Dec 99 newsletter [counterpane.com].
All of this was easily found with a Google search [google.com] that garned 24,000 hits.
Is that so? (Score:4, Interesting)
Sure, you can leave stuff out (images, JavaScript, Flash), but "at least quadruple"? If the page is simple enough then you can't just ditch a chunk of it.
Ooh, AND "[at] least quadruple surfing speeds" and "they found it boosted surfing speeds by between 100 and 500". Even the article isn't making any sense . .
Of course, if this turns out to be true than I will be the first to eat my cat (and the first to download it), but I'm sure this isn't even possible, right?
Just my 2 cents (actually, that was more like 5) . . .
Lets not get too excited (Score:2, Interesting)
What a load of crap (Score:5, Interesting)
A kid coding 780'000 lines of code in 18 months. All alone. In that time he have had to design and implement the whole shit including "every single media player built in".
It would require some sort of dial-up-server side module to compress and modify the contents of the data and this kind of system would most certainly be a lossy method for transferring data. It won't be possible to transfer binary data with this thing without corrupting the result completely.
And what kind of a piece of software would choke under the load of 7x56k modem ("At seven times it actually crashes so I have limited it to six.")?
This is just a cheap attempt to gather some attention.
Re:What a load of crap (Score:3, Informative)
As for his claims, well, I wasn't at the show this year, so I haven't seen his entry, unfortunately. They do sound fairly unbelievable, but you have to remember that they're being filtered through journalists, most of whom are really fairly tech-ignorant.
I can say though that the Young Scientist is a major and well respected competition. The quality of the winners varies a lot from year to year, as you'd expect, but it's not run by idiots likely to be taken in by a hoax. Two yeras ago they flew in a Maths professor from MIT to verify some claim, so don't just accept things blindly.
Of course, none of this prevents this guy from having stolen chunks of Mozilla or something, and then bolting some bits on.
Ok, let's think this through.... (Score:5, Interesting)
If it does require a server side piece, it's not a web browser, per se; but as a general question, is it worthwhile to look into "compressed" web pages, e.g., foo.html.zlib? (I tend to doubt the savings are that much for the "average" page, but shoving graphics into an archive might keep down the number of requests needed to fetch a whole page and its graphics.)
If it's not server side compression, the only thing I can think of (and fortunately smarter people than me will think of other things I'm sure) is that he's pre-fetching and caching pages to make the apparent speed faster.
So is the "secret" that he has some hueristic that sensibly guesses what links you'll click next, combined with regularly fetching, oh say, your most requested bookmarks? (In my case it might look like: slashdot -- New York Times -- slashdot -- sourceforge -- slashdot -- freshmeat -- eurekareport -- slashdot.)
In other words, is he mirroring sites locally in the background? And if so, how must bandwidth is wasted just sitting in the cache until it's stale?
(On the other hand, could I point his browser at
Re:Ok, let's think this through.... (Score:5, Informative)
If it does require a server side piece, it's not a web browser, per se; but as a general question, is it worthwhile to look into "compressed" web pages, e.g., foo.html.zlib?
This already exists, look for example at mod_gzip for Apache. This will compress pages before transmitting if the browser claims to support it. Mozilla does, I believe IE does too.
Re:Ok, let's think this through.... (Score:2)
At any rate it's a technology that basically anyone can use since all teh big browsers and servers support it.
Re:Ok, let's think this through.... (Score:2)
I could disclaim that I work for the same company, but I have squat to do with dial up services, so it's kinda pointless. I just know about it because I walk through tech support every day to my office.
--
Evan
Re:Ok, let's think this through.... (Score:3, Informative)
Sure is. So much so, that its already been done. Mozilla, for example sends a HTTP header Accept-encoding: gzip, deflate, compress;q=0.9. If the server understands that (e.g. Apache with mod_gzip), it's free to compress the data on the wire. IE (as of 5.5 anyway, don't know about 6.0) doesn't appear to send any "Accept-encoding" headers. I'd very surprised though, if this led to anything like a 400% speedup in anything but highly controlled test conditions.
I'd hazard a guess that this new browser is quietly doing some background-caching. What articles I could find about this, however, are short on detail and kinda long on BS (web browsing and watching DVD's at the same time is a revolutionary feature? riiight), so it's really difficult to tell what substance there is behind all this. Time will tell, though...
Re:Ok, let's think this through.... (Score:2)
Anyway, gzipping content can easily make for an 8x size reduction on a large page, especially if there's a lot of repetition in there, e.g. lots of tables. Whether this translates to a significant speedup in browsing speed depends heavily on the size of the page and the speed of the connection; certainly on a modem, going from 80k to 10k is very noticable
Apparently is was actually 1.5 Million lines... (Score:5, Informative)
There is no way.
Um...yeah... (Score:2)
This is definitely one of those "I'll believe it when I see it" articles.
no footprint ? (Score:4, Informative)
I however found out this thread in the news [google.com] but, mind you, it's based on the same story...
They bet that if it's possible, he may have either implemented some quick prefetch and/or pre-formatting subroutine...
Writing the whole thing? (Score:2, Insightful)
Maybe he found some compiler options that quadrupled the rendering speed of <somebrowser/>.
Maybe he is just a fraud, and could sneak into the competition after creating a nice looking theme for <somebrowser/>.
Maybe I'm just guessing and typing whatever comes to mind in <somebrowser/>.
Hardware requirements... (Score:2, Funny)
I'm sure that that's in there somewhere, oh, yeah, look...there it is commented out above line 53,425 in the code. Yep.
Pattern matching? (Score:5, Insightful)
Most web pages have a lot of static content in, especially menus etc. You could start rendering the page immediately from the cache from the last page and rerender afterwards as the new page starts to differ from the cached version.
As the page comes in, keep switching to the page that is closest to same structure in cache (ie predominantly on the HTML tags). Don't render the text until the initial few chars are confirmed by the version downloading, then progressively render that (ie show old version then modify words where they differ).
This would have the effect of progressively rendering the page as a whole much like those progressive GIFs. It would show a large speedup on pages that contain tables, as most browsers these days won't render a table until it has recieved the
This would be a 'faster' browser with no compression or pre-caching.
Phillip.
Re:Pattern matching? (Score:4, Insightful)
I think 90% of page traffic occurs on the top few websites through regular visitors, so in most cases the client will already have some data available.
So, Jim, I'm concerned with your performance... (Score:5, Funny)
Jim: Yes, I --
PHB: Jim...I'm concerned about your performance.
Jim: Er, wha--
PHB: You write, what, 30 maybe 80 "eL Oh Cee" a day? Right?
Jim: Well, the TPS and project plans take --
PHB: Says here, that this 16 year old kid can write 1500 "eL Oh Cee" a day. What do you think about that?
PHB: Don't laugh...this is serious.
Jim: Sorry. I ment --
PHB: Jim, maybe you need to put in more hours. Reconsider your work habbits.
Jim: I work till 10 most nights...
PHB: Jim, it's not the hours it's how efficiently you handle them. I expect todays TPS on my desk by noon, along with a status report on each programming task you've done today.
Jim: It's 11 --
PHB: That's it Jim! Keep up the good work. In the meantime, see if you can increase that "eL Oh Cee" to, say, about a hundread. It's good to make a good impression. Fine. Excellent. I knew I could count on you. I'll see you then! ... Brian...did you fill out that TPS report...
Science project? (Score:2, Insightful)
Hell, even reading the hypothesis of his project would be an improvement over what we have -- nothing.
What shitty news coverage. The media isn't skeptical enough when it comes to science. If this was some miracle dreamed up by a politician, the media would have torn him to shreds by now, digging up dirt on him, his family, his marital history... everything.
But when a miracle science story comes around, the media swallows it hook, line, and sinker. Unacceptable for this day and age.
Next year he'll win another big prize (Score:2)
My guesstimate is... (Score:2, Funny)
Adnan says a six-fold increase is about the maximum practical boost.
"At seven times it actually crashes so I have limited it to six."
Now that's a good debugging technique, no wonder the code has 780,000 lines!
To make the software more user friendly, it features a talking animated figure called Phoebe.
With these skills, I guess he'll be working in Redmond soon...
Very skeptical (Score:2)
No Wil ! (Score:2)
Stop playing with time-shifting, and go out and play with the kids...
Increases "surfing speeds" (Score:3, Interesting)
Irish Patent Office does not know about this (Score:5, Interesting)
Query
Application Date: 08/01/2003 -> 10/01/2003
Abstract: *internet*
Results: 0
Query
Date Of Grant: 08/01/2003 -> 10/01/2003
Abstract: *internet*
Results: One Result: 2000/0717 82661 Server-based electronic wallet system
Thats it, so it doesn't seem he applied for the patent in Ireland then...
P.S. The stars around "internet" are mine, I used them to indicate that I searched all abstracts that contained the word "internet"
Why only a two day window? (Score:3)
Duh (Score:2, Funny)
Then why the hell did you post it? (Score:5, Insightful)
Twits who make up bullshit stories like this thrive on attention. By posting it on a major site like slashdot, you give him exactly what he wants. Just use a little restraint, and try not to post the stories that are obvioulsy fake -- like this one, and the one about Masters of Orion 3 beign out soon (grin).
What methods are possible? Hard disk are cheap? (Score:2, Interesting)
Serverside caching [propel.com] could be used.
TCP/IP non-comformaty [slashdot.org] is the third option.
Assuming this is true, (ignoring the 1500 lines a day), what else could he be doing?
Judging by harddisk prices, client side cacheing algorythms would make sense. Cacheing many portal and search engine homepages is a powerful start. Combined with a central server that then reviews these popular pages for changes, and publishes a simple summary for the browser client to collect and compare with older summaries, then a browser can collect only updated portal pages for the cache, all optimizes portal renders.
Then less common homepages, such as the high school I attended, can be gleened from users typed-in webaddress history, and automatically cached as chron-job.
Creating cached copies of commonly used graphics on portal website can save a ton of bandwidth. Again a server based bot could rate the linkcount of graphics on portal sites, and if the graphic has changed, and then post this list for browsers to collect for caching. Searching HTML for imagefiles, that are already stored in the cache, and modify the page on the fly to call only the cached image would save bandwidth. e.g. caching all of slashdot's article catagory icons.
Then the tricky part, "which linked pages to cache while the user reads a page?", so that when a link is clicked, the pages renders fast. I would download the html from all of them, and while the reader reads, check for already cached images, and then start downloading image files.
-Mac Refugee, Paper MCSE, Linux Wanna be!
and first poster of the word "knoppix"
1500 lines a day (Score:5, Funny)
Then he discovered loops.
Possible explaination for LoC (Score:5, Informative)
I never really knew the true line count. I just remember the Borland one because I used to often do a global compile any time I wanted a half hour break ("Oh, the systems acting funny. Better do a global compile to make sure it's not a dependancy problem." If my boss came by and I wasn't there, he'd see the compile running on the screen).
ESAT Young Scientist Competition (Score:3, Informative)
http://www.esatys.com/ [esatys.com]
Is it possible he counted 780,000 loc because he was including libraries and component code etc. etc. The article is badly written and doesn't give a true representation of his work. He claimed on Irish TV that he had written a client-server pair. I'm still fairly suspicious myself, but it *is* possible.
some newsgroup discussion about it all (Score:3, Informative)
No data for patent (Score:3, Informative)
Yeah, right (Score:3, Interesting)
I call bullshit. That claim dosn't make any sense whatsoever, especialy if it's just software.
It seems (to me) Like he just threw together a bunch of MS APIs (such as the microsoft speach API for 'Phoebe', the windows media API for the DVD player and video players, probably even used IE to display pages).
At most he threw in an intelegent caching routine, such as pre-downloading linked pages or something. I also don't think he wrote 780kloc
FTL Browsing (Score:4, Funny)
Explains why it crashes at Warp 7 too, the dilithium code just can't take, keptin!
Some verification was done! (Score:3)
Sorry Folks... looks for real (Score:3, Insightful)
The Irish browser story: Ok folks, here's the scoop. I am just back from talking to one of MIT Media Lab Europe's researchers, who both checked out the browser and talked to Adnan. He says the browser is 'absolutely extraordinary'. He says that what Adnan has done is re-engineer the efficiency of how a browser operates, which allows it to run up to six times faster (but usually not that much faster -- two to four times faster is more common). So it's not managing bandwidth but managing the way the browser itself handles and presents information. The researcher (whom I know and will vouch for) says that instead of simply tinkering with existing code he went down to the socket layer and reworked it at the protocol level (now, many of you guys will know the significance of this better than me, I'm just reporting the conversation). He added that it is incredibly clever work and stunning that a 16 year old has done this (I am not scrimping on the superlatives because that is what was said). (NB: A conversation in a group ensued that this work perhaps suggests that because the browser market is a virtual monopoly, there's been little incentive to improve efficiency in this way -- indeed, it might be beneficial to product development to just eke out a leeeetle more efficiency now and then and advertise it as continuing innovation... but I leave that to further discussion among the well-informed).
And Adnan has indeed worked in all the existing media players AND a DVD player so you can watch a DVD while surfing. And incorporated in a voice agent that will speak web pages, for young children or for the sight-impaired. The improved efficiency angle got the notice of the few media reports done on this so far, but it's really not what Adnan himself was emphasising -- it's the whole package, said the MIT guy.
Not surprisingly Adnan now has more than one university interested in him. And he has apparently told the numerous companies who saw the browser in action and who wanted to commercialise it that, at least for now, he has no interest in commercialising it.
I will note that the MIT researcher had a big grin on his face and it was clear he found the whole project a pleasure to talk about. He also said he'd heard about the browser before he arrived at the Young Scientist exhibition and made a beeline to see it. Adnan apparently didn't really think it would necessarily win an award --the researcher told me it was clear that it HAD to win. So there you go. I'm sure we'll hear a lot more about all this soon.
And yes, he has copyrighted it.
Read More... [weblogs.com]
More technical details here (Score:3, Informative)
Some snippets:
"He says that what Adnan has done is re-engineer the efficiency of how a browser operates, which allows it to run up to six times faster (but usually not that much faster -- two to four times faster is more common). So it's not managing bandwidth but managing the way the browser itself handles and presents information. The researcher (whom I know and will vouch for) says that instead of simply tinkering with existing code he went down to the socket layer and reworked it at the protocol level (now, many of you guys will know the significance of this better than me, I'm just reporting the conversation). He added that it is incredibly clever work and stunning that a 16 year old has done this (I am not scrimping on the superlatives because that is what was said)."
So perhaps there is some truth in this after all.
newsQuakes [www.skep.tk]
Re:Great, yet another browser... (Score:5, Informative)
it is doubtful that he will - according to the article he has applied for a patent on it.
Re:Great, yet another browser... (Score:3, Insightful)
Re:Not necessary (Score:2, Offtopic)
2:00 am to 5:00 am?? That's when I do most of my browsing!!!!
Of course, not at Kuro5hin.org...
I uh, do, uh, anatomy research... and uh, changes in primate mating habits caused by the ubiquity of digital cameras.
No. Make it GMT (Score:2)
What's that you say? that's 6pm - 9pm EST? Shame.
Re:Different Calendar (Score:5, Funny)
Well, okay, they don't but it'd be nice if they did... instead of the year round crap.
Re:Not 6x -- six FOLD! That's 64x56k (Score:2, Informative)
Re:Paper clip feature too! (Score:2, Funny)
Re:It slices it dices (Score:5, Interesting)
The article discusses that he uses a simple modem, so that could perhaps mean that he just wrote some new method for transmitting more bits/second by a more accurate signal detection method.
Although, this has basically nothing to do with the browser at all, so it does not make any sense to me. Sounds like the article mixes apples and oranges, or perhaps the "student" is just laying out a smoke barrier so that noone will steal his ideas before he get the patent.
my 5 cents...