Your Tech Skills Have a Two Year Half-Life 289
itwbennett writes "Eric Bloom, an IT leadership coach and former CIO, has answered that eternal question 'does working on old software hurt your professional marketability' with a somewhat surprising 'no.' But, Bloom adds, 'a techie's skill set from a marketability perspective has a two year half-life. That is to say, that the exact set of skills you have today will only be half as marketable two years from now.'"
Depends... (Score:4, Insightful)
Depends really on how specific your skills are.
Knowing, for example Java or .NET programming languages won't decline in value that fast. Perhaps specialising in certain specific products will- and certainly the development environment will.
On a non-programming side- knowing the basics of computer hardware doesn't decline in value that fast. Perhaps specialising in certain models does.
Re: (Score:3)
I was going to make a similar but converse point... as a tech generalist, much of what I do is bleeding-edge. Old knowledge is as irrelevant to me as it would be to a potential employer.
Just as doctors are supposed to keep up to date on their skills through continuing education, technologists are expected to keep fresh on new tech trends.
Re:Depends... (Score:5, Interesting)
Re:Depends... (Score:4, Insightful)
I'd be wary about that "old knowledge". It may prove useful. There's LOTS of legacy software out there. I stay familiar with Fortran because it's still bloody good for numeric computations and it's uneconomic to translate old Fortran codes, which means I'm going to encounter it. I spent time learning about Intels iWARP chip (brilliant design, naff implementation) and Content Addressable Memory because these are ideas that have appeared multiple times and will therefore appear again. Understanding the principles now saves me time and effort for when they become important later on.
That's not to say I stay from the bleeding edge. I try to split my time 50:50 between the past that I may well encounter in the future (a trait that secured me my current job) and the future that I will certainly encounter in the future (a trait that secured me my jobs at NASA and Lightfleet). Both will come up, that is inevitable, but it's not possible to know in advance which one will come up first or in what way.
Generalizing is best done by making the fewest assumptions about the past, present and future that you can that will leave you enough time to learn the skills well.*
*This is important. 100 half-baked skills are of equal value to 100 highly-tuned future-only skills that turned out to be a dead-end. None whatsoever. Mastering a smaller set of transferable skills, legacy skills and future skills, thus being totally generalized, is the obvious ideal.
Re: (Score:2)
I don't think the point is that it won't be useful. I think, by definition, the article is simply saying that it will be half as useful.
It stands to reason: the older a technology is, the more people in the labor market know it (including the global labor market: India, etc.)
To be honest, I think the life of a permanent contract is an unstable one. I prefer the Japanese model (now also fading) of in-house training to fill strategic goals coupled with bi-lateral loyalty to the company. It's a much more susta
Re:Depends... (Score:5, Informative)
Software, OTOH, frequently changes drastically and constantly; it's engineered by man, and can be radically altered in any number of ways on whim, forcing a reinventing of the wheel sometimes even; a moving, morphing target, much of it probably driven as much as by planned obsolescence and profit as it is utter necessity. (Does Word really need to keep "evolving" to do what it does?) Sometimes I really wanna say "screw all this" and go start a goat farm.
Re: (Score:2)
Depends on what kind of doctor you are... recent advances in surgical techniques (laparoscopy, robotic surgery, etc) means surgeons have to learn new techniques all the time - and ones that require a lot more precision and practice than the latest HTML standards or OS APIs...
Re: (Score:2)
"as a tech generalist, much of what I do is bleeding-edge. Old knowledge is as irrelevant to me as it would be to a potential employer."
Then I'd bet you are not so much a "tech generalist" as you think. What I learnt about 20 years ago is basically as valuable now as it was then, and I've being building what I know now upon that from then on.
Maybe the fact that I almost don't do Microsoft can help to explain it. I certainly don't value so much what I learnt about DOS 5.2 or Windows 3.11 but what I learnt
Re:What I learnt about ____ years ago (Score:2)
You're a ways ahead of me but what I learned between 2004-2010 for Windows "help-desk" stuff is still good for some other 3-4 years. I purposely stayed away from the harder volatile server side stuff, because I like Durable Knowledge. So I'm backup Helpdesk and a "line" accounting administrator.
Re: (Score:2)
Old Knowledge is NOT irrelevant. It is useful for abstraction of concepts, which are longer lasting principles that span times. I have known people that work bleeding edge stuff, but don't understand concepts, and they scare the crap out of me, because they have no real understanding of what they are doing. They do, because the manual says to.
Re: (Score:2)
I think the point is that if you are an admin of a specific release of a product from a vendor, the further behind on the upgrade path you are the less useful your skills are.
Re: (Score:2, Interesting)
If you only admin one product, from one vendor you are a glorified user.
Re:Depends... (Score:4, Interesting)
Not sure about glorified. Users get to scream when the feature sets change. Admins can't. Users often get to practice in other environments, it's much harder for admins to. Users get to blame admins when things fall over. Admins get to.... ....well, turn into a paranoid, schizophrenic wreck of a human being.
Re: (Score:2)
Knowing Java ins't good enough anymore. For instance, a developer who just does AWT or SWING is going to limited use for potential employers. You have to keep up to date on the common frameworks. What's SpringSource, Hibernate, Apache, etc. up to lately in the Java Space? What about other languages that execute in the JVM (i.e. JRuby, Clojure).
Re: (Score:2)
I've never seen a job with JRuby, Clojure or Groovy requirement. (But certainly not in my country, Hungary.)
Re: (Score:2)
The strength of any language is not in it's syntax, but in the libraries and frameworks.
Sadly, must compilers do not really do 'DWIM', even if you got it 'conceptually right'.
CC.
Re:Depends... (Score:5, Interesting)
It also depends on when you're talking about. After the Dot-Com crash, Java programmers were hurt FAR, FAR worse than C or Fortran programmers. Shortly before Y2K, Fortran and Cobol programmers were in massive demand. (For those who argue Y2K was a hoax because nothing happened, I'd point out that after a large fortune and a larger army of coders went to work on fixing the bugs, you should have EXPECTED nothing to happen. Fixing problems after the disaster is too late.)
So the decay curve isn't a simple one. It has bounces and bottomless pits along the way.
However, and I can't stress this enough, staying current isn't merely a matter of learning the next feature of the old language set. To stay relevant, you MUST diversify. A coder should also be a damn good system admin and be capable of database admin duties as well. Being able to do tech writing as well won't hurt. You don't know what's going to be in demand tomorrow, you only know what was in demand when you last applied for work.
Programmers and systems admins shouldn't specialize on one OS either. As OS/2 demonstrated, the biggest thing out there in week 1 can be a forgotten memory by week 12. The market is slow at some times, fickle at others. You don't know how it'll be, the best thing you can do is hedge your bets. If you've covered (and stay current on) Linux, a BSD Unix variant, a SysV Unix variant, Windows Server, and at least one RTOS (doesn't matter which), you'll know 98% of everything you'll need. You can learn the specific lingo needed by a specific OS implementation quickly because that's only a 2% filler and not a 100% learn from scratch.
Although workplaces don't do sabbaticals (which is stupid), you should still plan on spending the equivalent of 1 study year for every 7 work years. (If you spend 1 hour a day practicing, relearning, or expanding your skills excluding any workside stuff, you're well in excess of what is required. I can't guarantee that an hour a day will make you invulnerable to downturns, but I can guarantee that there will never be a time, even in the worst recession, that your skills aren't in demand.)
Re:Depends... (Score:5, Interesting)
Half-life is still a pretty damn good analogy.
It is one of those mysterious, non-intuitive things about the subatomic world. You see, rather than ageing uniformly, atoms randomly decide whether to decay or not. Meaning that if you have a container filled with plutonium, after 24,100 years half of the atoms would have decayed, the supply in the container has decayed as a whole, but in reality half of the atoms there never decayed at all.
The result of the analogy is that every two years half of the programmers will be unmarketable (unless they acquire new skills) the other half however doesn't need to learn anything since their exact skill set is still in demand.
Re: (Score:3)
I've admined 386BSD, FreeBSD, NetBSD, OpenBSD, Linux (0.1 - 3.1), SunOS, Solaris, IRIX, HPUX, OSF/1, VxWorks, Windows 200x Server, Windows XP, Windows 7, Plan 9, Inferno, the PDP-11 and OS/X. Not sure I'd call RTAI or Xenomai distinct OS'. Montavista certainly wasn't. Dealt with all three.
This includes direct kernel work (hacking patches together to form the Functionally Overloaded Linux Kernel was damn hard, thanks to massive conflicts), writing drivers for a number of these, in addition to the usual insta
Re: (Score:2)
Actually, it will. Not that .NET or Java are going away within two years, but they'll evolve and develop. What you know today about them is only worth half as much in 2 years when new libraries are out and the next version of .NET requires you yet again to relearn half of what you know.
Re: (Score:2)
As I sit here working on a Java web app written almost 10 years ago, updated continuously since, but still with threads of old libraries and methodologies within, I think the half-life is a little bit of a weak comparison to make.
TFA is all about staying on top of your unnamed vendor's magic moving technology. It isn't about technology skills in general. Heck, it isn't mentioning anything specific either. Well, there is that one line about staying up with PL/SQL.
Re: (Score:3, Informative)
.Net is not a programming language. Two of the most popular languages used with .Net are C# and VB.Net. C# is new with .Net and is still around. VB.Net is an evolution of Visual Basic, which has existed prior to Windows 3.0. (I don't know exactly how old it is, but VB 1.0 was in text mode, and created apps for the then-current version of Windows.)
What about languages? (Score:2)
Will that be less valuable after having 2 years experience in the field?
Re: (Score:3)
I don't think the theory applies universally to all tech skills. C has endured well over the years. So has SQL. Other languages, not so much. I don't see many ads for Ada or Lisp these days. Your actual mileage may vary.
Re: (Score:2)
But how marketable is SQL ? Most of the people already know SQL. Lets say you apply for a programming job at a web-development company and they are all using fancy "noSQL" databases. The question is if you know all the new stuff and when to use it and when not to use it.
Re: (Score:2)
Again, which dialect? T-SQL isn't identical to PL-SQL. You are correct about NoSQL (though that's often just a fancy way to describe a subclass of key/value databases, which also includes the BerkeleyDB family, XML databases and a myriad of other styles). However, it's not limited to that. Hierarchical databases exist, as do "star" databases (data warehouses), object-oriented databases (sometimes considered a branch of NoSQL), indexed sequential databases are still a popular format,
SQL is also not a static
Re: (Score:3)
How marketable is SQL? There are two ways to look at it. Will SQL help you distinguish yourself from others and leave them in the dust? No. But just try to get hired without it.
There are lots of legacy databases out there, and you won't be talking to them without a fair understanding of SQL. Even the niftiest of whiz-bang query tools will generate flawed SQL every so often, leaving you on your own to figure it out.
IT is more than coders... (Score:5, Interesting)
Obviously, if you think of IT as just programers, what he is saying makes no reals sense, since staples like C, Java, and
Re: (Score:2)
If you have 'skills' using Office version X, it will probably not be as valuable in two years when a new and improved product, Office X+1 comes out.
That makes the big assumption that few of your Office X skills translate well to Office X+1. When it comes to the upgrade from Office 2007 to Office 2010, you might have a point (stupid ribbon...), but in general that's not the case. Someone who could operate Office 95 can get around Office 2007 just fine. Someone who learned pre-ANSI C, C++98, or Java 1.1 can deal with 99% of code written in C1x, C++11, and Java 7, with a short learning curve for anything significant that has changed. It's the same pat
Re: (Score:2)
It's not just the "barebone language". It's the various libraries and other tidbits that are considered "essential" today because the allow rapid development. The C standard didn't change in ages. Still, if our programmers didn't know their way around the various libraries we have collected in the past years (and we're still collecting, adding to, replacing and eliminating) they'd be worth less than half of what they are.
Various other things also apply. Security is one aspect that becomes more and more impo
Re: (Score:2)
I don't see many ads for Ada or Lisp these days
Look at Rolls Royce. They're hiring SPARK Ada programmers like crazy, as are a lot of other aerospace companies. It doesn't really have any competitors for systems where failure is not an option. As for Lisp, the last job offer I got to write Lisp was for an investment bank.
Re:What about languages? (Score:5, Insightful)
Suppose I know some amount (X) of C now (Just out of college)
Will that be less valuable after having 2 years experience in the field?
No, it wont. He's talking about *certain* IT skills. I'm going to go out on a limb and bet he's referring to the kind of tools you learn in a simple ITT-Tech type certification program.
Re: (Score:2)
Last time ITT Tech was covered by any forum I read, it was soundly ridiculed as not teaching any skills worth knowing, and the certification was denounced as being utterly worthless and accepted by nobody. ITT Tech probably taught Good Governance on Numenor.
Re:What about languages? (Score:5, Insightful)
One simple way to avoid that type of shit from entering the workplace, refuse to hire anyone that has a technical certificate of any kind along with those with a degree from the diploma mills such as ITT Tech and Conservative err Community Colleges. Community Colleges are a fancy way of saying trade school. Employers should only hire the truly educated and those are only from the major Universities.
I work in higher education, and not for a for-profit or a community college. Your belief that graduates from "the major Universities" are somehow better than those from other institutions, especially for something like application development, is hilarious to me.
Re: (Score:2)
"Employers should only hire the truly educated and those are only from the major Universities"
What a pompous ass thing to say. If/when you get out of mom's basement and into the real world, you may see that many jobs/careers are perfectly suited for trade school graduates. Like maybe, umm, trades? Technicians? It doesn't take a rocket scientist .. etc.
4 year CS does not fit for IT jobs apprenticeships (Score:3)
It jobs need apprenticeships not 4-5 years in a class room. IT is a trade and CS is the high level stuff.
Re: (Score:2)
Techs have to self educate themselves, unless they want to spend 4 years in university every 2 years of work. And if they can self educate, they don't need the training at uni...
Unless the business plan is to use em up, burn em out, send in another replaceable cog every two years...
Re: (Score:2)
Suppose I know some amount (X) of C now (Just out of college) Will that be less valuable after having 2 years experience in the field?
If you haven't learned anything new in your first two years as a professional c programmer, you might want to try another discipline.
If I haven't learned anything new in any particular two year period, I get bored. Best option there is to either shake something up with my current venue, or quit.
Re: (Score:2)
Suppose I know some amount (X) of C now (Just out of college)
Will that be less valuable after having 2 years experience in the field?
School related C skills without work experience... no... it won't be worth less in 2 years. It will be worth exactly the same... which is diddly.
Only the 2 years field experience will mean anything when you apply for another job. And that field experience will decline... usually because a great deal of knowledge around programming is not about knowing the language but the framework, modules and libraries your project uses. And those are continually changing.
Re: (Score:2)
With what libraries and languages what you worked in C? Won't those change? If you're a games person, are you up on DX9? DX10? 11? Database backends? SQL? NOSQL? Have your version control skills expanded to match existing systems? Still using CVS? SVN? Git? "The Cloud" ... have any of your applications been designed with that kind of focus in mind of starting and stopping at any point and being part of a model with dynamically changing resource allocations?
Evolving skills are a demonstration of the ability
Re: (Score:2)
No, the 2 year claim is bullshit.
Re: (Score:2)
Not less valuable, but less marketeable, as the article says. Both are different things. Also, I don't belive it. But the working environment around here (that I've already jumped out of) may be unusual.
TFA sounds absurd, as it claims that markeability depends on the specific version of softwre you have experience. Like if somebody would hire a person that knows JSF 3.1* (it claims that small numbers aren't as important, but puts some importance on them) but not 3.2*.
Have you ever seen a CV that tells versi
Re: (Score:2)
Yes, I have seen versions numbers for platforms on CV's. I have them on my CV and I look for them in applicant CV's. I'm a Linux admin who manages two student interns (Jr. Admins). I do the screening of my interns as well as helping to screen full-time co-workers. When reading CV's, I give a higher weight to those with version numbers. I'm not too worried about minor numbers (i.e. RHEL5.4 vs. RHEL5). I'm not too worried about older versions. Version numbers act as a shibboleth to weed out the posers from th
Re: (Score:2)
"When interviewing a potential Linux admin, I always ask what version and flavour of Linux that they have experience with."
When interviewing a potential Linux admin, I always ask a TECHNICAL QUESTION I know specific for the version and flavour of Linux that they claim they have experience with if I even give a damn about deep specific knowledge (which most of the time I find basically irrelevant). That's what talks about their skill, not their ability to retain some version numbers out of a fast google sea
Re: (Score:2)
In some cases, yes. C coding style recommendations have changed over the years. Some C dialects have died (K&R, for example) and others have grown. The standards have shifted, so those who have learned C99 will be at a disadvantage to those who know C1x for newer code -- though the reverse will be true for middle-aged code. Ancient code could be in any of a thousand dialects.
The market for C is growing, but the number of shifts from C to C++, C# or Java (or other languages) is also growing and the value
Re: (Score:3)
Depends who interviews you really. If you're interviewed by someone who was a developer once- sure.
However, it's just as common to have someone with a non-programming background being the person involved in running IT departments. (especially in manufacturing- if you're working at a mid-sized company in manufacturing- almost all the IT managers came from sales or accounting and know very little about computers).
Quite frankly- IT is not a career to take if you ever want a promotion. Sure, it can happen- b
Re: (Score:2)
Experience is more valuable than education. No doubt about that. Problem is, though, unless you're very lucky you won't get the experience without the education. HR first and foremost looks at your CV. Nothing in your CV that suggests you know a thing about programming, no chance to get programming experience. There's one position to fill and hundreds of applicants. HR will simply toss out everyone who doesn't have a relevant degree without even hearing the applicant.
Huh? (Score:3)
As a general rule I don't even list things on my CV (resume) that I have less than two years experience in, these days...
I'm willing to accept this is the case for startups wanting the latest buzzword filled technology, but a LOT of places are happy at a much slower pace.
Re: (Score:3)
On my CV, I list things that I have less than 2 years experience, but I put skill level qualifiers like "Novice" ,"Intermediate", and "Expert"
Tell that to a COBOL programmer ... (Score:3)
Re: (Score:2)
It is highly field dependent. I traded the front line trenches of IT-Security for a comfortable management position not even a year ago, and I'm already struggling to stay current with the various threats coming our way. I simply don't have the time anymore to concentrate on it as much as I used to. I'd guess in a year, what I knew a year ago is not only obsolete but simply laughable.
Of course, COBOL won't change in the foreseeable time, or maybe ever. For most, the reality will be somewhere in between.
Re: (Score:2)
I'm still employed using skills I learnt in 1980.
How many skills are you using that you learned in the 1980s but didn't use at all in the 1990s? I bet it's a much lower number.
I call bullshit (Score:5, Interesting)
I still program in Java which I've been doing since 1998. I also sometimes program in Python which I've been doing since 1997. Obviously some things about those languages have changed, but many things haven't.
OO languages are fairly similar to what they were 10 years ago. As is OO design, etc. There have been large changes to frameworks etc, but there is a significant "core skill set" which transfers over.
In my case, my skills have not become become less marketable at all over the last two years. Recently I spent two years out of work (voluntarily), and when I returned to the job market I had no problem whatsoever finding a job.
I think the half-life of skills is more like 15 years.
Re: (Score:3)
If you take the time to read the article, you'll see he's actually talking about how long your skills in customizing a particular release of software are viable, not about how long languages or operating systems remain relevant.
As many companies stick with the same release of software for even longer, I question his numbers, but I don't question the theory. The lifespan of customizable products is much shorter than the tool-related skillsets required to do that customization. Your skills as a programme
Re: (Score:2)
I read the entire article before commenting. It says nothing of the sort. I don't even know where you got what you're saying. Did you read the article?
From the article: "The longer answer is that, in my opinion, a techie’s skill set from a marketability perspective has a two ye
Re: (Score:3)
My 1979 APL skills gave me a huge leg up on learning the R language in 2008, except for the tax of unlearning elegance, and the odd rust flake or two.
Are we talking skill cycles or fashion cycles on the two year tau?
Re: (Score:2)
I'm not disagreeing with this point. I said that frameworks etc have changed significantly, but core languages have not.
A 10-year half life might be reasonable, but the author was claiming a 2-year half-life.
That long? Optimistic, aren't we? (Score:4, Interesting)
That said, I've been coding QA software in some VB-Form language since 1994. My pay during that time has only increased. This is the first year that I've had to do anything in a C-form language.
The unfortunate fact of the matter is that a lot of new technologies are horse puckey. C++ was an actual improvement over C. The .net platform, for all its many faults, has actually increased my productivity, but much of the rest, Windows Presentation Foundation, Python, Ocaml, Ruby, Silverlight, et. al are nifty, but nobody *needs* them. Frankly, if the world standardized on Java tomorrow, and we just used extensions thereof for different platforms and purposes, we could all concentrate on getting useful work done and quit dicking around with learning the latest obscure and allegedly more elegant syntax. The best language and syntax isn't the most logically consistent one, it's the one you know. In productivity terms, human factors trump formal systems elegance every time.
Re: (Score:2)
I used a few basics as a teenager, as well as C, C++ & Delphi. Then I tried Perl, which I absolutely love. I've tried a little Ruby, it was okay. Currently learning some lisp, and going to have a look at Python soon. At work I mostly use Perl/HTML/JavaScript/SQL, with a little legacy maintenance of a Delphi app that we've thankfully just sold off the source to someone else, so I can use whatever the hell I want for future desktop-only apps.
If you're going to stick with your "one size fits all" mentality
Re: (Score:2)
I disagree. Some of my best productivity gains came from learning a new language, then never using it. Instead, I'd use all the good ideas in my "normal" programming language.
I became a better Perl programmer after I learned Ruby. I became a better programmer (in all languages) after learning Lisp, Prolog, and Erlang.
I last wrote a Lisp or Prolog program in the late 90's, but I use those techniques every day.
Re: (Score:2)
Some of what you say is true, but there is no reason to get all monotheistic about it.
Different languages have their uses.
Re: (Score:2)
Probably. I picked Java as an example since it's portable and object-oriented, unlike straight C.
sounds about right (Score:3, Interesting)
This certainly fits my experience. I'm "over 39" and have specific tech skills that date back to the early 80s. Those are worthless. I continued doing highly technical work and staying current into the late 90s, when I went back to school to build up some of my non-technical skills. Not such a good idea as it sounded. I emerged from school several years later with just enough still-marketable skills to land a tech job that offered little opportunity to further advance my skills, then got laid off from that, took a retail job as a life raft.... and now my "freshest" marketable tech skills are a dozen years old, and close to worthless. I guess it's time to get out the paintbrushes and see if I can swing a new career as an artist; at least the half-life on those skills isn't as short.
Re: (Score:2)
I have some specific tech skills from the early 80's as well... and some are not remotely worthless.
For instance, I learned C in 1982.
Or isn't knowing a specific programming language considered a specific tech skill?
Re: (Score:2)
Anything involving Linux kernel programming (like device drivers).
It depends (Score:3)
Re: (Score:3)
Interesting. I learned C, Unix and RDBMSes back in the early 80s. I only use C at home for hobby projects, but I still use Unix and SQL professionally. I learned Java back around the turn of the century and it's still paying my mortgage. Franky, I'm disappointed I can't seem to find any new positions that use any of the technologies I've learned lately (like OSGi, SOA or NoSQL databases). It's different if you're a front-end guy, I guess — I have seen some places looking for jQuery and HTML5 expe
Re: (Score:2)
"I took a retail job as a life raft."
I would love to be a life raft and get paid for it.
Consider the source - no wonder it's garbage! (Score:5, Insightful)
Even COBOL refuses to die. C, C++ and it's variants are still everywhere (Objective C for Apple's iPhone App Store) decades later. Java has outlasted the fads of Ruby and Rails. HTML has been around ... well ... since the Internet. Javascript continues to be the #1 web scripting language.
So no, your skills don't have a half-life of "X" number of years.
Re: (Score:2)
Well, I actually agree with your point that skill marketability does not degrade so quickly. Perhaps more importantly, I think that good employers recognize that it is far more important to be able to quickly learn new skills than it is to already possess them; being largely self-taught and having a fairly wide skill set has impressed employers more than any single point on my resume in my experience.
I do have to take some exception with one of your points though:
Re: (Score:2)
"IT skills" is an oxymoron. Generic software engineering skills never lose value.
Re: (Score:2)
HTML has been around ... well ... since the Internet
Ignoring the fact that you are about two or three decades off with that, have you looked at HTML 1.0 recently? Even HTML 2.0 from 1994 didn't have CSS or any semantic markup. HTML 4 is still pretty relevant, but that's only from 1997 and web development for the past few years has really required JavaScript (which has changed significantly over the years by the way) and asynchronous HTML requests - which weren't mentioned at all in HTML 4.
In short, your post make me think that you've never done any deve
Re: (Score:3)
HTML 4 as written in 2011 is vastly different from HTML 4 written in 1997 (for instance, we tend not to write our sites for a specific version of a specific browser anymore). CSS as written in 2011 is vastly different from CSS written in 1996; for instance, before we had vendor prefixes we had to use hacks to present different CSS depending on browser and sometimes browser version.
Gross oversimplification (Score:2)
Depends on the field (Score:2)
On the other hand, I don't think this is true for all "techies." The tools for electrical design, for instance, haven't changed much since the introduction of 2D CAD tools for PCB layout in the 1980s. If you've been soldering, prototyping, debugging, and laying circuits out for the last 20 years, chances are pretty go
Technic / marketing (Score:3, Insightful)
"a techie's skill set from a marketability perspective has a two year half-life"
Well, a marketie's skill set from a technical perspective has a zero year half-life.
Programming only, I suppose (Score:2)
Bullshit (Score:5, Interesting)
Do I not really understand him... (Score:2)
...or is it most of the poster here, who bash him?
Did he say that C/Java/whatever decays, or only the worth of people who use this skills?
I can imagine an interpretation of his statement, which would make sense. In my youth I coded in BASIC, Forth, PASCAL... This was somewhen in the middle ages. I was ok for that time, but today those skills are decayed to nothingness. I made some money in JAVA projects. Only a few years ago. Since then I didn't use JAVA at all. I image it would be much more difficult for m
This is one persons opinion. (Score:2)
With exactly zero evidence to back it up. The faster we ignore this entire story, the better.
Look at windows xp that lasted 5 years and win 7 (Score:2)
Windows 7 may go just as long.
Some industrial systems still have windows 9.X, ISA cards and other older stuff.
2 years is to quick and lot's of places may do long testing time of new OS's / software any ways before roll outs.
Re: (Score:2)
speak for yourself nodejs boy (Score:2)
Totally flawed analogy. The figure might hold true for latest fashion in development technology, but its insane to think that fortran skills for instance will be half as marketable 2.5 years from now. They will probably have declined by a few percent but difference in value of 40 year old tech versus 41 year old tech is negligable. Its more like the value of a technology falls by 100/(2+years-since-hot) percent every year.
If all you have is coding skills ... (Score:2)
Read the "6 ways": this guy is incompetent (Score:2)
Every single statement referenced the "software vendor". Every software vendor's goal is to lock you in to not thinking and just buying your way out of any problem. Saying you have technology skills because you know some software from some vendor is like saying you can play guitar since you've got such high scores on the XBOX/PS3/Wii for Rock Band. Even if you know something from that "software vendor" inside & out, you don't know shit unless you understand the fundamentals under the hood of what the
Marketable vrs. Useful (Score:2)
Huh ? (Score:3, Insightful)
Being able to think makes you valuable. (Score:5, Insightful)
As a programmer, I can say that programming itself, that is, *how* to write code, in terms of methodology -- is a skill that will never leave you once you have acquired it (so long as you keep using it).
Almost any programmer worth their salt can learn a new language in a few weeks, if not days. Granted it may take more time to develop understanding of any idioms or warts the language may have, but you can learn that stuff on the fly, unless you're writing HA/mission critical code, in which case, there'd better be a review process, and it's reasonable to expect that someone on the team will be an expert in the technology being used.
So I'd say unless you've given up programming entirely and have moved on to a different career, your skills are still valuable, and will stay reasonably "fresh" even if you're writing code in a 30-year-old language (as the article says), as long as you actually think while you write code, and aren't just a copy/paste/munge wizard, not that there's anything wrong with that, for certain kinds of things.
This of course doesn't even consider the (imho) much more valuable part of being a software developer: being able to converse with non-technical people, in whatever human language you use, and then translate that into some sort of actionable programming work. That's often more than half the battle. Then of course there is testing, testing and testing.
The article isn't completely wrong, but (like much of the "IT industry") I think it missed the point of what skills are actually important to doing software development. Knowing how to use a specific bit of kit is pretty far down on the list, I think, for any reasonably competent programmer/technologist.
I treat anything with the word "marketability" in it with suspicion.
uh oh (Score:2)
oh, shit, I still use a keyboard! (Score:2)
OMG, so my skill of using a keyboard then has put me into a complete untenable position, career wise. I've been using a keyboard for over 28 years now. What do you people use today, do you talk into your mouse?
Generic skills seem to last (Score:2)
I design and configure GUI automated testing systems. The particulars change. The principles don't. I've started to design server and virtual machine environments to control the systems more precisely and easily. Think virtual machines are going away? Doubtful. I'm sure the particulars of these too will change, but the principles won't.
Yes and no. (Score:2)
The fact I can troubleshoot classic MacOS 7.6.1 up through 9.2.2 and a number of old-world PPC related hardware issues over the phone without being anywhere near the machine in question is hardly Buzzword Compliant in this day and age.
The fact that I learned basic troubleshooting out of self defense in that environment, however, gave me a great baseline for dealing with hardware and basic software issues in the general sense. While any classic MacOS-related "certifications" may be long useless, the fact th
old programming skills for new SoC+Card platforms (Score:2)
Back in the "olden" days, I grew up programming basic, assembly, and later in "C" on Vic20 and C64.
Now work for a major chip manufacturer, and some of my duties involves developing bootloaders, firmware, apps, and cryptographic libraries for SoC (System on Chip) and smart card platforms. Those old programming skills allow me to develop on platforms with extreme memory, storage, and performance limitations. I noticed that this is a challenge for those young wipper-snappers that only developed on platforms
Being current is about knowing the defects (Score:2)
Being current in some area of software today is about knowing its defects. You can read the manual about how it's supposed to work, but knowing what actually works is essential to high productivity.
I've been thinking about this recently in connection with Mozilla Jetpack, which is a library for making add-ons for Firefox, etc. There are two websites, a blog, a forum, a Google group, a development committee, two completely different sets of development tools, "hack sessions", and an IRC channel. There's
Boy, am I glad I picked EE... (Score:2)
Skills != Knowledge (Score:2)
Based on my experience, two years is a bit of an exaggeration. Just look at Windows XP. The software is ten years old and still in wide spread use. The real danger is missing out on major trends. In my own career, I almost missed out on virtualization. My employer did not have any plans on virtualizing and that is where the industry went. If I had stayed put, my career would have been dead.
On the other hand, my knowledge of IT allowed me to make a move into a better position with a company that did no
Re: (Score:2)
Getting out the C?O wing sounds like a forward move to me.
Re: (Score:2)
If you ever make it to CIO then- I'll swap jobs with you.
Re: (Score:2)
Re: (Score:2)
Actually, I'd guess applying Moore's law means you have to work twice as fast to keep up with the decay.