theodp writes "U.S. tech talent shortage discussions tend to focus on getting more young people to go to college to become CS grads. Nothing wrong with that, writes Anil Dash, but let's not forget about education which teaches mid-level programming as a skilled trade, suitable for apprenticeship and advancement in a way that parallels traditional trade skills like HVAC or welding. Dash encourages less of a focus on 'the next Zuckerberg' in favor of encouraging solid middle-class tech jobs that are primarily focused on creating and maintaining tech infrastructure in non-tech companies. Dash also suggests 'changing the conversation about recruiting technologists from the existing narrow priesthood of highly-skilled experts constantly chasing new technologies to productive workers getting the most out of widely-deployed platforms and frameworks.'"
Migrate from GitHub to SourceForge quickly and easily with this tool. Check out all of SourceForge’s recent improvements.×
An anonymous reader writes "What was taught to you about computers in High School? Computer use and computer science in schools are regular headlines, but what 'normal' do we compare it to? It's not a shared reference. A special class with Commodore PETs was set up just after I graduated, and I'm only starting to grey. Everybody younger has had progressive levels of exposure. What was 'normal' for our 40-, 30-, and 20-year olds here? And how well did it work for you, and your classmates?" For that matter, what's it like now — if you're in middle or high school now, or know students who are, what's the tech curriculum like?
theodp writes "HBS lecturer Robert C. Pozen says it's high time for management to stop emphasizing hours over results. By viewing those employees who come in over the weekend or stay late in the evening as more 'committed' and 'dedicated' to their work, as a UC Davis study showed, managers create a perverse incentive to not be efficient and get work done during normal business hours. 'It's an unfortunate reality that efficiency often goes unrewarded in the workplace,' writes Pozen. 'Focusing on results rather than hours will help you accomplish more at work and leave more time for the rest of your life.'"
pacopico writes "Much has been made about Facebook hitting 1 billion users. But Businessweek has the inside story detailing how the site actually copes with this many people and the software Facebook has invented that pushes the limits of computer science. The story quotes database guru Mike Stonebraker saying, 'I think Facebook has the hardest information technology problem on the planet.' To keep Facebooking moving fast, Mark Zuckerberg apparently instituted a program called Boot Camp in which engineers spend six-weeks learning every bit of Facebook's code."
ptorrone writes "Adafruit, the NYC based open-source hardware company led by Ladyada released their open-source Raspberry Pi WebIDE alpha today. Its goal is to be 'The easiest way to develop code on your Raspberry Pi.' To get up and running head on over to learn.adafruit.com/webide and follow the installation and setup instructions. It uses Bitbucket, and any code changes you make will be synced to your Bitbucket account. Adafruit chose Bitbucket over GitHub because they offer free secure accounts, which is very important for a Web-based IDE."
Talcyon writes "I'm a 40-year-old developer, and it's become apparent that my .NET skillset is woefully out of date after five years of doing various bits of support. I tried the 'Management' thing last year, but that was a failure as I'm just not a people person, and a full-on development project this year has turned into a disaster area. I'm mainly a VB.NET person with skills from the .NET 2.0 era. Is that it? Do I give up a career in technology now? Or turn around and bury myself in a support role, sorting out issues with other people's/companies' software? I've been lurking around Slashdot for many years now, and this question occasionally comes up, but it pays to get the opinions of others. Do I retrain and get back up to speed, or am I too old?"
Mark Hachman writes in Slash Datacenter that the Sparc T5 chip Oracle announced earlier this year apparently won't be ready until sometime in 2013. John Fowler, executive vice president, Systems, Oracle, presented at Oracle Open World a chart outlining highlights of Oracle's plans for the future. "But Fowler also skipped over some bad news: an apparent delay for the Sparc T5. A year ago, Oracle’s Sun division announced the Sparc T4—and according to Fowler, Oracle chief Larry Ellison set a very high bar for the next iteration: double the performance while maintaining app compatibility on an annual basis. Apparently, that didn’t quite happen with the T5; Oracle had the opportunity to announce a T5-based server, and didn’t. That’s a bit of bad news for the Sun design team, which already had to watch Intel’s Xeon chief, Diane Bryant, give the preceding keynote. ... As detailed at this year’s Hot Chips conference, the T5 combines 16 CPU cores running at 3.6 GHz on a 28-nm manufacturing process. Continuing the trend of hardware acceleration of specific functions, Sun executives claimed the chip would lead in on-chip encryption acceleration, with support for asymmetric (public key) encryption, symmetric encryption, hashing up to SHA-512, plus a hardware random number generator."
Nerval's Lobster writes "Google seems determined to press forward with Google Glass technology, filing a patent for a Google Glass wristwatch. As pointed out by CNET, the timepiece includes a camera and a touch screen that, once flipped up, acts as a secondary display. In the patent, Google refers to the device as a 'smart-watch. Whether or not a Google Glass wristwatch ever appears on the marketplace — just because a tech titan patents a particular invention doesn't mean it's bound for store shelves anytime soon — the appearance of augmented-reality accessories brings up a handful of interesting issues for everyone from app developers to those tasked with handling massive amounts of corporate data.For app developers, augmented-reality devices raise the prospect of broader ecosystems and spiraling complexity. It's one thing to build an app for smartphones and tablets — but what if that app also needs to handle streams of data ported from a pair of tricked-out sunglasses or a wristwatch, or send information in a concise and timely way to a tiny screen an inch in front of someone's left eye?"
CowboyRobot writes "'UML too complex? Flowcharts too old school? Mind maps offer a simple way to capture designs and weave them together elegantly.' The quickest way to begin designing a program is to simply write down the steps in normal text, but this method breaks down with more complex projects. UML can be a useful format for larger projects but can be difficult to get right, especially when trying to use it with a less conventional project. The middle ground are 'Mind Maps,' 'a diagrammatic representation of loosely connected ideas. They are a central tool in brainstorming sessions. Mind map tools help capture ideas and then mush them around until you have the structure you want.'"
angry tapir writes "Nokia and Oracle have joined forces on mapping, with details of the deal to be announced at the Oracle OpenWorld conference. To differentiate its smartphones from the competition, Nokia is betting big on location as well as imaging technology. Oracle is expected to add Nokia's mapping technology to its applications. Part of Nokia's location strategy is signing deals for the use of its Navteq mapping technology with as many companies as possible. Besides the deal with Oracle, Nokia has recently announced contracts with car makers BMW, Mercedes, Volkswagen and Korean Hyundai, which will all use Navteq map data in some of their vehicles. Garmin will also start using Nokia data on transit services and walking routes to power a new Urban Guidance feature, which will be available as part of its Navigon app for Android and iOS. Nokia's most important partner on navigation, though, is Microsoft. All smartphones based on Windows Phone 8 will have Nokia's Drive application as standard, while Microsoft's Bing Maps geographical search engine uses Nokia data."
Nerval's Lobster writes "Back in May, I took a look at three cloud management platforms: RightScale, Scalr, and enStratus. Perhaps the biggest surprise was that people from two of those companies—RightScale and Scalr—took note of the article and replied in the comments, offering some clarification on their offerings. (And they were very civil: thank you!) What I'd like to do next is re-visit these platforms, but focus directly on the APIs that the three offer—not so much coding, but a high-level picture of them. How do they stack up? What features do they have? How do they fit with standards? And what can you expect from the long-term?"
Nerval's Lobster writes "Oracle CEO Larry Ellison used his opening keynote at Oracle Open World (OOW) to unveil several initiatives to accelerate the cloud, including its own private cloud, Infrastructure-as-a-Service, and its latest database version—which, coincidentally, can be stored in memory within Oracle's latest Exadata database machines. Ellison also paid tribute to Oracle hardware partner Fujitsu, which had earlier announced 'Project Athena': a server designed with a UltraSPARC chip that (he claimed) can run the Oracle database 'faster than any microprocessor on the planet.' Ellison opened OpenWorld with four key announcements: that Oracle is now offering infrastructure as a service; that it will complement the IaaS offering by allowing customers to run that same infrastructure behind their corporate firewall as a private cloud; the launch of Oracle database 12C (where the 'c' stands for 'cloud'); and, finally, the new Exadata servers, which barely use disk drives at all in-favor of in-memory storage, with flash memory as a fallback."
First time accepted submitter abhi2012 writes "Noah Kagan, a former Facebook product manager, has written a brutally honest article about how and why he got fired from Facebook in 2006 and what he learned from it. The experience must be particularly painful, given that it eventually cost Kagan a $100 million fortune."
theodp writes "Blogger Floopsy complains that he would love to RTFM, but can't do so if no one will WTFM. 'You spend hours, days, months, perhaps years refining your masterpiece,' Floopsy laments to creators of otherwise excellent programming language, framework, and projects. 'It is an expression of your life's work, heart and soul. Why, then, would you shortchange yourself by providing poor or no documentation for the rest of us?' One problem with new program languages, a wise CS instructor of mine noted in the early look-Ma-no-documentation days of C++, is that their creators are not typically professional writers and shy away from the effort it takes to produce even less-than-satisfactory manuals. But without these early efforts, he explained, the language or technology may never gain enough traction for the Big Dogs like O'Reilly to come in and write the professional-caliber books that are necessary for truly widespread adoption. So, how important is quality documentation to you as a creator or potential user of new technologies? And how useful do you find the documentation that tech giants like Google (Go), Twitter (Bootstrap), Facebook (iOS 6 Facebook Integration), Microsoft (Windows Store apps), and Apple (Create Apps for IOS 6) produce to promote their nascent technologies? Is it useful on its own, or do you have to turn to other 'store-bought' documentation to really understand how to get things done?"