mikejuk writes "Is it possible that we have been wasting our time typing programs. Could voice recognition, with a little help from an invented spoken language, be the solution we didn't know we needed? About two years ago Tavis Rudd, developed a bad case of RSI caused by typing lots of code using Emacs. It was so severe that he couldn't code. As he puts it: 'Desperate, I tried voice recognition'. The Dragon Naturally Speaking system used by Rudd supported standard language quite well, but it wasn't adapted to program editing commands. The solution was to use a Python speech extension, DragonFly, to program custom commands. OK, so far so good, but ... the commands weren't quite what you might have expected. Instead of English words for commands he used short vocalizations — you have to hear it to believe it. Now programming sounds like a conversation with R2D2. The advantage is that it is faster and the recognition is easier — it also sounds very cool and very techie. it is claimed that the system is faster than typing. So much so that it is still in use after the RSI cleared up."
Follow Slashdot stories on Twitter
msheekhah writes "I have a friend who, when he gets out of college, has been promised a job at well known electronics company with a salary around $70k. However, he wants to instead go work for Blizzard or some other game company as a game programmer. I've read enough on here and on other tech websites to know that he should take the job he's been offered. Can you share with me your experiences so I can give him real life examples to convince him to take this job? If your experience is contrary to mine, I'd appreciate that input as well."
An anonymous reader writes "This article takes a look at whether web apps will ever match desktop and mobile apps in terms of performance and usability. Jo Rabin, who's leading the push by web standards body W3C to get web app performance up to scratch, is optimistic web apps will eventually be the default choice for building the majority of commercial and business apps, while the article weighs up just how much web technologies need to be improved before this could happen. Quoting: 'Native apps are generally first to gain access to new platform-specific hardware features — be it navigating using a phone's GPS and accelerometer or taking pictures with a phone's camera. But if a particular hardware feature becomes popular, standards to implement that feature in the browser will always follow, Rabin said. Work is taking place within W3C to standardise APIs for web technologies to access many of the features found on modern smartphones. Ongoing work this year includes setting out a system-level API to allow a web app to manage a device's contacts book, a messaging API for sending and receiving SMS and MMS, new mechanisms for capturing photos and recordings, new event triggers that could handle mouse, pen and touch inputs, a new push API to allow web apps to receive messages in the background, new media queries for responsive web design, an API for exchanging information using NFC and precise control over resource loading times in a web document.'"
An anonymous reader writes "Here in the U.S., 'being professional' means giving at least two week's notice when leaving a job. Is this an outmoded notion? We've all heard stories about (or perhaps experienced) a quick escort to the parking lot upon giving the normal notice, and I've never heard of a company giving a two-week notice to an employee that's being laid off or fired. A generation ago, providing a lengthy notice was required to get a glowing reference, but these days does a reference hold water any more? Once you're reached the point where you know it's time to leave, under what circumstances would you just up and walk out or give only a short notice?"
New submitter Lemeowski writes "Critics have been pounding GitHub recently, claiming it is hosting tons of code with no explicit software license. The debate was thrust into the limelight last year when James Governor of RedMonk issued an acclaimed tweet about young developers being 'about POSS — post open source software,' meaning they disliked or avoided licensing and governance. Red Hat's IP attorney Richard Fontana explores the complaint saying there is a positive aspect of the POSS and GitHub phenomenon: Developers are, for the first time in the history of free software, helping inform each other about licensing and aiding in the selection process. The result is that it's becoming easier to suggest legal improvements to GitHub-hosted repositories."
Nerval's Lobster writes "Oracle CEO Larry Ellison thinks that Apple will collapse without Steve Jobs at the helm. In a televised interview with CBS News, scheduled to air August 13, Ellison called the deceased Jobs 'brilliant' and compared him to iconic creators such as Thomas Edison and Pablo Picasso. When asked about Apple's future now that Jobs is dead, Ellison didn't hold back: 'We already know, we saw — we conducted the experiment, it's been done.' Raising his hand above his head, presumably to indicate the rise of Apple's fortunes during Jobs' initial reign, Ellison said: 'We saw Apple with Steve Jobs.' Then he lowered his hand: "We saw Apple without Steve Jobs." In other words, the period following Jobs' ouster, when the company's revenues declined and it launched whole portfolios of consumer products that failed. 'We saw Apple with Steve Jobs,' Ellison continued, raising his hand above his head again — this time, to suggest that incandescent period following Jobs' return to the company, when it released the iPod, iPhone, iPad, and a variety of bestselling PCs. 'And now, we're going to see Apple without Steve Jobs,' he finished, and his hand fell."
An anonymous reader writes "Simple DirectMedia Layer 2.0 has finally been released. The cross-platform multimedia layer used by hundreds of cross-platform games has seen its first major release in years. The SDL 2.0 release has many new features including GL3 and OpenGL ES rendering support, a new 2D rendering API, better full-screen / multi-window support, multiple input support, Android and iOS support, power management, and other new functionality. SDL 2.0 can be downloaded from libsdl.org."
First time accepted submitter jovius writes "The Matriculation Examination Board of Finland has just opened an international hacking contest to find flaws and exploits in Digabi Live — the Live Debian based operating system to be used in the all-digital final exams by the year 2016. The contest ends on 1st of September, and the winners are about to scoop hefty hardware prizes, also available as cash."
theodp writes "Bret Victor's The Future of Programming (YouTube video; Vimeo version) should probably be required viewing this fall for all CS majors — and their professors. For his recent DBX Conference talk, Victor took attendees back to the year 1973, donning the uniform of an IBM systems engineer of the times, delivering his presentation on an overhead projector. The '60s and early '70s were a fertile time for CS ideas, reminds Victor, but even more importantly, it was a time of unfettered thinking, unconstrained by programming dogma, authority, and tradition. 'The most dangerous thought that you can have as a creative person is to think that you know what you're doing,' explains Victor. 'Because once you think you know what you're doing you stop looking around for other ways of doing things and you stop being able to see other ways of doing things. You become blind.' He concludes, 'I think you have to say: "We don't know what programming is. We don't know what computing is. We don't even know what a computer is." And once you truly understand that, and once you truly believe that, then you're free, and you can think anything.'"
reifman writes "Nothing sucks more than finding an 'Error establishing database connection' on your blog hours after the fact, but it's not easy to find inexpensive, simple monitoring solutions which support smartphone notifications. I wrote MonitorApp, a free, open source software applet which sends notifications to your iPhone (or Android) if anything goes wrong with your web site or services. This tutorial describes how to install and configure MonitorApp for your own purposes. The only cost is a $4.99 mobile application called Pushover — which links MonitorApp to your phone. Pushover also links with Nagios, a more complex open source option — but ironically, Nagios' website was down when I looked for it last month."
hypnosec writes "NVidia has now open-sourced the operating system that powers the gaming console to encourage its modification and further development. Powered by NVidia's homegrown Tegra 4 processor, the console runs Android, which shouldn't surprise many as the company moves ahead with its open-sourcing intentions. The GPU company has said that the SHIELD is an 'open gaming platform' that allows for 'an open ecosystem,' enabling developers to develop content as well as applications that takes advantage of the underlying hardware and which can be enjoyed on bigger displays as well as mobile screen." Playing with it isn't without risks (like potentially voiding the warranty), but NVIDIA's blog post says they're also providing a recovery image to fall back to.
gspec writes "I am an engineer with about 14 years experience in the industry. Lately I have been interviewing with a few companies hoping to land a better position. In almost all those interviews, I was asked these types of question: 'Have you been a leader in a project?' or 'Why after these many years, you are not in a management? Do you lack leadership skills?' Sometimes these questions discourage me and make me feel like an underachiever. I found an article in which the author talked about exactly this, and I agree with him. I think in this modern society, especially in the U.S., we overvalue the leaders and undervalue the followers to the point that we forget that leaders cannot do any good if they do not have good followers."
reifman writes "Zillow quietly released boundary data for more than 7,000 neighborhoods in the U.S. via the Creative Commons attribute-sharealike license but few people know how to integrate this data into their applications. This tutorial describes how to import the data and integrate it with Google Maps and HTML5 Geolocation."
chicksdaddy writes "Two researchers at the Black Hat Briefings security conference Thursday said Smart TVs from electronics giant Samsung are rife with vulnerabilities in the underlying operating system and Java-based applications. Those vulnerabilities could be used to steal sensitive information on the device owner, or even spy on the television's surroundings using an integrated webcam. Speaking in Las Vegas, Aaron Grattafiori and Josh Yavor, both security engineers at the firm ISEC Partners, described Smart TVs as Linux boxes outfitted with a Webkit-based browser. They demonstrated how vulnerabilities in SmartHub, the Java-based application that is responsible for many of the Smart TV's interactive features, could be exploited by a local or remote attacker to surreptitiously activate and control an embedded webcam on the SmartTV, launch drive-by download attacks and steal local user credentials and those of connected devices, browser history, cache and cookies as well as credentials for the local wireless network. Samsung has issued patches for many of the affected devices and promises more changes in its next version of the Smart TV. This isn't the first time Smart TVs have been shown to be vulnerable. In December, researchers at the firm ReVuln also disclosed a vulnerability in the Smart TV's firmware that could be used to launch remote attacks."
msmoriarty writes "Google's Don Dodge, GitHub's Tom Preston-Werner, New Relic's Lew Cirne and others recently got together in San Francisco on a panel called 'The Developer is King: The Power Behind the Throne.' According to coverage of the event, the panelists all agreed that programmers — both independent ones and those employed by companies — have more power, and thus opportunities, than ever. Even the marketing power of developers was acknowledged: 'The only way to convince a developer is by giving them a demo and showing them how it's better,' said Preston-Werner. 'The beauty is, you plant these seeds around the world, and those people will evangelize it for you. Because another thing that developers are great at is telling other developers what works for them.'"
twofishy writes "Something I've noticed amongst financial service companies in London is a growing use of Java in preference to C/C++ for exchange systems, High Frequency Trading and over low-latency work. InfoQ has a good written panel discussion with Peter Lawrey, Martin Thompson, Todd L. Montgomery and Andy Piper. From the article: 'Often the faster an algorithm can be put into the market, the more advantage it has. Many algorithms have a shelf life and quicker time to market is key in taking advantage of that. With the community around Java and the options available, it can definitely be a competitive advantage, as opposed to C or C++ where the options may not be as broad for the use case. Sometimes, though, pure low latency can rule out other concerns. I think currently, the difference in performance between Java and C++ is so close that it's not a black and white decision based solely on speed. Improvements in GC techniques, JIT optimizations, and managed runtimes have made traditional Java weaknesses with respect to performance into some very compelling strengths that are not easy to ignore.'"
Nerval's Lobster writes "Developer and editor Jeff Cogswell asks: When it comes to implementing a CouchDB installation, do you roll your own, or go with a service that provides a hosted version of the database? He takes a look at some of the technologies present in CouchDB that can greatly influence that decision. His conclusion? Like all things, it's a little complicated. 'If you're going to be self-hosting—unless you're working on a really small system—don't use the basic CouchDB for anything,' he writes. 'If you want scalability, either go with Couchbase or BigCouch, or wait until Cloudant's BigCouch merger into CouchDB is officially available.' But going with a host also creates its own things to watch for, including potential issues with replication and eventual consistency."
snydeq writes "Taming technology is sometimes more art than science, but the difference can sometimes be hard to discern, writes Deep End's Paul Venezia. 'You've probably come across colleagues who were extremely skilled at their jobs — system administrators who can bend a zsh shell to their every whim, or developers who can write lengthy functions that compile without a whimper the first time. You've probably also come across colleagues who were extremely talented — who could instantly visualize a new infrastructure addition and sketch it out to extreme detail on a whiteboard while they assembled it in their head, for example, or who could devise a new, elegant UI without breaking a sweat. The truly gifted among us exhibit both of those traits, but most fall into one category or another. There is a difference between skill and talent. Such is true in many vocations, of course, but IT can present a stark contrast between the two.'"Assuming Venezia is correct, which do you think is more important?
theodp writes "In the movie Groundhog Day, a weatherman finds himself living the same day over and over again. It's a tale to which software-designers-of-a-certain-age can relate. Like Philip Greenspun, who wrote in 1999, 'One of the most painful things in our culture is to watch other people repeat earlier mistakes. We're not fond of Bill Gates, but it still hurts to see Microsoft struggle with problems that IBM solved in the 1960s.' Or Dave Winer, who recently observed, 'We marvel that the runtime environment of the web browser can do things that we had working 25 years ago on the Mac.' And then there's Scott Locklin, who argues in a new essay that one of the problems with modern computer technology is that programmers don't learn from the great masters. 'There is such a thing as a Beethoven or Mozart of software design,' Locklin writes. 'Modern programmers seem more familiar with Lady Gaga. It's not just a matter of taste and an appreciation for genius. It's a matter of forgetting important things.' Hey, maybe it's hard to learn from computer history when people don't acknowledge the existence of someone old enough to have lived it, as panelists reportedly did at an event held by Mark Zuckerberg's FWD.us last Friday!"