Forgot your password?

Want to read Slashdot from your mobile device? Point it at and keep reading!


US Senator Blasts Microsoft's H-1B Push As It Lays 18,000 Off Workers 514

Posted by Unknown Lamer
from the good-for-nothing-lazy-programmers dept.
dcblogs (1096431) writes On the floor of U.S. Senate Thursday, Sen. Jeff Sessions delivered a scalding and sarcastic attack on the use of highly skilled foreign workers by U.S. corporations that was heavily aimed at Microsoft, a chief supporter of the practice. Sessions' speech began as a rebuttal to a recent New York Times op-ed column by Microsoft founder Bill Gates, investor Warren Buffett and Sheldon Adelson ... But the senator's attack on "three of our greatest masters of the universe," and "super billionaires," was clearly primed by Microsoft's announcement, also on Thursday, that it was laying off 18,000 employees. "What did we see in the newspaper today?" said Sessions, "News from Microsoft. Was it that they are having to raise wages to try to get enough good, quality engineers to do the work? Are they expanding or are they hiring? No, that is not what the news was, unfortunately. Not at all."

Math, Programming, and Language Learning 236

Posted by Soulskill
from the i-before-e-except-SyntaxError: dept.
An anonymous reader writes: There's often debate amongst modern programmers about how much math a professional developer should know, and to what extent programming is math. Learning to program is often viewed as being on a spectrum between learning math and learning spoken/written languages. But in a new article, Jeremy Kun argues that the spectrum should be formulated another way: Human language -> Mathematics -> Programming. "Having studied all three subjects, I'd argue that mathematics falls between language and programming on the hierarchy of rigor. ... [T]he hierarchy of abstraction is the exact reverse, with programming being the most concrete and language being the most abstract. Perhaps this is why people consider mathematics a bridge between human language and programming. Because it allows you to express more formal ideas in a more concrete language, without making you worry about such specific hardware details like whether your integers are capped at 32 bits or 64. Indeed, if you think that the core of programming is expressing abstract ideas in a concrete language, then this makes a lot of sense. This is precisely why learning mathematics is 'better' at helping you learn the kind of abstract thinking you want for programming than language. Because mathematics is closer to programming on the hierarchy. It helps even more that mathematics and programming readily share topics."

Microsoft's Missed Opportunities: Memo From 1997 161

Posted by Unknown Lamer
from the hyper-cube-os dept.
New submitter gthuang88 (3752041) writes In the 1990s, Microsoft was in position to own the software and devices market. Here is Nathan Myhrvold's previously unpublished 1997 memo on expanding Microsoft Research to tackle problems in software testing, operating systems, artificial intelligence, and applications. Those fields would become crucial in the company's competition with Google, Apple, Amazon, and Oracle. But research didn't do enough to make the company broaden its businesses. While Microsoft Research was originally founded to ensure the company's future, the organization only mapped out some possible futures. And now Microsoft is undergoing the biggest restructuring in its history. At least F# and LINQ saw the light of day.

ChickTech Brings Hundreds of Young Women To Open Source 158

Posted by Soulskill
from the more-engineers-more-cool-stuff dept.
ectoman writes: is running an interview with Jennifer Davidson of ChickTech, a non-profit organization whose mission is to create communities of support for women and girls pursuing (or interested in pursuing) careers in tech. "In the United States, many girls are brought up to believe that 'girls can't do math' and that science and other 'geeky' topics are for boys," Davidson said. "We break down that idea." Portland, OR-based ChickTech is quickly expanding throughout the United States—to cities like Corvallis and San Francisco—thanks to the "ChickTech: High School" initiative, which gathers hundreds of young women for two-day workshops featuring open source technologies. "We fill a university engineering department with 100 high school girls—more girls than many engineering departments have ever seen," Davidson said. "The participants can look around the building and see that girls from all backgrounds are just as excited about tech as they are."
Hardware Hacking

SRI/Cambridge Opens CHERI Secure Processor Design 59

Posted by Unknown Lamer
from the dreaming-of-hurd/coyotos dept.
An anonymous reader writes with some exciting news from the world of processor design: Robert Watson at Cambridge (author of Capsicum) has written a blog post on SRI/Cambridge's recent open sourcing of the hardware and software for the DARPA-sponsored CHERI processor — including laser cutting directions for an FPGA-based tablet! Described in their paper The CHERI Capability Model: Reducing Risk in an age of RISC, CHERI is a 64-bit RISC processor able to boot and run FreeBSD and open-source applications, but has a Clang/LLVM-managed fine-grained, capability-based memory protection model within each UNIX process. Drawing on ideas from Capsicum, they also support fine-grained in-process sandboxing using capabilities. The conference talk was presented on a CHERI tablet running CheriBSD, with a video of the talk by student Jonathan Woodruff (slides).

Although based on the 64-bit MIPS ISA, the authors suggest that it would also be usable with other RISC ISAs such as RISC-V and ARMv8. The paper compares the approach with several other research approaches and Intel's forthcoming Memory Protection eXtensions (MPX) with favorable performance and stronger protection properties.
The processor "source code" (written in Bluespec Verilog) is available under a variant of the Apache license (modified for application to hardware). Update: 07/16 20:53 GMT by U L : If you have any questions about the project, regular Slashdot contributor TheRaven64 is one of the authors of the paper, and is answering questions.

Is the Software Renaissance Ending? 170

Posted by Soulskill
from the da-vinci-code dept.
An anonymous reader writes Writer and former software engineer Matt Gemmell adds his voice to the recent rumblings about writing code as a profession. Gemmell worries that the latest "software Renaissance," which was precipitated by the explosion of mobile devices, is drawing to a close. "Small shops are closing. Three-person companies are dropping back to sole proprietorships all over the place. Products are being acquired every week, usually just for their development teams, and then discarded. The implacable, crushing wheels of industry, slow to move because of their size, have at last arrived on the frontier. Our frontier, or at least yours now. I've relinquished my claim." He also pointed out the cumulative and intractable harm being done by software patents, walled-garden app stores, an increasingly crowded market, and race-to-the-bottom pricing. He says that while the available tools make it a fantastic time to develop software, actually being an independent developer may be less sustainable than ever.

Ask Slashdot: Future-Proof Jobs? 506

Posted by Soulskill
from the robot-overlord-exterminator dept.
An anonymous reader writes: My niece, who is graduating from high school, has asked me for some career advice. Since I work in data processing, my first thought was to recommend a degree course in computer science or computer engineering. However, after reading books by Jeremy Rifkin (The Third Industrial Revolution) and Ray Kurzweil (How to Create a Mind), I now wonder whether a career in information technology is actually better than, say, becoming a lawyer or a construction worker. While the two authors differ in their political persuasions (Rifkin is a Green leftist and Kurzweil is a Libertarian transhumanist), both foresee an increasingly automated future where most of humanity would become either jobless or underemployed by the middle of the century. While robots take over the production of consumer hardware, Big Data algorithms like the ones used by Google and IBM appear to be displacing even white collar tech workers. How long before the only ones left on the payroll are the few "rockstar" programmers and administrators needed to maintain the system? Besides politics and drug dealing, what jobs are really future-proof? Would it be better if my niece took a course in the Arts, since creativity is looking to be one of humanity's final frontiers against the inevitable Rise of the Machines?

Today In Year-based Computer Errors: Draft Notices Sent To Men Born In the 1800s 205

Posted by timothy
from the pa-dmv-never-did-me-any-favors-either dept.
sandbagger (654585) writes with word of a Y2K-style bug showing up in Y2K14: "The glitch originated with the Pennsylvania Department of Motor Vehicles during an automated data transfer of nearly 400,000 records. The records of males born between 1993 and 1997 were mixed with those of men born a century earlier. The federal agency didn't know it because the state uses a two-digit code to indicate birth year." I wonder where else two-digit years are causing problems; I still see lots of paper forms that haven't made the leap yet to four digits.

Prof. Andy Tanenbaum Retires From Vrije University 136

Posted by timothy
from the congratulations-and-good-wishes dept.
When Linus Torvalds first announced his new operating system project ("just a hobby, won't be big and professional like gnu"), he aimed the announcement at users of Minix for a good reason: Minix (you can download the latest from the Minix home page) was the kind of OS that tinkerers could afford to look at, and it was intended as an educational tool. Minix's creator, Professor Andrew Stuart "Andy" Tanenbaum, described his academic-oriented microkernel OS as a hobby, too, in the now-famous online discussion with Linus and others. New submitter Thijssss (655388) writes with word that Tanenbaum, whose educational endeavors led indirectly to the birth of Linux, is finally retiring. "He has been at the Vrije Universiteit for 43 years, but everything must eventually end."

Ask Slashdot: Best Dedicated Low Power Embedded Dev System Choice? 183

Posted by samzenpus
from the what-should-I-get? dept.
An anonymous reader writes "I'm a Solaris user which is not well supported by the OSS toolchains. I'd like to have a dedicated Linux based dev system which has good support for ARM, MSP430 and other MCU lines and draws very little (5-10 watts max) power. The Beaglebone Black has been suggested. Is there a better choice? This would only be used for software development and testing for embedded systems."

Will Google's Dart Language Replace Javascript? (Video) 180

Posted by Roblimo
from the shall-we-play-darts-or-javascripts-this-evening-at-the-pub? dept.
Seth Ladd, Google Web engineer and Chrome Developer Advocate, is today's interviewee. He's talking about Dart, which Wikipedia says is 'an open-source Web programming language developed by Google.' The Wikipedia article goes on to say Dart was unveiled at the GOTO conference in Aarhus, October 10–12, 2011, and that the goal of Dart is 'ultimately to replace JavaScript as the lingua franca of web development on the open web platform.' A bold aim, indeed. Last month (June, 2014), InfoWorld ran an article by Paul Krill headlined, Google's Go language on the rise, but Dart is stalling. Seth Ladd, unlike Paul Krill, is obviously rah-rah about Dart -- which is as it should be, since that's his job -- and seems to think it has a growing community and a strong place in the future of Web programming. For more about Dart, scroll down to watch Tim Lord's video interview with Seth -- or read the transcript, if you prefer. (Alternate Video Link)

Normal Humans Effectively Excluded From Developing Software 608

Posted by Unknown Lamer
from the elitism-at-its-finest dept.
theodp (442580) writes Over at Alarming Development, Jonathan Edwards has an interesting rant entitled Developer Inequality and the Technical Debt Crisis. The heated complaints that the culture of programming unfairly excludes some groups, Edwards feels, is a distraction from a bigger issue with far greater importance to society.

"The bigger injustice," Edwards writes, "is that programming has become an elite: a vocation requiring rare talents, grueling training, and total dedication. The way things are today if you want to be a programmer you had best be someone like me on the autism spectrum who has spent their entire life mastering vast realms of arcane knowledge — and enjoys it. Normal humans are effectively excluded from developing software. The real injustice of developer inequality is that it doesn't have to be this way." Edwards concludes with a call to action, "The web triumphalists love to talk about changing the world. Well if you really want to change the world, empower regular people to build web apps. Disrupt web programming! Who's with me?" Ed Finkler, who worries about his own future as a developer in The Developer's Dystopian Future, seconds that emotion. "I think about how I used to fill my time with coding," Finkler writes. "So much coding. I was willing to dive so deep into a library or framework or technology to learn it. My tolerance for learning curves grows smaller every day. New technologies, once exciting for the sake of newness, now seem like hassles. I'm less and less tolerant of hokey marketing filled with superlatives. I value stability and clarity."

US Tech Firms Recruiting High Schoolers (And Younger) 253

Posted by Soulskill
from the there-oughta-be-a-law-enforcing-the-laws-we-already-have dept.
ShaunC writes: Is there a glut of qualified American tech workers, or isn't there? Some companies like Facebook and Airbnb are now actively courting and recruiting high school students as young as 13 with promises of huge stipends and salaries. As one student put it, "It's kind of insane that you can make more than the U.S. average income in a summer." Another who attended a Facebook-sponsored trip said he'd "forego college for a full-time job" if it were offered. Is Silicon Valley taking advantage of naive young workers?

Python Bumps Off Java As Top Learning Language 415

Posted by Soulskill
from the from-college-import-education dept.
itwbennett writes: Python has surpassed Java as the top language used to introduce U.S. students to programming and computer science, according to a recent survey posted by the Association for Computing Machinery (ACM). Eight of the top 10 computer science departments now use Python to teach coding, as well as 27 of the top 39 schools, indicating that it is the most popular language for teaching introductory computer science courses, according to Philip Guo, a computer science researcher who compiled the survey for ACM."

Indie Game Developers Talk About Why They Struck Out On Their Own 49

Posted by timothy
from the you-can-hire-someone-to-flog-you dept.
Nerval's Lobster (2598977) writes Technology writer Jon Brodkin sat down with a group of indie game developers (as well as a professor at the University of Southern California's game-design program) to talk about why they decided to launch their own small studios rather than stick with comfortable (albeit stressful) jobs at major firms like Disney or Zynga. The answer, as you'd expect, boils down to control. "Working for a bigger company is a good way to gain experience, and learn how games are made," said Graham Smith, one of the co-founders of Toronto-based DrinkBox Studios. "It's also nice to have a steady salary coming in as you learn the ropes. On the flip side, depending on the company, you might not have much control over the game's design, or even be making the types of games that you enjoy playing." But startups come with their own challenges, not the least of which is the prospect of an economic downturn quickly wiping you out, or not making your Kickstarter goal.

"The Street finds its own uses for technology." -- William Gibson