Nerval's Lobster writes "A compiler can take your C++ loops and create vectorized assembly code for you. It's obviously important that you RTFM and fully understand compiler options (especially since the defaults may not be what you want or think you're getting), but even then, do you trust that the compiler is generating the best code for you? Developer and editor Jeff Cogswell compares the g++ and Intel compilers when it comes to generating vectorized code, building off a previous test that examined the g++ compiler's vectorization abilities, and comes to some definite conclusions. 'The g++ compiler did well up against the Intel compiler,' he wrote. 'I was troubled by how different the generated assembly code was between the 4.7 and 4.8.1 compilers—not just with the vectorization but throughout the code.' Do you agree?"
Slashdot stories can be listened to in audio form via an RSS feed, as read by our own robotic overlord.
hessian writes "According to a news release from Purdue University, 'Researchers are developing computers capable of "approximate computing" to perform calculations good enough for certain tasks that don't require perfect accuracy, potentially doubling efficiency and reducing energy consumption. "The need for approximate computing is driven by two factors: a fundamental shift in the nature of computing workloads, and the need for new sources of efficiency," said Anand Raghunathan, a Purdue Professor of Electrical and Computer Engineering, who has been working in the field for about five years. "Computers were first designed to be precise calculators that solved problems where they were expected to produce an exact numerical value. However, the demand for computing today is driven by very different applications. Mobile and embedded devices need to process richer media, and are getting smarter – understanding us, being more context-aware and having more natural user interfaces. ... The nature of these computations is different from the traditional computations where you need a precise answer."' What's interesting here is that this is how our brains work."
Hugh Pickens DOT Com writes "Chuong Nguyen reports that Apple is forcing developers to adopt iOS 7's visual UI for their apps, and has advised iOS developers that all apps submitted after February 1, 2014 must be optimized for iOS 7 and built using Xcode 5 ... 'It's likely that Apple is more anxious than ever for developers to update their apps to fit in visually and mechanically with iOS 7, as it's the largest change in the history of Apple's mobile software,' says Matthew Panzarino. 'iOS 7 introduced a much more complex physical language while stripping out many of the visual cues that developers had relied on to instruct users. For better or worse, this has created a new aesthetic that many un-updated apps did not reflect.' Most app developers have been building apps optimized towards iOS 7 since Apple's World Wide Developer Conference in June 2013. Apple has been on a push over the past couple of years to encourage developers to support the latest editions of its OS faster than ever. To do this, it's made a habit of pointing out the adoption rates of new versions of iOS, which are extremely high. Nearly every event mentions iOS 7 adoption, which now tops 76% of all iOS users, and Apple publishes current statistics. In order to optimize apps for the new operating system, they must be built with the latest version of Xcode 5 which includes 64-bit support and access to new features like backgrounding APIs."
tsu doh nimh writes "Security experts have long opined that one way to make software more secure is to hold software makers liable for vulnerabilities in their products. This idea is often dismissed as unrealistic and one that would stifle innovation in an industry that has been a major driver of commercial growth and productivity over the years. But a new study released this week presents perhaps the clearest economic case yet for compelling companies to pay for information about security vulnerabilities in their products. Stefan Frei, director of research at NSS Labs, suggests compelling companies to purchase all available vulnerabilities at above black-market prices, arguing that even if vendors were required to pay $150,000 per bug, it would still come to less than two-tenths of one percent of these companies' annual revenue (PDF). To ensure that submitted bugs get addressed and not hijacked by regional interests, Frei also proposes building multi-tiered, multi-region vulnerability submission centers that would validate bugs and work with the vendor and researchers. The questions is, would this result in a reduction in cybercrime overall, or would it simply hamper innovation? As one person quoted in the article points out, a majority of data breaches that cost companies tens of millions of dollars have far more to do with other factors unrelated to software flaws, such as social engineering, weak and stolen credentials, and sloppy server configurations."
theodp writes "On the final day of Computer Science Education Week, the Hour of Code bravado continues. Around 12:30 a.m. Sunday (ET), Code.org was boasting that in just 6 days, students of its tutorials have "written" more than 10x the number of lines of code in Microsoft Windows. "Students of the Code.org tutorials have written 507,152,775 lines of code. Is this a lot? By comparison, the Microsoft Windows operating system has roughly 50 million lines of code." Code.org adds, "In total, 15,481,846 students have participated in the Hour of Code. Of this group, 6,872,757 of them used the tutorials by Code.org, and within the Code.org tutorial, they've written 507,152,775 lines of code." On YouTube, however, a playlist of the Code.org tutorial videos has distinctly lower numbers, with only 2,246 views of the Code.org Wrap Up video reported as of this writing. So, any thoughts on why the big disconnect, and how close the stats might reflect reality? Code.org does explain that an 'Hour of Code' is not necessarily an 'hour of code' ("Not everybody finishes an Hour of Code tutorial. Some students spend one hour. Some spend 10 minutes. Some spend days. Instead of counting how many students 'finish one hour'; or how much time they spent, this [LOC] is our simplest measure of progress"). So, with millions being spent on efforts to get Code.org into the nation's schools — New York and Chicago have already committed their 1.5 million K-12 students — is it important to get a better understanding of what the Hour of Code usage stats actually represent — and what their limitations might be — and not just accept as gospel reports like AllThingsD's 15 Million Students Learned to Program This Week, Thanks to Hour of Code ("every other school family in the U.S. has a child that has done the Hour of Code")?"
mikejuk writes "Google's Dart just reached version 1.0, but now it seems that it has aspirations to being an international standard. The question is will this make any difference to the language's future? Given that Google effectively owns Dart, what advantage does standardization bring? The answer to what Google thinks it brings is indicated in the Chromium blog: 'The new standardization process is an important step towards a future where Dart runs natively in web browsers.' and this seems reasonable. A standard is something that would be required before other browser makers decided to fall in line and support native Dart. It is probably a necessary but far from sufficient condition, however, with Microsoft, Apple and Mozilla having other interests to further. Last but not least, having the backing of a standard might just encourage possible users to believe that the language won't sink if Google gets distracted with other projects and decides that Dart is dispensable. However, a strong open source development community capable of supporting Dart without Google's input would be a better reassurance. If you want to help, Google would like you to join the committee. After all, it still doesn't have a Vice Chair. So can we expect to see ECMA CoffeeScript or TypeScript in the near future? Probably not."
theodp writes "Code.org, backed by Bill Gates and Mark Zuckerberg, boasts in a blog post that thanks to this week's Hour of Code, which featured a Blockly tutorial narrated by Gates and Zuckerberg, 'More students have participated in computer science in U.S. schools in the last three days than in the last 100 years.' Taking note of the impressive numbers being put up on the Hour of Code Leaderboards ('12,522,015 students have done the Hour of Code and written 406,022,512 lines of code'), the Seattle Times adds that 'More African American and Hispanic kids learned about the subject in two days than in the entire history of computer science,' and reports that the cities of Chicago and New York have engaged Code.org to offer CS classes in their schools. So, isn't it a tad hyperbolic to get so excited over kids programming with blocks? 'Yes, we can all agree that this week's big Hour of Code initiative is a publicity stunt,' writes the Mercury News' Mike Cassidy, 'but you know what? A publicity stunt is exactly what we need.'"
New submitter KDE Community writes that the KDE project has released KDevelop 4.6.0 as the latest version of the free and open source integrated development environment. "KDevelop 4.6.0 improves debugging support with GDB. The GDB integration improvements include some operations now going into effect immediately rather than needing to re-run the program, improved debugging from external terminals, and a CPU registers toolview. KDevelopers' CPU registers toolview also allows for showing and editing all user-mode registers and general purpose flags for x86/x86_64/ARMv7 platforms. Other KDevelop 4.6.0 changes include greater language support within the PHP plug-in, Python language support improvements, more C++11 language support, improved project management, and a clean-up to the IDE's user-interface."
dmiller1984 writes "The Chicago Public Schools, the third-largest public school system in the United States, announced a five-year plan today that would add at least one computer science course to every CPS high school, and elevate computer science to a core requirement instead of an elective. CPS announced this through a partnership with code.org, stating that the non-profit would provide free curriculum, professional development, and stipends for teachers."
An anonymous reader writes "Ekisto is an interactive network visualization of three online communities: StackOverflow, Github and Friendfeed. Ekisto tries to map our online habitats using graph algorithms and the city as a metaphor. A graph layout algorithm arranges users in 2D space based on their similarity. Cosine similarity is computed based on the users' network (Friendfeed), collaborate, watch, fork and follow relationships (Github), or based on the tags of posts contributed by users (StackOverflow). The height of each user represents the normalized value of the user's Pagerank (Github, Friendfeed) or their reputation points (StackOverflow)."
the agent man writes "The Hour of Code event taking place December 9-15 has produced a number of tutorials with the goal to excite 10 millions kids to code. It's really interesting to contrast the different pedagogical approaches behind the roughly 30 tutorials. The University of Colorado's 'Make a 3D Game' tutorial wants to excite kids to code by focusing less on coding. This pedagogy is based on the idea that coding alone, without non-coding creativity, has a hard time attracting kids who are skeptical of computer science, including a high percentage of girls who think 'programming is hard and boring.' Instead, the 'Make a 3D Game' activity has the kids create sharable 3D shapes and 3D worlds in their browsers, which they then want to bring to life — through coding. There is evidence that this strategy works. The article talks about the research exploring how kids get excited through game design, and how they can later leverage coding skills acquired to make science simulations. You can try the activity by yourself or with your kids, if you're curious."
Jah-Wren Ryel writes "What do you get when you train a Markov chain on the King James Bible and a copy of Structure and Interpretation of Computer Programs? King James Programming — a tumblr of auto-generated pseudo-scripture (or pseudo-compsci lessons). Some examples: -- 'The LORD is the beginning (or prefix) of the code for the body of the procedure.' -- 'More precisely, if P and Q are polynomials, let O1 be the order of blessed.' -- ''In APL all data are represented as arrays, and there shall they see the Son of man, in whose sight I brought them out.'"
Nerval's Lobster writes "'Building on many lessons learned from spreadsheets, functional languages and model-driven application development environments, Reactive Programming (RP) is a recent approach to creating applications,' Val Huber, CTO of Espresso Logic, writes in a new column. 'In the RP paradigm, you can create complex applications from a series of simple declarative statements. Many of the details of implementation and work of gluing together various sorts of application constructs can be hidden because of the declarative approach.' He goes on to argue that RP makes maintenance easier and applications more adaptable, because RP offers automatic dependency management and reuse; at the same time, he also acknowledges that RP has some shortcomings. What say you?"
An anonymous reader writes "I have been trying to make the move to the Seattle, WA area. I liked the weather, the nature, the scenery and the tech environment. However, for whatever reason it seems like interviews are hard to come by. As a MS Stack software developer in LA, I barely had to do anything and recruiters always come knocking, either via LinkedIn or from past connections. Not to mention in general I got phone interviews for easily .8 of the positions I applied for. I wanted to finally make the move and fulfill a live long dream to live in Seattle. So I have been applying for positions in the greater Seattle, WA (King County) area. So far the ratio of positions applied to phone interviews is a dismal .1. Which is terrible considering the economy was much worse when I was actively looking for job in LA. Something isn't right because I am still getting offers for interviews here in SoCal, but not much from where I really wanted to be. What could I be doing wrong? Why such a contrast? Is the IT market in Seattle in poor shape? Or may be I just lack the proper connections in a new area? Am I just being screened out immediately for not being local? Or is it the prevalence of bigger corporations vs. smaller startups? And frankly as nice as the city is I can't move unless there's a healthy IT market to thrive by. I hope someone can point me in the right direction."
sl4shd0rk writes "In 2012, Oracle took Google to court over Java. In the balance hung the legalities of writing code to mimic the functionality of copyrighted software. The trial was set to determine how all future software would be written (and by whom). Oracle's entire case boiled down to an inadvertent 9 lines of code; an argument over a simple and basic comparison of a range of numbers. The presiding judge (who had some background in writing software) didn't buy it stating he had 'written blocks of code like rangeCheck a hundred times before.' A victory for more than just Google. This week, however, Microsoft, EMC, Oracle and Netapp have filed for appeal and seek to reverse the ruling. It's not looking good as the new bevy of judges Indicating they may side with Oracle on the issue."