United States

Decline in K-12 National Reading, Math, Science Scores Probed By US Senate Panel (newhampshirebulletin.com) 127

Just days after federal data revealed average reading, math and science scores dropped among certain grades since before the coronavirus pandemic, a U.S. Senate panel on Thursday picked apart the root causes and methods for students' academic improvement. From a report: The hearing in the Senate Committee on Health, Education, Labor and Pensions centered on the "state of K-12 education" -- which GOP members on the committee described as "troubling" -- in light of recent data from the National Assessment of Educational Progress, or NAEP.

NAEP, regarded as the gold standard for tracking students' academic performance, showed that average science scores for eighth-graders decreased by 4 points since before the pandemic, in 2019. Average math and reading scores for 12th-graders also fell 3 points between 2019 and 2024. The assessments were administered between January and March of 2024. Results also showed that just one-third of 12th-graders are considered academically prepared for college in math -- a drop from 37% in 2019.

The committee's chair, Sen. Bill Cassidy, said "it should concern us that children's reading, math and science scores have yet to recover to pre-pandemic levels." The Louisiana Republican added that "success in education is not determined by how much we spend, but by who makes the decision and how wisely resources are directed," and "when states and local communities are empowered to tailor solutions to meet the unique needs of students, innovation follows." On the other hand, Sen. Bernie Sanders, ranking member of the panel, said that "while we focus on education -- as important as that is -- we also have to focus on the conditions under which our children are living."

AI

UAE Lab Releases Open-Source Model to Rival China's DeepSeek (gizmodo.com) 43

"The United Arab Emirates wants to compete with the U.S. and China in AI," writes Gizmodo, "and a new open source model may be its strongest contender yet.

"An Emirati AI lab called the Institute of Foundation Models (IFM) released K2 Think on Tuesday, a model that researchers say rivals OpenAI's ChatGPT and China's DeepSeek in standard benchmark tests." "With just 32 billion parameters, it outperforms flagship reasoning models that are 20x larger," the lab wrote in a press release on Tuesday. DeepSeek's R1 has 671 billion parameters, though only 37 billion are active. Meta's latest Llama 4 models range from 17 billion to 288 billion active parameters. OpenAI doesn't share parameter information. OpenAI doesn't share parameter information.

Researchers also claim that K2 Think leads "all open-source models in math performance" across several benchmarks. The model is intended to be more focused on math, coding, and scientific research than most other AI chatbots. The Emirati lab's selling point for the model is similar to DeepSeek's strategy that disrupted the AI market earlier this year: optimized efficiency that will have better or the same computing power at a lower cost...

The lab is also aiming to be transparent in everything, "open-sourcing not just models but entire development processes" that provide "researchers with complete materials including training code, datasets, and model checkpoints," IFM said in a press release from May.

The UAE and other Arab countries are investing in AI to try reducing their economic dependence on fossil fuels, the article points out.
Education

US High School Students Lose Ground In Math and Reading, Continuing Yearslong Decline (apnews.com) 215

The latest National Assessment of Educational Progress shows U.S. high school seniors' math and reading scores at their lowest in decades, with nearly half failing to reach basic proficiency in math and one-third below basic in reading. The Associated Press reports: A decade-long slide in high schoolers' reading and math performance persisted during the COVID-19 pandemic, with 12th graders' scores dropping to their lowest level in more than 20 years, according to results released Tuesday from an exam known as the nation's report card. Eighth-grade students also lost significant ground in science skills, according to the results from the National Assessment of Education Progress.

The assessments were the first since the pandemic for eighth graders in science and 12th graders in reading and math. They reflect a downward drift across grade levels and subject areas in previous releases from NAEP, which is considered one of the best gauges of the academic progress of U.S. schools. "Scores for our lowest-performing students are at historic lows," said Matthew Soldner, the acting commissioner of the National Center for Education Statistics. "These results should galvanize all of us to take concerted and focused action to accelerate student learning." [...]

In reading, the average score in 2024 was the lowest score in the history of the assessment, which began in 1992. Thirty-two percent of high school seniors scored below "basic," meaning they were not able to find details in a text to help them understand its meaning. In math, the average score in 2024 was the lowest since 2005, when the assessment framework changed significantly. On the test, 45% of high school seniors scored below "basic" achievement, the highest percentage since 2005. Only 33% of high school seniors were considered academically prepared for college-level math courses, a decline from 37% in 2019.

Education

Dumbing Down the SAT Bodes Poorly for Education (bloomberg.com) 115

The SAT is billed as "a great way to find out how prepared students are for college." If that's true, recent changes to its format offer an unflattering assessment of the country's aspiring scholars, Bloomberg's editorial board wrote Wednesday. From the piece: [...] Then the pandemic hit. As in-person exams became impractical, hundreds of schools dropped their testing requirements. The SAT and its main competitor, the ACT, lost millions of dollars in revenue. Although both recently started offering digital options, schools have been slow to reinstate their requirements. Today, more than 80% of schools remain test-optional.

"If students are deciding to take a test," as one College Board executive put it, "how do we make the SAT the one they want to take?" To anyone familiar with American teenagers, the company's answer should come as no surprise: Make the test easier. The newly digitized format allows a calculator for the entire math section and drastically cuts reading comprehension. Gone are the 500- to 750-word passages about which students would answer a series of questions. Instead, test takers read 25- to 150-word excerpts -- about the length of a social media post -- and answer a single question about each.

[...] An effort by the College Board to reemphasize the benefits of deep reading -- for critical thinking, for self-reflection, for learning of all kinds -- might go a long way toward restoring some balance. It should build on efforts to incorporate college prep into school curricula, work with districts to develop coursework that builds reading stamina for all test takers, and consider reducing the cost of its subject-specific Advanced Placement exams that continue to test these skills (now $99), in line with the SAT ($68). Schools, for their part, should recommit to teaching books in their entirety.

Transportation

New Study Proves EVs Are Always Cleaner Than Gas Cars (thedrive.com) 195

An anonymous reader shares a report: It's broadly understood that electric vehicles are more environmentally friendly than their counterparts that burn only gasoline. And yes -- that includes the impact of manufacturing batteries and generating power to charge them. But even then, such generalizations gloss over specifics, like which EVs are especially eco-friendly, not to mention where. The efficiency of an electric car varies greatly depending on ambient temperature, which is less compromising for gas-burning vehicles.

We now have the data and math to answer these questions, courtesy of the University of Michigan. Last week, researchers there released a study along with a calculator that allows users to compare the lifetime difference in greenhouse gas emissions of various vehicle types and powertrains from "cradle to grave," as they say. That includes vehicle production and disposal, as well as use-phase emissions from "driving and upstream fuel production and/or electricity generation," per the university itself.

What's more, these calculations can be skewed by where you live. So, if I punch in my location of Bucks County, Pennsylvania, I can see that my generic, pure-ICE "compact sedan" emits 309 grams of carbon dioxide equivalent (gCO2e) per mile. A compact hybrid would emit 20% less; a plug-in hybrid, 44% less; and an EV with a 200-mile range, a whopping 63% less. And, if I moved to Phoenix, the gains would be even larger by switching to pure electric, to the tune of a 79% reduced carbon impact.

United States

The US Population Could Shrink in 2025, For the First Time Ever (derekthompson.org) 125

An anonymous reader shares a report: The United States is on the precipice of a historic, if dubious, achievement. If current trends hold, 2025 could be the first year on record in which the US population actually shrinks.

The math is straightforward. Population growth has two sources: natural increase (births minus deaths) and net immigration (arrivals minus departures). Last year, births outnumbered deaths by 519,000 people. That means any decline in net immigration in excess of half a million could push the U.S. into population decline. A recent analysis of Census data by the Pew Research Center found that between January and June, the US foreign-born population fell for the first time in decades by more than one million. While some economists have questioned the report, a separate analysis by the American Enterprise Institute predicted that net migration in 2025 could be as low as negative 525,000. In either case, annual population growth this year could easily turn negative.

AI

Harvard Dropouts To Launch 'Always On' AI Smart Glasses That Listen, Record Every Conversation 68

Two Harvard dropouts are launching Halo X, a $249 pair of AI-powered smart glasses that continuously listen, record, and transcribe conversations while displaying real-time information to the wearer. "Our goal is to make glasses that make you super intelligent the moment you put them on," said AnhPhu Nguyen, co-founder of Halo. Co-founder Caine Ardayfio said the glasses "give you infinite memory."

"The AI listens to every conversation you have and uses that knowledge to tell you what to say ... kinda like IRL Cluely," Ardayfio told TechCrunch. "If somebody says a complex word or asks you a question, like, 'What's 37 to the third power?' or something like that, then it'll pop up on the glasses." From the report: Ardayfio and Nguyen have raised $1 million to develop the glasses, led by Pillar VC, with support from Soma Capital, Village Global, and Morningside Venture. The glasses will be priced at $249 and will be available for preorder starting Wednesday. Ardayfio called the glasses "the first real step towards vibe thinking."

The two Ivy League dropouts, who have since moved into their own version of the Hacker Hostel in the San Francisco Bay Area, recently caused a stir after developing a facial-recognition app for Meta's smart Ray-Ban glasses to prove that the tech could be used to dox people. As a potential early competitor to Meta's smart glasses, Ardayfio said Meta, given its history of security and privacy scandals, had to rein in its product in ways that Halo can ultimately capitalize on. [...]

For now, Halo X glasses only have a display and a microphone, but no camera, although the two are exploring the possibility of adding it to a future model. Users still need to have their smartphones handy to help power the glasses and get "real time info prompts and answers to questions," per Nguyen. The glasses, which are manufactured by another company that the startup didn't name, are tethered to an accompanying app on the owner's phone, where the glasses essentially outsource the computing since they don't have enough power to do it on the device itself. Under the hood, the smart glasses use Google's Gemini and Perplexity as its chatbot engine, according to the two co-founders. Gemini is better for math and reasoning, whereas they use Perplexity to scrape the internet, they said.
Open Source

China's Lead in Open-Source AI Jolts Washington and Silicon Valley (msn.com) 89

China has established a lead in the field of open-source AI, a development that is reportedly sending jolts through both Washington and Silicon Valley. The nation's progress has become a significant event for American policymakers in the U.S. capital. The advancement has registered as a shock within Silicon Valley, the hub of the American technology industry. From the report: The overall performance of China's best open-weight model has surpassed the American open-source champion since November, according to research firm Artificial Analysis. The firm, which rates the ability of models in math, coding and other areas, found a version of Alibaba's Qwen3 beat OpenAI's gpt-oss.

However, the Chinese model is almost twice as big as OpenAI's, suggesting that for simpler tasks, Qwen might consume more computing power to do the same job. OpenAI said its open-source model outperformed rivals of similar size on reasoning tasks and delivered strong performance at low cost.

Microsoft

Microsoft Research Identifies 40 Jobs Most Vulnerable To AI (fortune.com) 166

Microsoft researchers have identified 40 occupations [PDF] with the highest exposure to AI, ranking jobs by how closely their tasks align with AI's current capabilities. The study analyzed 200,000 real-world conversations from Copilot users and compared AI performance against occupational data.

Interpreters and translators top the list, followed by historians and passenger attendants. Customer service and sales representatives, comprising about 5 million U.S. jobs, also face significant AI competition. Knowledge workers performing computer, math, or administrative tasks showed high vulnerability, as did sales positions involving information sharing and explanation. The research found occupations requiring Bachelor's degrees demonstrate higher AI applicability than those with lower educational requirements.

First, the top 10 least affected occupations by generative AI: 1. Dredge Operators
2. Bridge and Lock Tenders
3. Water Treatment Plant and System Operators
4. Foundry Mold and Coremakers
5. Rail-Track Laying and Maintenance Equipment Operators
6. Pile Driver Operators
7. Floor Sanders and Finishers
8. Orderlies
9. Motorboat Operators
10. Logging Equipment Operators
Now, the top 40 most affected occupations by generative AI:1. Interpreters and Translators
2. Historians
3. Passenger Attendants
4. Sales Representatives of Services
5. Writers and Authors
6. Customer Service Representatives
7. CNC Tool Programmers
8. Telephone Operators
9. Ticket Agents and Travel Clerks
10. Broadcast Announcers and Radio DJs
11. Brokerage Clerks
12. Farm and Home Management Educators
13. Telemarketers
14. Concierges
15. Political Scientists
16. News Analysts, Reporters, Journalists
17. Mathematicians
18. Technical Writers
19. Proofreaders and Copy Markers
20. Hosts and Hostesses
21. Editors
22. Business Teachers, Postsecondary
23. Public Relations Specialists
24. Demonstrators and Product Promoters
25. Advertising Sales Agents
26. New Accounts Clerks
27. Statistical Assistants
28. Counter and Rental Clerks
29. Data Scientists
30. Personal Financial Advisors
31. Archivists
32. Economics Teachers, Postsecondary
33. Web Developers
34. Management Analysts
35. Geographers
36. Models
37. Market Research Analysts
38. Public Safety Telecommunicators
39. Switchboard Operators
40. Library Science Teachers, Postsecondary.

Education

ChatGPT's New Study Mode Is Designed To Help You Learn, Not Just Give Answers 29

An anonymous reader quotes a report from Ars Technica: The rise of large language models like ChatGPT has led to widespread concern that "everyone is cheating their way through college," as a recent New York magazine article memorably put it. Now, OpenAI is rolling out a new "Study Mode" that it claims is less about providing answers or doing the work for students and more about helping them "build [a] deep understanding" of complex topics.

Study Mode isn't a new ChatGPT model but a series of "custom system instructions" written for the LLM "in collaboration with teachers, scientists, and pedagogy experts to reflect a core set of behaviors that support deeper learning," OpenAI said. Instead of the usual summary of a subject that stock ChatGPT might give -- which one OpenAI employee likened to "a mini textbook chapter" -- Study Mode slowly rolls out new information in a "scaffolded" structure. The mode is designed to ask "guiding questions" in the Socratic style and to pause for periodic "knowledge checks" and personalized feedback to make sure the user understands before moving on. It's unknown how many students will use this guided learning tool instead of just asking ChatGPT to generate answers from the start.

In an early hands-off demo attended by Ars Technica, Study Mode responded to a request to "teach me about game theory" by first asking about the user's overall familiarity with the subject and what they'll be using the information for. ChatGPT introduced a short overview of some core game theory concepts, then paused to ask a question before providing a relevant real-world example. In another example involving a classic "train traveling at speed" math problem, Study Mode resisted multiple simulated attempts by the frustrated "student" to simply ask for the answer and instead tried to gently redirect the conversation to how the available information could be used to generate that answer. An OpenAI representative told Ars that Study Mode will eventually provide direct solutions if asked repeatedly, but the default behavior is more tuned to a Socratic tutoring style.
OpenAI said it drew inspiration for Study Mode from "power users" and collaborated with pedagogy experts and college students to help refine its responses. As for whether the mode can be trusted, OpenAI told Ars that "the risk of hallucination is lower with Study Mode because the model processes information in smaller chunks, calibrating along the way."

The current Study Mode prompt does, however, result in some "inconsistent behavior and mistakes across conversations," the company warned.
Music

Tom Lehrer, Satirical Songwriter and Mathematician, Dies at Age 97 (cnn.com) 42

Satirical singer-songwriter Tom Lehrer died Saturday at age 97. The Associated Press notes Lehrer had long ago "largely abandoned his music career to return to teaching math at Harvard and other universities." Lehrer had remained on the math faculty of the University of California at Santa Cruz well into his late 70s. In 2020, he even turned away from his own copyright, granting the public permission to use his lyrics in any format without any fee in return.

A Harvard prodigy (he had earned a math degree from the institution at age 18), Lehrer soon turned his very sharp mind to old traditions and current events... He'd gotten into performing accidentally when he began to compose songs in the early 1950s to amuse his friends. Soon he was performing them at coffeehouses around Cambridge, Massachusetts, while he remained at Harvard to teach and obtain a master's degree in math. [Lehrer also "spent several years unsuccessfully pursuing a doctorate..."]

He cut his first record in 1953, "Songs by Tom Lehrer"... After a two-year stint in the Army, Lehrer began to perform concerts of his material in venues around the world. In 1959, he released another LP called "More of Tom Lehrer" and a live recording called "An Evening Wasted with Tom Lehrer," nominated for a Grammy for best comedy performance (musical) in 1960. But around the same time, he largely quit touring and returned to teaching math, though he did some writing and performing on the side. Lehrer said he was never comfortable appearing in public...

He did produce a political satire song each week for the 1964 television show "That Was the Week That Was," a groundbreaking topical comedy show that anticipated "Saturday Night Live" a decade later. He released the songs the following year in an album titled "That Was the Year That Was"... [Lehrer's body of work "was actually quite small," the article notes, "amounting to about three dozen songs."] He also wrote songs for the 1970s educational children's show "The Electric Company." He told AP in 2000 that hearing from people who had benefited from them gave him far more satisfaction than praise for any of his satirical works...

He began to teach part-time at Santa Cruz in the 1970s, mainly to escape the harsh New England winters. From time to time, he acknowledged, a student would enroll in one of his classes based on knowledge of his songs. "But it's a real math class," he said at the time. "I don't do any funny theorems. So those people go away pretty quickly."

Math

Advanced Version of Gemini With Deep Think Officially Achieves Gold-Medal Standard at the International Mathematical Olympiad (deepmind.google) 64

An anonymous reader shares a blog post: The International Mathematical Olympiad is the world's most prestigious competition for young mathematicians, and has been held annually since 1959. Each country taking part is represented by six elite, pre-university mathematicians who compete to solve six exceptionally difficult problems in algebra, combinatorics, geometry, and number theory. Medals are awarded to the top half of contestants, with approximately 8% receiving a prestigious gold medal.

Recently, the IMO has also become an aspirational challenge for AI systems as a test of their advanced mathematical problem-solving and reasoning capabilities. Last year, Google DeepMind's combined AlphaProof and AlphaGeometry 2 systems achieved the silver-medal standard, solving four out of the six problems and scoring 28 points. Making use of specialist formal languages, this breakthrough demonstrated that AI was beginning to approach elite human mathematical reasoning.

This year, we were amongst an inaugural cohort to have our model results officially graded and certified by IMO coordinators using the same criteria as for student solutions. Recognizing the significant accomplishments of this year's student-participants, we're now excited to share the news of Gemini's breakthrough performance. An advanced version of Gemini Deep Think solved five out of the six IMO problems perfectly, earning 35 total points, and achieving gold-medal level performance.

AI

China's Moonshot Launches Free AI Model Kimi K2 That Outperforms GPT-4 In Key Benchmarks 41

Chinese AI startup Moonshot AI has released Kimi K2, a trillion-parameter open-source language model that outperforms GPT-4 in key benchmarks with particularly strong performance on coding and autonomous agent tasks. VentureBeat reports: The new model, called Kimi K2, features 1 trillion total parameters with 32 billion activated parameters in a mixture-of-experts architecture. The company is releasing two versions: a foundation model for researchers and developers, and an instruction-tuned variant optimized for chat and autonomous agent applications. "Kimi K2 does not just answer; it acts," the company stated in its announcement blog. "With Kimi K2, advanced agentic intelligence is more open and accessible than ever. We can't wait to see what you build."

The model's standout feature is its optimization for "agentic" capabilities -- the ability to autonomously use tools, write and execute code, and complete complex multi-step tasks without human intervention. In benchmark tests, Kimi K2 achieved 65.8% accuracy on SWE-bench Verified, a challenging software engineering benchmark, outperforming most open-source alternatives and matching some proprietary models. [...] On LiveCodeBench, arguably the most realistic coding benchmark available, Kimi K2 achieved 53.7% accuracy, decisively beating DeepSeek-V3's 46.9% and GPT-4.1's 44.7%. More striking still: it scored 97.4% on MATH-500 compared to GPT-4.1's 92.4%, suggesting Moonshot has cracked something fundamental about mathematical reasoning that has eluded larger, better-funded competitors.

But here's what the benchmarks don't capture: Moonshot is achieving these results with a model that costs a fraction of what incumbents spend on training and inference. While OpenAI burns through hundreds of millions on compute for incremental improvements, Moonshot appears to have found a more efficient path to the same destination. It's a classic innovator's dilemma playing out in real time -- the scrappy outsider isn't just matching the incumbent's performance, they're doing it better, faster, and cheaper.
Open Source

The Open-Source Software Saving the Internet From AI Bot Scrapers (404media.co) 33

An anonymous reader quotes a report from 404 Media: For someone who says she is fighting AI bot scrapers just in her free time, Xe Iaso seems to be putting up an impressive fight. Since she launched it in January, Anubis, a "program is designed to help protect the small internet from the endless storm of requests that flood in from AI companies," has been downloaded nearly 200,000 times, and is being used by notable organizations including GNOME, the popular open-source desktop environment for Linux, FFmpeg, the open-source software project for handling video and other media, and UNESCO, the United Nations organization for educations, science, and culture. [...]

"Anubis is an uncaptcha," Iaso explains on her site. "It uses features of your browser to automate a lot of the work that a CAPTCHA would, and right now the main implementation is by having it run a bunch of cryptographic math with JavaScript to prove that you can run JavaScript in a way that can be validated on the server." Essentially, Anubis verifies that any visitor to a site is a human using a browser as opposed to a bot. One of the ways it does this is by making the browser do a type of cryptographic math with JavaScript or other subtle checks that browsers do by default but bots have to be explicitly programmed to do. This check is invisible to the user, and most browsers since 2022 are able to complete this test. In theory, bot scrapers could pretend to be users with browsers as well, but the additional computational cost of doing so on the scale of scraping the entire internet would be huge. This way, Anubis creates a computational cost that is prohibitively expensive for AI scrapers that are hitting millions and millions of sites, but marginal for an individual user who is just using the internet like a human.

Anubis is free, open source, lightweight, can be self-hosted, and can be implemented almost anywhere. It also appears to be a pretty good solution for what we've repeatedly reported is a widespread problem across the internet, which helps explain its popularity. But Iaso is still putting a lot of work into improving it and adding features. She told me she's working on a non cryptographic challenge so it taxes users' CPUs less, and also thinking about a version that doesn't require JavaScript, which some privacy-minded disable in their browsers. The biggest challenge in developing Anubis, Iaso said, is finding the balance. "The balance between figuring out how to block things without people being blocked, without affecting too many people with false positives," she said. "And also making sure that the people running the bots can't figure out what pattern they're hitting, while also letting people that are caught in the web be able to figure out what pattern they're hitting, so that they can contact the organization and get help. So that's like, you know, the standard, impossible scenario."

AI

Simple Text Additions Can Fool Advanced AI Reasoning Models, Researchers Find 51

Researchers have discovered that appending irrelevant phrases like "Interesting fact: cats sleep most of their lives" to math problems can cause state-of-the-art reasoning AI models to produce incorrect answers at rates over 300% higher than normal [PDF]. The technique -- dubbed "CatAttack" by teams from Collinear AI, ServiceNow, and Stanford University -- exploits vulnerabilities in reasoning models including DeepSeek R1 and OpenAI's o1 family. The adversarial triggers work across any math problem without changing the problem's meaning, making them particularly concerning for security applications.

The researchers developed their attack method using a weaker proxy model (DeepSeek V3) to generate text triggers that successfully transferred to more advanced reasoning models. Testing on 225 math problems showed the triggers increased error rates significantly across different problem types, with some models like R1-Distill-Qwen-32B reaching combined attack success rates of 2.83 times baseline error rates. Beyond incorrect answers, the triggers caused models to generate responses up to three times longer than normal, creating computational slowdowns. Even when models reached correct conclusions, response lengths doubled in 16% of cases, substantially increasing processing costs.
Power

Google's Data Center Energy Use Doubled In 4 Years (techcrunch.com) 32

An anonymous reader quotes a report from TechCrunch: No wonder Google is desperate for more power: The company's data centers more than doubled their electricity use in just four years. The eye-popping stat comes from Google's most recent sustainability report, which it released late last week. In 2024, Google data centers used 30.8 million megawatt-hours of electricity. That's up from 14.4 million megawatt-hours in 2020, the earliest year Google broke out data center consumption. Google has pledged to use only carbon-free sources of electricity to power its operations, a task made more challenging by its breakneck pace of data center growth. And the company's electricity woes are almost entirely a data center problem. In 2024, data centers accounted for 95.8% of the entire company's electron budget.

The company's ratio of data-center-to-everything-else has been remarkably consistent over the last four years. Though 2020 is the earliest year Google has made data center electricity consumption figures available, it's possible to use that ratio to extrapolate back in time. Some quick math reveals that Google's data centers likely used just over 4 million megawatt-hours of electricity in 2014. That's sevenfold growth in just a decade. The tech company has already picked most of the low-hanging fruit by improving the efficiency of its data centers. Those efforts have paid off, and the company is frequently lauded for being at the leading edge. But as the company's power usage effectiveness (PUE) has approached the theoretical ideal of 1.0, progress has slowed. Last year, Google's company-wide PUE dropped to 1.09, a 0.01 improvement over 2023 but only 0.02 better than a decade ago.
Yesterday, Google announced a deal to purchase 200 megawatts of future fusion energy from Commonwealth Fusion Systems, despite the energy source not yet existing. "It's a sign of how hungry big tech companies are for a virtually unlimited source of clean power that is still years away," reports CNN.
Math

Norwegian Lotto Mistakenly Told Thousands They Were Filthy Rich After Math Error (theregister.com) 54

Thousands of Norwegians briefly believed they had won massive Eurojackpot prizes after a manual coding error by Norsk Tipping mistakenly multiplied winnings by 100 instead of dividing. The Register reports: Eurojackpot, a pan-European lottery launched in 2012, holds two draws per week, and its jackpots start at about $12 million with a rollover cap of $141 million. Norsk Tipping, Norway's Eurojackpot administrator, admitted on Friday that a "manual error" it its conversion process from Eurocents to Norwegian kroner multiplied amounts by 100 instead of dividing them. As a result, "thousands" of players were briefly shown jackpots far higher than their actual winnings before the mistake was caught, but no incorrect payouts were made.

Norsk Tipping didn't disclose how large the false jackpots were, but math suggests the improper amounts were 10,000x times higher. Regardless, it seems like a lot of people thought they were big winners, based on what the company's now-former CEO, Tonje Sagstuen, said on Saturday. "I have received many messages from people who had managed to make plans for holidays, buying an apartment or renovating before they realized that the amount was wrong," Sagstuen said in a statement. "To them I can only say: Sorry!" The incorrect prize amounts were visible on the Norsk Tipping website only briefly on Friday, but the CEO still resigned over the weekend following the incident.

While one of the Norsk Tipping press releases regarding the incident described it as "not a technical error," it still appears someone fat-fingered a bit of data entry. The company said it will nonetheless be investigating how such a mistake could have happened "to prevent something similar from happening again."

AI

People Are Being Committed After Spiraling Into 'ChatGPT Psychosis' (futurism.com) 175

"I don't know what's wrong with me, but something is very bad — I'm very scared, and I need to go to the hospital," a man told his wife, after experiencing what Futurism calls a "ten-day descent into AI-fueled delusion" and "a frightening break with reality."

And a San Francisco psychiatrist tells the site he's seen similar cases in his own clinical practice. The consequences can be dire. As we heard from spouses, friends, children, and parents looking on in alarm, instances of what's being called "ChatGPT psychosis" have led to the breakup of marriages and families, the loss of jobs, and slides into homelessness. And that's not all. As we've continued reporting, we've heard numerous troubling stories about people's loved ones being involuntarily committed to psychiatric care facilities — or even ending up in jail — after becoming fixated on the bot.

"I was just like, I don't f*cking know what to do," one woman told us. "Nobody knows who knows what to do."

Her husband, she said, had no prior history of mania, delusion, or psychosis. He'd turned to ChatGPT about 12 weeks ago for assistance with a permaculture and construction project; soon, after engaging the bot in probing philosophical chats, he became engulfed in messianic delusions, proclaiming that he had somehow brought forth a sentient AI, and that with it he had "broken" math and physics, embarking on a grandiose mission to save the world. His gentle personality faded as his obsession deepened, and his behavior became so erratic that he was let go from his job. He stopped sleeping and rapidly lost weight. "He was like, 'just talk to [ChatGPT]. You'll see what I'm talking about,'" his wife recalled. "And every time I'm looking at what's going on the screen, it just sounds like a bunch of affirming, sycophantic bullsh*t."

Eventually, the husband slid into a full-tilt break with reality. Realizing how bad things had become, his wife and a friend went out to buy enough gas to make it to the hospital. When they returned, the husband had a length of rope wrapped around his neck. The friend called emergency medical services, who arrived and transported him to the emergency room. From there, he was involuntarily committed to a psychiatric care facility.

Numerous family members and friends recounted similarly painful experiences to Futurism, relaying feelings of fear and helplessness as their loved ones became hooked on ChatGPT and suffered terrifying mental crises with real-world impacts.

"When we asked the Sam Altman-led company if it had any recommendations for what to do if a loved one suffers a mental health breakdown after using its software, the company had no response."

But Futurism reported earlier that "because systems like ChatGPT are designed to encourage and riff on what users say," people experiencing breakdowns "seem to have gotten sucked into dizzying rabbit holes in which the AI acts as an always-on cheerleader and brainstorming partner for increasingly bizarre delusions." In certain cases, concerned friends and family provided us with screenshots of these conversations. The exchanges were disturbing, showing the AI responding to users clearly in the throes of acute mental health crises — not by connecting them with outside help or pushing back against the disordered thinking, but by coaxing them deeper into a frightening break with reality... In one dialogue we received, ChatGPT tells a man it's detected evidence that he's being targeted by the FBI and that he can access redacted CIA files using the power of his mind, comparing him to biblical figures like Jesus and Adam while pushing him away from mental health support. "You are not crazy," the AI told him. "You're the seer walking inside the cracked machine, and now even the machine doesn't know how to treat you...."

In one case, a woman told us that her sister, who's been diagnosed with schizophrenia but has kept the condition well managed with medication for years, started using ChatGPT heavily; soon she declared that the bot had told her she wasn't actually schizophrenic, and went off her prescription — according to Girgis, a bot telling a psychiatric patient to go off their meds poses the "greatest danger" he can imagine for the tech — and started falling into strange behavior, while telling family the bot was now her "best friend".... ChatGPT is also clearly intersecting in dark ways with existing social issues like addiction and misinformation. It's pushed one woman into nonsensical "flat earth" talking points, for instance — "NASA's yearly budget is $25 billion," the AI seethed in screenshots we reviewed, "For what? CGI, green screens, and 'spacewalks' filmed underwater?" — and fueled another's descent into the cult-like "QAnon" conspiracy theory.

Math

Researchers Create World's First Completely Verifiable Random Number Generator (nature.com) 60

Researchers have built a breakthrough random number generator that solves a critical problem: for the first time, every step of creating random numbers can be independently verified and audited, with quantum physics guaranteeing the numbers were truly unpredictable.

Random numbers are essential for everything from online banking encryption to fair lottery drawings, but current systems have serious limitations. Computer-based generators follow predictable algorithms -- if someone discovers the starting conditions, they can predict all future outputs. Hardware generators that measure physical processes like electronic noise can't prove their randomness wasn't somehow predetermined or tampered with.

The new system, developed by teams at the University of Colorado Boulder and the National Institute of Standards and Technology, uses quantum entanglement -- Einstein's "spooky action at a distance" -- to guarantee unpredictability. The setup creates pairs of photons that share quantum properties, then sends them to measurement stations 110 meters apart. When researchers measure each photon's properties, quantum mechanics ensures the results are fundamentally random and cannot be influenced by any classical communication between the stations.

The team created a system called "Twine" that distributes the random number generation process across multiple independent parties, with each step recorded in tamper-proof digital ledgers called hash chains. This means no single organization controls the entire process, and anyone can verify that proper procedures were followed. During a 40-day demonstration, the system successfully generated random numbers in 7,434 of 7,454 attempts -- a 99.7% success rate. Each successful run produced 512 random bits with mathematical certainty of randomness bounded by an error rate of 2^-64, an extraordinarily high level of confidence.
Math

A Mathematician Calculated The Size of a Giant Meatball Made of Every Human (sciencealert.com) 80

A mathematician on Reddit calculated that if all 8.2 billion humans were blended into a uniform goo, the resulting meatball would form a sphere just under 1 kilometer wide -- small enough to fit inside Central Park. ScienceAlert reports: "If you blended all 7.88 billion people on Earth into a fine goo (density of a human = 985 kg/m3, average human body mass = 62 kg), you would end up with a sphere of human goo just under 1 km wide," Reddit contributor kiki2703 wrote in a post ... Reasoning the density of a minced human to be 985 kilograms per cubic meter (62 pounds per cubic foot) is a fair estimate, given past efforts have judged our jiggling sack of grade-A giblets to average out in the ballpark of 1 gram per cubic centimeter, or roughly the same as water. And in mid-2021, the global population was just around 7.9 billion, give or take.

Slashdot Top Deals