Connect with us

Funding & Investment in Travel

A new paradigm for AI: How ‘thinking as optimization’ leads to better general-purpose models

Published

on


Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now


Researchers at the University of Illinois Urbana-Champaign and the University of Virginia have developed a new model architecture that could lead to more robust AI systems with more powerful reasoning capabilities. 

Called an energy-based transformer (EBT), the architecture shows a natural ability to use inference-time scaling to solve complex problems. For the enterprise, this could translate into cost-effective AI applications that can generalize to novel situations without the need for specialized fine-tuned models.

The challenge of System 2 thinking

In psychology, human thought is often divided into two modes: System 1, which is fast and intuitive, and System 2, which is slow, deliberate and analytical. Current large language models (LLMs) excel at System 1-style tasks, but the AI industry is increasingly focused on enabling System 2 thinking to tackle more complex reasoning challenges.

Reasoning models use various inference-time scaling techniques to improve their performance on difficult problems. One popular method is reinforcement learning (RL), used in models like DeepSeek-R1 and OpenAI’s “o-series” models, where the AI is rewarded for producing reasoning tokens until it reaches the correct answer. Another approach, often called best-of-n, involves generating multiple potential answers and using a verification mechanism to select the best one. 

However, these methods have significant drawbacks. They are often limited to a narrow range of easily verifiable problems, like math and coding, and can degrade performance on other tasks such as creative writing. Furthermore, recent evidence suggests that RL-based approaches might not be teaching models new reasoning skills, instead just making them more likely to use successful reasoning patterns they already know. This limits their ability to solve problems that require true exploration and are beyond their training regime.

Energy-based models (EBM)

The architecture proposes a different approach based on a class of models known as energy-based models (EBMs). The core idea is simple: Instead of directly generating an answer, the model learns an “energy function” that acts as a verifier. This function takes an input (like a prompt) and a candidate prediction and assigns a value, or “energy,” to it. A low energy score indicates high compatibility, meaning the prediction is a good fit for the input, while a high energy score signifies a poor match.

Applying this to AI reasoning, the researchers propose in a paper that devs should view “thinking as an optimization procedure with respect to a learned verifier, which evaluates the compatibility (unnormalized probability) between an input and candidate prediction.” The process begins with a random prediction, which is then progressively refined by minimizing its energy score and exploring the space of possible solutions until it converges on a highly compatible answer. This approach is built on the principle that verifying a solution is often much easier than generating one from scratch.

This “verifier-centric” design addresses three key challenges in AI reasoning. First, it allows for dynamic compute allocation, meaning models can “think” for longer on harder problems and shorter on easy problems. Second, EBMs can naturally handle the uncertainty of real-world problems where there isn’t one clear answer. Third, they act as their own verifiers, eliminating the need for external models.

Unlike other systems that use separate generators and verifiers, EBMs combine both into a single, unified model. A key advantage of this arrangement is better generalization. Because verifying a solution on new, out-of-distribution (OOD) data is often easier than generating a correct answer, EBMs can better handle unfamiliar scenarios.

Despite their promise, EBMs have historically struggled with scalability. To solve this, the researchers introduce EBTs, which are specialized transformer models designed for this paradigm. EBTs are trained to first verify the compatibility between a context and a prediction, then refine predictions until they find the lowest-energy (most compatible) output. This process effectively simulates a thinking process for every prediction. The researchers developed two EBT variants: A decoder-only model inspired by the GPT architecture, and a bidirectional model similar to BERT.

Energy-based transformer (source: GitHub)

The architecture of EBTs make them flexible and compatible with various inference-time scaling techniques. “EBTs can generate longer CoTs, self-verify, do best-of-N [or] you can sample from many EBTs,” Alexi Gladstone, a PhD student in computer science at the University of Illinois Urbana-Champaign and lead author of the paper, told VentureBeat. “The best part is, all of these capabilities are learned during pretraining.”

EBTs in action

The researchers compared EBTs against established architectures: the popular transformer++ recipe for text generation (discrete modalities) and the diffusion transformer (DiT) for tasks like video prediction and image denoising (continuous modalities). They evaluated the models on two main criteria: “Learning scalability,” or how efficiently they train, and “thinking scalability,” which measures how performance improves with more computation at inference time.

During pretraining, EBTs demonstrated superior efficiency, achieving an up to 35% higher scaling rate than Transformer++ across data, batch size, parameters and compute. This means EBTs can be trained faster and more cheaply. 

At inference, EBTs also outperformed existing models on reasoning tasks. By “thinking longer” (using more optimization steps) and performing “self-verification” (generating multiple candidates and choosing the one with the lowest energy), EBTs improved language modeling performance by 29% more than Transformer++. “This aligns with our claims that because traditional feed-forward transformers cannot dynamically allocate additional computation for each prediction being made, they are unable to improve performance for each token by thinking for longer,” the researchers write.

For image denoising, EBTs achieved better results than DiTs while using 99% fewer forward passes. 

Crucially, the study found that EBTs generalize better than the other architectures. Even with the same or worse pretraining performance, EBTs outperformed existing models on downstream tasks. The performance gains from System 2 thinking were most substantial on data that was further out-of-distribution (different from the training data), suggesting that EBTs are particularly robust when faced with novel and challenging tasks.

The researchers suggest that “the benefits of EBTs’ thinking are not uniform across all data but scale positively with the magnitude of distributional shifts, highlighting thinking as a critical mechanism for robust generalization beyond training distributions.”

The benefits of EBTs are important for two reasons. First, they suggest that at the massive scale of today’s foundation models, EBTs could significantly outperform the classic transformer architecture used in LLMs. The authors note that “at the scale of modern foundation models trained on 1,000X more data with models 1,000X larger, we expect the pretraining performance of EBTs to be significantly better than that of the Transformer++ recipe.”

Second, EBTs show much better data efficiency. This is a critical advantage in an era where high-quality training data is becoming a major bottleneck for scaling AI. “As data has become one of the major limiting factors in further scaling, this makes EBTs especially appealing,” the paper concludes. 

Despite its different inference mechanism, the EBT architecture is highly compatible with the transformer, making it possible to use them as a drop-in replacement for current LLMs. 

“EBTs are very compatible with current hardware/inference frameworks,” Gladstone said, including speculative decoding using feed-forward models on both GPUs or TPUs. He said he is also confident they can run on specialized accelerators such as LPUs and optimization algorithms such as FlashAttention-3, or can be deployed through common inference frameworks like vLLM.

For developers and enterprises, the strong reasoning and generalization capabilities of EBTs could make them a powerful and reliable foundation for building the next generation of AI applications. “Thinking longer can broadly help on almost all enterprise applications, but I think the most exciting will be those requiring more important decisions, safety or applications with limited data,” Gladstone said.



Source link
Continue Reading
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Funding & Investment in Travel

North Korea’s ‘Benidorm’ resort bans foreign visitors – despite bid to bring tourists | World | News

Published

on


International visitors have been banned from North Korea‘s massive new beach resort following its grand opening. The Wonsan Kalma complex, unveiled by leader Kim Jong-un at the end of June and dubbed the North Korean Benidorm, boasts a capacity for nearly 20,000 guests and includes accommodation, a shoreline, sporting venues, and restaurants.

Kim declared it would be remembered as “one of the greatest successes this year” and hailed the location as “the proud first step” towards advancing tourism. However, only North Koreans can experience the facilities. DPR Korea Tour, a platform operated by the nation’s tourism officials, announced that the resort “is temporarily not receiving foreign tourists”.

No additional information was provided regarding the reasons behind the ban or how long it would last. Shortly after its launch, a limited number of Russians were the only foreign tourists to visit.

North Korea may have stopped international visitor access after a Russian journalist wrote a damning story about the Wonsan Kalma resort.

Accompanied the Russian foreign minister, the journalist suggested the people at the resort were government operatives rather than genuine guests.

Kim has been pushing to make North Korea a tourist destination as part of efforts to revive the isolated country’s struggling economy.

Wonsan Kalma, with a 2.5 mile beach, is one of Kim’s most-discussed tourism projects, and state media reported North Korea will also confirm plans to build large tourism areas in other parts of the country.

Photos shared by state media show the leader taking in the views and watching someone go down a slide.

Despite the dangers Westerners may face if allowed to visit, Brits have voiced their desire to go.

Holiday planners On The Beach opened a link for people to express their interest, which racked up more than 250 sign-ups from Brits within a month.



Source link

Continue Reading

Funding & Investment in Travel

Amazon has a dual voltage hot air brush on sale for 15% off

Published

on


If you’ve ever landed in a dreamy European Airbnb only to realize your blow dryer won’t work (or worse, fries itself in the outlet — #ToAlltheDysonsWeveLostBefore), you know the struggle is real.

Voltage mismatches are the silent saboteurs of great hair days abroad, so if you want to avoid them at all costs, a versatile blowout brush that adapts to both U.S. and international power standards is a great tool to have.

Amazon has the Aima Beauty Hot Air Brush on sale for 15% off with a clickable coupon, and if you’re a Prime member you can get it as early as tomorrow with free overnight shipping.

Aima Beauty Worldwide Travel Hair Dryer Brush

$42 NOW FOR $36

Aima Beauty Worldwide Travel Hair Dryer Brush

Apply the clickable 15% off coupon by checking the box next to the orange “Coupon” banner beneath the regular retail price.

Buy it now at Amazon

The lightweight, oval-shaped hot air brush is designed to do it all: heat up fast, dry, straighten, volumize, and smooth, so you don’t have to pack multiple tools or gamble on the hotel hair dryer.

According to Aima and dozens of happy Amazon customers, the compact, 10-inch x 3.4-inch brush — which weighs less than a pound — is said to fit easily into a suitcase or backpack without hogging space or adding bulk.

It features 360-degree airflow and negative ion technology, which the company says helps reduce frizz, boosts shine, and protects against heat damage — all of which are ideal for bouncing back from long flights or humid climates.

The nylon pins and tufted bristles are built for detangling and gripping hair without tugging, and the ergonomic handle contributes to more effortless styling, even in a cramped hotel bathroom.

With two heat and speed settings, you can tailor the airflow to your hair type — lower temp and slower speed for fine strands, higher temp and faster speed for thicker textures. And thanks to the 360-degree swivel cord, you won’t get tangled up mid-style.

But the real flex with this dryer is its dual voltage, which means you can adjust its voltage from 110-240V (110-120V for the U.S., Canada, Japan, etc. 220-240V for the E.U., U.K., Australia, etc.) with a simple screwdriver turn. So, whether you’re in Paris, Tokyo, or a remote Greek island, you can plug in with confidence.

Aima even includes a free European plug adapter in the box. Depending on your destination, you may need to purchase other adapters separately.

If you’ve ever been burned by hair drying across the pond, literally or figuratively, needed to buy a last-minute hair dryer abroad (and then leave it behind), or gone days without styling because your tools just didn’t work, this brush is your redemption arc.

Grab it with the 15% off coupon and saves yourself time, money, and suitcase space.

Other travel deals to shop at Amazon right now

If you purchase a product or register for an account through a link on our site, we may receive compensation. By using this site, you consent to our User Agreement and agree that your clicks, interactions, and personal information may be collected, recorded, and/or stored by us and social media and other third-party partners in accordance with our Privacy Policy.



Source link

Continue Reading

Funding & Investment in Travel

Dozens dead after tourist boat capsizes in Vietnam

Published

on


Rescuers search for missing people after tourist boat capsizes in Vietnam

At least 34 people have died and several are still missing after a tourist boat capsized in Vietnam during bad weather.

The incident took place in Ha Long Bay, a popular tourist destination in the north of the country.

Most of the passengers were reportedly Vietnamese families visiting from the capital Hanoi.

Heavy rain has been hindering the search for survivors, rescuers say, but so far 11 people have been pulled from the water alive.

The vessel, named Wonder Seas, was carrying 53 people when it capsized after encountering a sudden storm, a statement from the Vietnamese Border Guards and navy said.

An eyewitness told AFP news agency that the sky darkened around 14:00 local time on Saturday (07:00 GMT).

There were “hailstones as big as toes with torrential rain, thunderstorm and lightning”, he said.

Getty Images Emergency services stand on top of a capsized tourist boat in Ha Long Bay, Quang Ninh province, Vietnam. Getty Images

At least one passenger was able to escape from an air pocket in the upturned vessel

A 10-year-old boy was rescued after being trapped in an air pocket in the upturned hull, local media say.

“I took a deep breath… dived, then swam up. I even shouted for help, then I was pulled up by a boat”, the boy – who had been travelling with his parents – told state media outlet VietnamNet.

Of the bodies so far recovered, at least eight were children, VNExpress reports.

Rescue efforts are set to continue into the night to find the many still missing.

Prime Minister Pham Minh Chinh sent his condolences to the families of the dead.

Authorities will investigate the cause of the accident and “strictly handle violations”, a government statement said.

Ha Long Bay in Quang Ninh province is dotted with hundreds of tiny islets, attracting 4 million tourists in 2019, and is a Unesco World Heritage site.



Source link

Continue Reading

Trending

Copyright © 2025 AISTORIZ. For enquiries email at prompt@travelstoriz.com