maths Archives - AI News https://www.artificialintelligence-news.com/news/tag/maths/ Artificial Intelligence News Fri, 25 Apr 2025 14:07:33 +0000 en-GB hourly 1 https://wordpress.org/?v=6.8.1 https://www.artificialintelligence-news.com/wp-content/uploads/2020/09/cropped-ai-icon-32x32.png maths Archives - AI News https://www.artificialintelligence-news.com/news/tag/maths/ 32 32 LG EXAONE Deep is a maths, science, and coding buff https://www.artificialintelligence-news.com/news/lg-exaone-deep-maths-science-and-coding-buff/ https://www.artificialintelligence-news.com/news/lg-exaone-deep-maths-science-and-coding-buff/#respond Tue, 18 Mar 2025 12:49:26 +0000 https://www.artificialintelligence-news.com/?p=104905 LG AI Research has unveiled EXAONE Deep, a reasoning model that excels in complex problem-solving across maths, science, and coding. The company highlighted the global challenge in creating advanced reasoning models, noting that currently, only a handful of organisations with foundational models are actively pursuing this complex area. EXAONE Deep aims to compete directly with […]

The post LG EXAONE Deep is a maths, science, and coding buff appeared first on AI News.

]]>
LG AI Research has unveiled EXAONE Deep, a reasoning model that excels in complex problem-solving across maths, science, and coding.

The company highlighted the global challenge in creating advanced reasoning models, noting that currently, only a handful of organisations with foundational models are actively pursuing this complex area. EXAONE Deep aims to compete directly with these leading models, showcasing a competitive level of reasoning ability.

LG AI Research has focused its efforts on dramatically improving EXAONE Deep’s reasoning capabilities in core domains. The model also demonstrates a strong ability to understand and apply knowledge across a broader range of subjects.

The performance benchmarks released by LG AI Research are impressive:

  • Maths: The EXAONE Deep 32B model outperformed a competing model, despite being only 5% of its size, in a demanding mathematics benchmark. Furthermore, the 7.8B and 2.4B versions achieved first place in all major mathematics benchmarks for their respective model sizes.
  • Science and coding: In these areas, the EXAONE Deep models (7.8B and 2.4B) have secured the top spot across all major benchmarks.
  • MMLU (Massive Multitask Language Understanding): The 32B model achieved a score of 83.0 on the MMLU benchmark, which LG AI Research claims is the best performance among domestic Korean models.

The capabilities of the EXAONE Deep 32B model have already garnered international recognition.

Shortly after its release, it was included in the ‘Notable AI Models’ list by US-based non-profit research organisation Epoch AI. This listing places EXAONE Deep alongside its predecessor, EXAONE 3.5, making LG the only Korean entity with models featured on this prestigious list in the past two years.

Maths prowess

EXAONE Deep has demonstrated exceptional mathematical reasoning skills across its various model sizes (32B, 7.8B, and 2.4B). In assessments based on the 2025 academic year’s mathematics curriculum, all three models outperformed global reasoning models of comparable size.

The 32B model achieved a score of 94.5 in a general mathematics competency test and 90.0 in the American Invitational Mathematics Examination (AIME) 2024, a qualifying exam for the US Mathematical Olympiad.

In the AIME 2025, the 32B model matched the performance of DeepSeek-R1—a significantly larger 671B model. This result showcases EXAONE Deep’s efficient learning and strong logical reasoning abilities, particularly when tackling challenging mathematical problems.

The smaller 7.8B and 2.4B models also achieved top rankings in major benchmarks for lightweight and on-device models, respectively. The 7.8B model scored 94.8 on the MATH-500 benchmark and 59.6 on AIME 2025, while the 2.4B model achieved scores of 92.3 and 47.9 in the same evaluations.

Science and coding excellence

EXAONE Deep has also showcased remarkable capabilities in professional science reasoning and software coding.

The 32B model scored 66.1 on the GPQA Diamond test, which assesses problem-solving skills in doctoral-level physics, chemistry, and biology. In the LiveCodeBench evaluation, which measures coding proficiency, the model achieved a score of 59.5, indicating its potential for high-level applications in these expert domains.

The 7.8B and 2.4B models continued this trend of strong performance, both securing first place in the GPQA Diamond and LiveCodeBench benchmarks within their respective size categories. This achievement builds upon the success of the EXAONE 3.5 2.4B model, which previously topped Hugging Face’s LLM Readerboard in the edge division.

Enhanced general knowledge

Beyond its specialised reasoning capabilities, EXAONE Deep has also demonstrated improved performance in general knowledge understanding.

The 32B model achieved an impressive score of 83.0 on the MMLU benchmark, positioning it as the top-performing domestic model in this comprehensive evaluation. This indicates that EXAONE Deep’s reasoning enhancements extend beyond specific domains and contribute to a broader understanding of various subjects.

LG AI Research believes that EXAONE Deep’s reasoning advancements represent a leap towards a future where AI can tackle increasingly complex problems and contribute to enriching and simplifying human lives through continuous research and innovation.

See also: Baidu undercuts rival AI models with ERNIE 4.5 and ERNIE X1

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post LG EXAONE Deep is a maths, science, and coding buff appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/lg-exaone-deep-maths-science-and-coding-buff/feed/ 0
Qwen2-Math: A new era for AI maths whizzes https://www.artificialintelligence-news.com/news/qwen2-math-new-era-ai-maths-whizzes/ https://www.artificialintelligence-news.com/news/qwen2-math-new-era-ai-maths-whizzes/#respond Fri, 09 Aug 2024 12:46:18 +0000 https://www.artificialintelligence-news.com/?p=15689 Alibaba Cloud’s Qwen team has unveiled Qwen2-Math, a series of large language models specifically designed to tackle complex mathematical problems. These new models – built upon the existing Qwen2 foundation – demonstrate remarkable proficiency in solving arithmetic and mathematical challenges, and outperform former industry leaders. The Qwen team crafted Qwen2-Math using a vast and diverse […]

The post Qwen2-Math: A new era for AI maths whizzes appeared first on AI News.

]]>
Alibaba Cloud’s Qwen team has unveiled Qwen2-Math, a series of large language models specifically designed to tackle complex mathematical problems.

These new models – built upon the existing Qwen2 foundation – demonstrate remarkable proficiency in solving arithmetic and mathematical challenges, and outperform former industry leaders.

The Qwen team crafted Qwen2-Math using a vast and diverse Mathematics-specific Corpus. This corpus comprises a rich tapestry of high-quality resources, including web texts, books, code, exam questions, and synthetic data generated by Qwen2 itself.

Rigorous evaluation on both English and Chinese mathematical benchmarks – including GSM8K, Math, MMLU-STEM, CMATH, and GaoKao Math – revealed the exceptional capabilities of Qwen2-Math. Notably, the flagship model, Qwen2-Math-72B-Instruct, surpassed the performance of proprietary models such as GPT-4o and Claude 3.5 in various mathematical tasks.

“Qwen2-Math-Instruct achieves the best performance among models of the same size, with RM@8 outperforming Maj@8, particularly in the 1.5B and 7B models,” the Qwen team noted.

This superior performance is attributed to the effective implementation of a math-specific reward model during the development process.

Further showcasing its prowess, Qwen2-Math demonstrated impressive results in challenging mathematical competitions like the American Invitational Mathematics Examination (AIME) 2024 and the American Mathematics Contest (AMC) 2023.

To ensure the model’s integrity and prevent contamination, the Qwen team implemented robust decontamination methods during both the pre-training and post-training phases. This rigorous approach involved removing duplicate samples and identifying overlaps with test sets to maintain the model’s accuracy and reliability.

Looking ahead, the Qwen team plans to expand Qwen2-Math’s capabilities beyond English, with bilingual and multilingual models in the pipeline.  This commitment to inclusivity aims to make advanced mathematical problem-solving accessible to a global audience.

“We will continue to enhance our models’ ability to solve complex and challenging mathematical problems,” affirmed the Qwen team.

You can find the Qwen2 models on Hugging Face here.

See also: Paige and Microsoft unveil next-gen AI models for cancer diagnosis

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Qwen2-Math: A new era for AI maths whizzes appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/qwen2-math-new-era-ai-maths-whizzes/feed/ 0
Bill Gates: AI will be teaching kids literacy within 18 months https://www.artificialintelligence-news.com/news/bill-gates-ai-teaching-kids-literacy-within-18-months/ https://www.artificialintelligence-news.com/news/bill-gates-ai-teaching-kids-literacy-within-18-months/#respond Mon, 24 Apr 2023 15:35:06 +0000 https://www.artificialintelligence-news.com/?p=12985 AI chatbots could be used to improve children’s reading and writing skills within the next 18 months, according to Microsoft co-founder Bill Gates. In a fireside chat at the ASU+GSV Summit in San Diego, Gates explained that the “AIs will get to that ability, to be as good a tutor as any human ever could.” […]

The post Bill Gates: AI will be teaching kids literacy within 18 months appeared first on AI News.

]]>
AI chatbots could be used to improve children’s reading and writing skills within the next 18 months, according to Microsoft co-founder Bill Gates.

In a fireside chat at the ASU+GSV Summit in San Diego, Gates explained that the “AIs will get to that ability, to be as good a tutor as any human ever could.”

AI chatbots such as OpenAI’s ChatGPT and Google’s Bard have developed rapidly in recent months and can now compete with human-level intelligence on some standardised tests.

Teaching writing skills has traditionally been difficult for computers, as they lack the cognitive ability to replicate human thought processes, Gates said. However, AI chatbots are able to recognise and recreate human-like language.

New York Times tech columnist Kevin Roose has already used ChatGPT to improve his writing, using the AI’s ability to quickly search through online style guides. Some academics have also been impressed by chatbots’ ability to summarise and offer feedback on text or even to write full essays.

The technology must improve before it can become a viable tutor, and Gates said that AI must get better at reading and recreating human language to better motivate students.

While it may be surprising that chatbots are expected to excel at reading and writing before maths, the latter is often used to develop AI technology and chatbots have difficulties with mathematical calculations.

If a solved math equation already exists within the datasets that the chatbot is trained on, it can provide the answer. However, calculating its own solution is more complex and requires improved reasoning abilities, Gates explained.

Gates is confident that the technology will improve within the next two years and he believes that it could help make private tutoring available to a wide range of students who may not otherwise be able to afford it.

While some free versions of chatbots already exist, Gates expects that more advanced versions will be available for a fee, although he believes that they will be more affordable and accessible than one-on-one tutoring with a human instructor.

You can watch the full talk with Bill Gates below:

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The event is co-located with Digital Transformation Week.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Bill Gates: AI will be teaching kids literacy within 18 months appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/bill-gates-ai-teaching-kids-literacy-within-18-months/feed/ 0