open-source Archives - AI News https://www.artificialintelligence-news.com/news/tag/open-source/ Artificial Intelligence News Thu, 24 Apr 2025 11:40:58 +0000 en-GB hourly 1 https://wordpress.org/?v=6.8.1 https://www.artificialintelligence-news.com/wp-content/uploads/2020/09/cropped-ai-icon-32x32.png open-source Archives - AI News https://www.artificialintelligence-news.com/news/tag/open-source/ 32 32 Deep Cogito open LLMs use IDA to outperform same size models https://www.artificialintelligence-news.com/news/deep-cogito-open-llms-use-ida-outperform-same-size-models/ https://www.artificialintelligence-news.com/news/deep-cogito-open-llms-use-ida-outperform-same-size-models/#respond Wed, 09 Apr 2025 08:03:15 +0000 https://www.artificialintelligence-news.com/?p=105246 Deep Cogito has released several open large language models (LLMs) that outperform competitors and claim to represent a step towards achieving general superintelligence. The San Francisco-based company, which states its mission is “building general superintelligence,” has launched preview versions of LLMs in 3B, 8B, 14B, 32B, and 70B parameter sizes. Deep Cogito asserts that “each […]

The post Deep Cogito open LLMs use IDA to outperform same size models appeared first on AI News.

]]>
Deep Cogito has released several open large language models (LLMs) that outperform competitors and claim to represent a step towards achieving general superintelligence.

The San Francisco-based company, which states its mission is “building general superintelligence,” has launched preview versions of LLMs in 3B, 8B, 14B, 32B, and 70B parameter sizes. Deep Cogito asserts that “each model outperforms the best available open models of the same size, including counterparts from LLAMA, DeepSeek, and Qwen, across most standard benchmarks”.

Impressively, the 70B model from Deep Cogito even surpasses the performance of the recently released Llama 4 109B Mixture-of-Experts (MoE) model.   

Iterated Distillation and Amplification (IDA)

Central to this release is a novel training methodology called Iterated Distillation and Amplification (IDA). 

Deep Cogito describes IDA as “a scalable and efficient alignment strategy for general superintelligence using iterative self-improvement”. This technique aims to overcome the inherent limitations of current LLM training paradigms, where model intelligence is often capped by the capabilities of larger “overseer” models or human curators.

The IDA process involves two key steps iterated repeatedly:

  • Amplification: Using more computation to enable the model to derive better solutions or capabilities, akin to advanced reasoning techniques.
  • Distillation: Internalising these amplified capabilities back into the model’s parameters.

Deep Cogito says this creates a “positive feedback loop” where model intelligence scales more directly with computational resources and the efficiency of the IDA process, rather than being strictly bounded by overseer intelligence.

“When we study superintelligent systems,” the research notes, referencing successes like AlphaGo, “we find two key ingredients enabled this breakthrough: Advanced Reasoning and Iterative Self-Improvement”. IDA is presented as a way to integrate both into LLM training.

Deep Cogito claims IDA is efficient, stating the new models were developed by a small team in approximately 75 days. They also highlight IDA’s potential scalability compared to methods like Reinforcement Learning from Human Feedback (RLHF) or standard distillation from larger models.

As evidence, the company points to their 70B model outperforming Llama 3.3 70B (distilled from a 405B model) and Llama 4 Scout 109B (distilled from a 2T parameter model).

Capabilities and performance of Deep Cogito models

The newly released Cogito models – based on Llama and Qwen checkpoints – are optimised for coding, function calling, and agentic use cases.

A key feature is their dual functionality: “Each model can answer directly (standard LLM), or self-reflect before answering (like reasoning models),” similar to capabilities seen in models like Claude 3.5. However, Deep Cogito notes they “have not optimised for very long reasoning chains,” citing user preference for faster answers and the efficiency of distilling shorter chains.

Extensive benchmark results are provided, comparing Cogito models against size-equivalent state-of-the-art open models in both direct (standard) and reasoning modes.

Across various benchmarks (MMLU, MMLU-Pro, ARC, GSM8K, MATH, etc.) and model sizes (3B, 8B, 14B, 32B, 70B,) the Cogito models generally show significant performance gains over counterparts like Llama 3.1/3.2/3.3 and Qwen 2.5, particularly in reasoning mode.

For instance, the Cogito 70B model achieves 91.73% on MMLU in standard mode (+6.40% vs Llama 3.3 70B) and 91.00% in thinking mode (+4.40% vs Deepseek R1 Distill 70B). Livebench scores also show improvements.

Here are benchmarks of 14B models for a medium-sized comparison:

Benchmark comparison of medium 14B size large language models from Deep Cogito compared to Alibaba Qwen and DeepSeek R1

While acknowledging benchmarks don’t fully capture real-world utility, Deep Cogito expresses confidence in practical performance.

This release is labelled a preview, with Deep Cogito stating they are “still in the early stages of this scaling curve”. They plan to release improved checkpoints for the current sizes and introduce larger MoE models (109B, 400B, 671B) “in the coming weeks / months”. All future models will also be open-source.

(Photo by Pietro Mattia)

See also: Alibaba Cloud targets global AI growth with new models and tools

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Deep Cogito open LLMs use IDA to outperform same size models appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/deep-cogito-open-llms-use-ida-outperform-same-size-models/feed/ 0
DeepSeek V3-0324 tops non-reasoning AI models in open-source first https://www.artificialintelligence-news.com/news/deepseek-v3-0324-tops-non-reasoning-ai-models-open-source-first/ https://www.artificialintelligence-news.com/news/deepseek-v3-0324-tops-non-reasoning-ai-models-open-source-first/#respond Tue, 25 Mar 2025 13:10:20 +0000 https://www.artificialintelligence-news.com/?p=104986 DeepSeek V3-0324 has become the highest-scoring non-reasoning model on the Artificial Analysis Intelligence Index in a landmark achievement for open-source AI. The new model advanced seven points in the benchmark to surpass proprietary counterparts such as Google’s Gemini 2.0 Pro, Anthropic’s Claude 3.7 Sonnet, and Meta’s Llama 3.3 70B. While V3-0324 trails behind reasoning models, […]

The post DeepSeek V3-0324 tops non-reasoning AI models in open-source first appeared first on AI News.

]]>
DeepSeek V3-0324 has become the highest-scoring non-reasoning model on the Artificial Analysis Intelligence Index in a landmark achievement for open-source AI.

The new model advanced seven points in the benchmark to surpass proprietary counterparts such as Google’s Gemini 2.0 Pro, Anthropic’s Claude 3.7 Sonnet, and Meta’s Llama 3.3 70B.

While V3-0324 trails behind reasoning models, including DeepSeek’s own R1 and offerings from OpenAI and Alibaba, the achievement highlights the growing viability of open-source solutions in latency-sensitive applications where immediate responses are critical.

DeepSeek V3-0324 represents a new era for open-source AI

Non-reasoning models – which generate answers instantly without deliberative “thinking” phases – are essential for real-time use cases like chatbots, customer service automation, and live translation. DeepSeek’s latest iteration now sets the standard for these applications, eclipsing even leading proprietary tools.

Benchmark results of DeepSeek V3-0324 in the Artificial Analysis Intelligence Index showing a landmark achievement for non-reasoning open-source AI models.

“This is the first time an open weights model is the leading non-reasoning model, a milestone for open source,” states Artificial Analysis. The model’s performance edges it closer to proprietary reasoning models, though the latter remain superior for tasks requiring complex problem-solving.

DeepSeek V3-0324 retains most specifications from its December 2024 predecessor, including:  

  • 128k context window (capped at 64k via DeepSeek’s API)
  • 671 billion total parameters, necessitating over 700GB of GPU memory for FP8 precision
  • 37 billion active parameters
  • Text-only functionality (no multimodal support) 
  • MIT License

“Still not something you can run at home!” Artificial Analysis quips, emphasising its enterprise-grade infrastructure requirements.

Open-source AI is bringing the heat

While proprietary reasoning models like DeepSeek R1 maintain dominance in the broader Intelligence Index, the gap is narrowing.

Three months ago, DeepSeek V3 nearly matched Anthropic’s and Google’s proprietary models but fell short of surpassing them. Today, the updated V3-0324 not only leads open-source alternatives but also outperforms all proprietary non-reasoning rivals.

“This release is arguably even more impressive than R1,” says Artificial Analysis.

DeepSeek’s progress signals a shift in the AI sector, where open-source frameworks increasingly compete with closed systems. For developers and enterprises, the MIT-licensed V3-0324 offers a powerful, adaptable tool—though its computational costs may limit accessibility.

“DeepSeek are now driving the frontier of non-reasoning open weights models,” declares Artificial Analysis.

With R2 on the horizon, the community awaits another potential leap in AI performance.

(Photo by Paul Hanaoka)

See also: Hugging Face calls for open-source focus in the AI Action Plan

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post DeepSeek V3-0324 tops non-reasoning AI models in open-source first appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/deepseek-v3-0324-tops-non-reasoning-ai-models-open-source-first/feed/ 0
Hugging Face calls for open-source focus in the AI Action Plan https://www.artificialintelligence-news.com/news/hugging-face-open-source-focus-ai-action-plan/ https://www.artificialintelligence-news.com/news/hugging-face-open-source-focus-ai-action-plan/#respond Thu, 20 Mar 2025 17:41:39 +0000 https://www.artificialintelligence-news.com/?p=104946 Hugging Face has called on the US government to prioritise open-source development in its forthcoming AI Action Plan. In a statement to the Office of Science and Technology Policy (OSTP), Hugging Face emphasised that “thoughtful policy can support innovation while ensuring that AI development remains competitive, and aligned with American values.” Hugging Face, which hosts […]

The post Hugging Face calls for open-source focus in the AI Action Plan appeared first on AI News.

]]>
Hugging Face has called on the US government to prioritise open-source development in its forthcoming AI Action Plan.

In a statement to the Office of Science and Technology Policy (OSTP), Hugging Face emphasised that “thoughtful policy can support innovation while ensuring that AI development remains competitive, and aligned with American values.”

Hugging Face, which hosts over 1.5 million public models across various sectors and serves seven million users, proposes an AI Action Plan centred on three interconnected pillars:

  1. Hugging Face stresses the importance of strengthening open-source AI ecosystems.  The company argues that technical innovation stems from diverse actors across institutions and that support for infrastructure – such as the National AI Research Resource (NAIRR), and investment in open science and data – allows these contributions to have an additive effect and accelerate robust innovation.
  1. The company prioritises efficient and reliable adoption of AI. Hugging Face believes that spreading the benefits of the technology by facilitating its adoption along the value chain requires actors across sectors of activity to shape its development. It states that more efficient, modular, and robust AI models require research and infrastructural investments to enable the broadest possible participation and innovation—enabling diffusion of technology across the US economy.
  1. Hugging Face also highlights the need to promote security and standards. The company suggests that decades of practices in open-source software cybersecurity, information security, and standards can inform safer AI technology. It advocates for promoting traceability, disclosure, and interoperability standards to foster a more resilient and robust technology ecosystem.

Open-source is key for AI advancement in the US (and beyond)

Hugging Face underlines that modern AI is built on decades of open research, with commercial giants relying heavily on open-source contributions. Recent breakthroughs – such as OLMO-2 and Olympic-Coder – demonstrate that open research remains a promising path to developing systems that match the performance of commercial models, and can often surpass them, especially in terms of efficiency and performance in specific domains.

“Perhaps most striking is the rapid compression of development timelines,” notes the company, “what once required over 100B parameter models just two years ago can now be accomplished with 2B parameter models, suggesting an accelerating path to parity.”

This trend towards more accessible, efficient, and collaborative AI development indicates that open approaches to AI development have a critical role to play in enabling a successful AI strategy that maintains technical leadership and supports more widespread and secure adoption of the technology.

Hugging Face argues that open models, infrastructure, and scientific practices constitute the foundation of AI innovation, allowing a diverse ecosystem of researchers, companies, and developers to build upon shared knowledge.

The company’s platform hosts AI models and datasets from both small actors (e.g., startups, universities) and large organisations (e.g., Microsoft, Google, OpenAI, Meta), demonstrating how open approaches accelerate progress and democratise access to AI capabilities.

“The United States must lead in open-source AI and open science, which can enhance American competitiveness by fostering a robust ecosystem of innovation and ensuring a healthy balance of competition and shared innovation,” states Hugging Face.

Research has shown that open technical systems act as force multipliers for economic impact, with an estimated 2000x multiplier effect. This means that $4 billion invested in open systems could potentially generate $8 trillion in value for companies using them.

These economic benefits extend to national economies as well. Without any open-source software contributions, the average country would lose 2.2% of its GDP. Open-source drove between €65 billion and €95 billion of European GDP in 2018 alone, a finding so significant that the European Commission cited it when establishing new rules to streamline the process for open-sourcing government software.

This demonstrates how open-source impact translates directly into policy action and economic advantage at the national level, underlining the importance of open-source as a public good.

Practical factors driving commercial adoption of open-source AI

Hugging Face identifies several practical factors driving the commercial adoption of open models:

  • Cost efficiency is a major driver, as developing AI models from scratch requires significant investment, so leveraging open foundations reduces R&D expenses.
  • Customisation is crucial, as organisations can adapt and deploy models specifically tailored to their use cases rather than relying on one-size-fits-all solutions.
  • Open models reduce vendor lock-in, giving companies greater control over their technology stack and independence from single providers.
  • Open models have caught up to and, in certain cases, surpassed the capabilities of closed, proprietary systems.

These factors are particularly valuable for startups and mid-sized companies, which can access cutting-edge technology without massive infrastructure investments. Banks, pharmaceutical companies, and other industries have been adapting open models to specific market needs—demonstrating how open-source foundations support a vibrant commercial ecosystem across the value chain.

Hugging Face’s policy recommendations to support open-source AI in the US

To support the development and adoption of open AI systems, Hugging Face offers several policy recommendations:

  • Enhance research infrastructure: Fully implement and expand the National AI Research Resource (NAIRR) pilot. Hugging Face’s active participation in the NAIRR pilot has demonstrated the value of providing researchers with access to computing resources, datasets, and collaborative tools.
  • Allocate public computing resources for open-source: The public should have ways to participate via public AI infrastructure. One way to do this would be to dedicate a portion of publicly-funded computing infrastructure to support open-source AI projects, reducing barriers to innovation for smaller research teams and companies that cannot afford proprietary systems.
  • Enable access to data for developing open systems: Create sustainable data ecosystems through targeted policies that address the decreasing data commons. Publishers are increasingly signing data licensing deals with proprietary AI model developers, meaning that quality data acquisition costs are now approaching or even surpassing computational expenses of training frontier models, threatening to lock out small open developers from access to quality data.  Support organisations that contribute to public data repositories and streamline compliance pathways that reduce legal barriers to responsible data sharing.
  • Develop open datasets: Invest in the creation, curation, and maintenance of robust, representative datasets that can support the next generation of AI research and applications. Expand initiatives like the IBM AI Alliance Trusted Data Catalog and support projects like IDI’s AI-driven Digitization of the public collections in the Boston Public Library.
  • Strengthen rights-respecting data access frameworks: Establish clear guidelines for data usage, including standardised protocols for anonymisation, consent management, and usage tracking.  Support public-private partnerships to create specialised data trusts for high-value domains like healthcare and climate science, ensuring that individuals and organisations maintain appropriate control over their data while enabling innovation.    
  • Invest in stakeholder-driven innovation: Create and support programmes that enable organisations across diverse sectors (healthcare, manufacturing, education) to develop customised AI systems for their specific needs, rather than relying exclusively on general-purpose systems from major providers. This enables broader participation in the AI ecosystem and ensures that the benefits of AI extend throughout the economy.
  • Strengthen centres of excellence: Expand NIST’s role as a convener for AI experts across academia, industry, and government to share lessons and develop best practices.  In particular, the AI Risk Management Framework has played a significant role in identifying stages of AI development and research questions that are critical to ensuring more robust and secure technology deployment for all. The tools developed at Hugging Face, from model documentation to evaluation libraries, are directly shaped by these questions.
  • Support high-quality data for performance and reliability evaluation: AI development depends heavily on data, both to train models and to reliably evaluate their progress, strengths, risks, and limitations. Fostering greater access to public data in a safe and secure way and ensuring that the evaluation data used to characterise models is sound and evidence-based will accelerate progress in both performance and reliability of the technology.

Prioritising efficient and reliable AI adoption

Hugging Face highlights that smaller companies and startups face significant barriers to AI adoption due to high costs and limited resources. According to IDC, global AI spending will reach $632 billion in 2028, but these costs remain prohibitive for many small organisations.

For organisations adopting open-source AI tools, it brings financial returns. 51% of surveyed companies currently utilising open-source AI tools report positive ROI, compared to just 41% of those not using open-source.

However, energy scarcity presents a growing concern, with the International Energy Agency projecting that data centres’ electricity consumption could double from 2022 levels to 1,000 TWh by 2026, equivalent to Japan’s entire electricity demand. While training AI models is energy-intensive, inference, due to its scale and frequency, can ultimately exceed training energy consumption.

Ensuring broad AI accessibility requires both hardware optimisations and scalable software frameworks.  A range of organisations are developing models tailored to their specific needs, and US leadership in efficiency-focused AI development presents a strategic advantage. The DOE’s AI for Energy initiative further supports research into energy-efficient AI, facilitating wider adoption without excessive computational demands.

With its letter to the OSTP, Hugging Face advocates for an AI Action Plan centred on open-source principles. By taking decisive action, the US can secure its leadership, drive innovation, enhance security, and ensure the widespread benefits of AI are realised across society and the economy.

See also: UK minister in US to pitch Britain as global AI investment hub

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Hugging Face calls for open-source focus in the AI Action Plan appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/hugging-face-open-source-focus-ai-action-plan/feed/ 0
Gemma 3: Google launches its latest open AI models https://www.artificialintelligence-news.com/news/gemma-3-google-launches-its-latest-open-ai-models/ https://www.artificialintelligence-news.com/news/gemma-3-google-launches-its-latest-open-ai-models/#respond Wed, 12 Mar 2025 09:08:41 +0000 https://www.artificialintelligence-news.com/?p=104758 Google has launched Gemma 3, the latest version of its family of open AI models that aim to set a new benchmark for AI accessibility. Built upon the foundations of the company’s Gemini 2.0 models, Gemma 3 is engineered to be lightweight, portable, and adaptable—enabling developers to create AI applications across a wide range of […]

The post Gemma 3: Google launches its latest open AI models appeared first on AI News.

]]>
Google has launched Gemma 3, the latest version of its family of open AI models that aim to set a new benchmark for AI accessibility.

Built upon the foundations of the company’s Gemini 2.0 models, Gemma 3 is engineered to be lightweight, portable, and adaptable—enabling developers to create AI applications across a wide range of devices.  

This release comes hot on the heels of Gemma’s first birthday, an anniversary underscored by impressive adoption metrics. Gemma models have achieved more than 100 million downloads and spawned the creation of over 60,000 community-built variants. Dubbed the “Gemmaverse,” this ecosystem signals a thriving community aiming to democratise AI.  

“The Gemma family of open models is foundational to our commitment to making useful AI technology accessible,” explained Google.

Gemma 3: Features and capabilities

Gemma 3 models are available in various sizes – 1B, 4B, 12B, and 27B parameters – allowing developers to select a model tailored to their specific hardware and performance requirements. These models promise faster execution, even on modest computational setups, without compromising functionality or accuracy.

Here are some of the standout features of Gemma 3:  

  • Single-accelerator performance: Gemma 3 sets a new benchmark for single-accelerator models. In preliminary human preference evaluations on the LMArena leaderboard, Gemma 3 outperformed rivals including Llama-405B, DeepSeek-V3, and o3-mini.
  • Multilingual support across 140 languages: Catering to diverse audiences, Gemma 3 comes with pretrained capabilities for over 140 languages. Developers can create applications that connect with users in their native tongues, expanding the global reach of their projects.  
  • Sophisticated text and visual analysis: With advanced text, image, and short video reasoning capabilities, developers can implement Gemma 3 to craft interactive and intelligent applications—addressing an array of use cases from content analysis to creative workflows.  
  • Expanded context window: Offering a 128k-token context window, Gemma 3 can analyse and synthesise large datasets, making it ideal for applications requiring extended content comprehension.
  • Function calling for workflow automation: With function calling support, developers can utilise structured outputs to automate processes and build agentic AI systems effortlessly.
  • Quantised models for lightweight efficiency: Gemma 3 introduces official quantised versions, significantly reducing model size while preserving output accuracy—a bonus for developers optimising for mobile or resource-constrained environments.

The model’s performance advantages are clearly illustrated in the Chatbot Arena Elo Score leaderboard. Despite requiring just a single NVIDIA H100 GPU, the flagship 27B version of Gemma 3 ranks among the top chatbots, achieving an Elo score of 1338. Many competitors demand up to 32 GPUs to deliver comparable performance.

Google Gemma 3 performance illustrated on benchmark against both open source and proprietary AI models in the Chatbot Arena Elo Score leaderboard.

One of Gemma 3’s strengths lies in its adaptability within developers’ existing workflows.  

  • Diverse tooling compatibility: Gemma 3 supports popular AI libraries and tools, including Hugging Face Transformers, JAX, PyTorch, and Google AI Edge. For optimised deployment, platforms such as Vertex AI or Google Colab are ready to help developers get started with minimal hassle.  
  • NVIDIA optimisations: Whether using entry-level GPUs like Jetson Nano or cutting-edge hardware like Blackwell chips, Gemma 3 ensures maximum performance, further simplified through the NVIDIA API Catalog.  
  • Broadened hardware support: Beyond NVIDIA, Gemma 3 integrates with AMD GPUs via the ROCm stack and supports CPU execution with Gemma.cpp for added versatility.

For immediate experiments, users can access Gemma 3 models via platforms such as Hugging Face and Kaggle, or take advantage of the Google AI Studio for in-browser deployment.

Advancing responsible AI  

“We believe open models require careful risk assessment, and our approach balances innovation with safety,” explains Google.  

Gemma 3’s team adopted stringent governance policies, applying fine-tuning and robust benchmarking to align the model with ethical guidelines. Given the models enhanced capabilities in STEM fields, it underwent specific evaluations to mitigate risks of misuse, such as generating harmful substances.

Google is pushing for collective efforts within the industry to create proportionate safety frameworks for increasingly powerful models.

To play its part, Google is launching ShieldGemma 2. The 4B image safety checker leverages Gemma 3’s architecture and outputs safety labels across categories such as dangerous content, explicit material, and violence. While offering out-of-the-box solutions, developers can customise the tool to meet tailored safety requirements.

The “Gemmaverse” isn’t just a technical ecosystem, it’s a community-driven movement. Projects such as AI Singapore’s SEA-LION v3, INSAIT’s BgGPT, and Nexa AI’s OmniAudio are testament to the power of collaboration within this ecosystem.  

To bolster academic research, Google has also introduced the Gemma 3 Academic Program. Researchers can apply for $10,000 worth of Google Cloud credits to accelerate their AI-centric projects. Applications open today and remain available for four weeks.  

With its accessibility, capabilities, and widespread compatibility, Gemma 3 makes a strong case for becoming a cornerstone in the AI development community.

(Image credit: Google)

See also: Alibaba Qwen QwQ-32B: Scaled reinforcement learning showcase

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Gemma 3: Google launches its latest open AI models appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/gemma-3-google-launches-its-latest-open-ai-models/feed/ 0
Endor Labs: AI transparency vs ‘open-washing’ https://www.artificialintelligence-news.com/news/endor-labs-ai-transparency-vs-open-washing/ https://www.artificialintelligence-news.com/news/endor-labs-ai-transparency-vs-open-washing/#respond Mon, 24 Feb 2025 18:15:45 +0000 https://www.artificialintelligence-news.com/?p=104605 As the AI industry focuses on transparency and security, debates around the true meaning of “openness” are intensifying. Experts from open-source security firm Endor Labs weighed in on these pressing topics. Andrew Stiefel, Senior Product Marketing Manager at Endor Labs, emphasised the importance of applying lessons learned from software security to AI systems. “The US […]

The post Endor Labs: AI transparency vs ‘open-washing’ appeared first on AI News.

]]>
As the AI industry focuses on transparency and security, debates around the true meaning of “openness” are intensifying. Experts from open-source security firm Endor Labs weighed in on these pressing topics.

Andrew Stiefel, Senior Product Marketing Manager at Endor Labs, emphasised the importance of applying lessons learned from software security to AI systems.

“The US government’s 2021 Executive Order on Improving America’s Cybersecurity includes a provision requiring organisations to produce a software bill of materials (SBOM) for each product sold to federal government agencies.”

An SBOM is essentially an inventory detailing the open-source components within a product, helping detect vulnerabilities. Stiefel argued that “applying these same principles to AI systems is the logical next step.”  

“Providing better transparency for citizens and government employees not only improves security,” he explained, “but also gives visibility into a model’s datasets, training, weights, and other components.”

What does it mean for an AI model to be “open”?  

Julien Sobrier, Senior Product Manager at Endor Labs, added crucial context to the ongoing discussion about AI transparency and “openness.” Sobrier broke down the complexity inherent in categorising AI systems as truly open.

“An AI model is made of many components: the training set, the weights, and programs to train and test the model, etc. It is important to make the whole chain available as open source to call the model ‘open’. It is a broad definition for now.”  

Sobrier noted the lack of consistency across major players, which has led to confusion about the term.

“Among the main players, the concerns about the definition of ‘open’ started with OpenAI, and Meta is in the news now for their LLAMA model even though that’s ‘more open’. We need a common understanding of what an open model means. We want to watch out for any ‘open-washing,’ as we saw it with free vs open-source software.”  

One potential pitfall, Sobrier highlighted, is the increasingly common practice of “open-washing,” where organisations claim transparency while imposing restrictions.

“With cloud providers offering a paid version of open-source projects (such as databases) without contributing back, we’ve seen a shift in many open-source projects: The source code is still open, but they added many commercial restrictions.”  

“Meta and other ‘open’ LLM providers might go this route to keep their competitive advantage: more openness about the models, but preventing competitors from using them,” Sobrier warned.

DeepSeek aims to increase AI transparency

DeepSeek, one of the rising — albeit controversial — players in the AI industry, has taken steps to address some of these concerns by making portions of its models and code open-source. The move has been praised for advancing transparency while providing security insights.  

“DeepSeek has already released the models and their weights as open-source,” said Andrew Stiefel. “This next move will provide greater transparency into their hosted services, and will give visibility into how they fine-tune and run these models in production.”

Such transparency has significant benefits, noted Stiefel. “This will make it easier for the community to audit their systems for security risks and also for individuals and organisations to run their own versions of DeepSeek in production.”  

Beyond security, DeepSeek also offers a roadmap on how to manage AI infrastructure at scale.

“From a transparency side, we’ll see how DeepSeek is running their hosted services. This will help address security concerns that emerged after it was discovered they left some of their Clickhouse databases unsecured.”

Stiefel highlighted that DeepSeek’s practices with tools like Docker, Kubernetes (K8s), and other infrastructure-as-code (IaC) configurations could empower startups and hobbyists to build similar hosted instances.  

Open-source AI is hot right now

DeepSeek’s transparency initiatives align with the broader trend toward open-source AI. A report by IDC reveals that 60% of organisations are opting for open-source AI models over commercial alternatives for their generative AI (GenAI) projects.  

Endor Labs research further indicates that organisations use, on average, between seven and twenty-one open-source models per application. The reasoning is clear: leveraging the best model for specific tasks and controlling API costs.

“As of February 7th, Endor Labs found that more than 3,500 additional models have been trained or distilled from the original DeepSeek R1 model,” said Stiefel. “This shows both the energy in the open-source AI model community, and why security teams need to understand both a model’s lineage and its potential risks.”  

For Sobrier, the growing adoption of open-source AI models reinforces the need to evaluate their dependencies.

“We need to look at AI models as major dependencies that our software depends on. Companies need to ensure they are legally allowed to use these models but also that they are safe to use in terms of operational risks and supply chain risks, just like open-source libraries.”

He emphasised that any risks can extend to training data: “They need to be confident that the datasets used for training the LLM were not poisoned or had sensitive private information.”  

Building a systematic approach to AI model risk  

As open-source AI adoption accelerates, managing risk becomes ever more critical. Stiefel outlined a systematic approach centred around three key steps:  

  1. Discovery: Detect the AI models your organisation currently uses.  
  2. Evaluation: Review these models for potential risks, including security and operational concerns.  
  3. Response: Set and enforce guardrails to ensure safe and secure model adoption.  

“The key is finding the right balance between enabling innovation and managing risk,” Stiefel said. “We need to give software engineering teams latitude to experiment but must do so with full visibility. The security team needs line-of-sight and the insight to act.”  

Sobrier further argued that the community must develop best practices for safely building and adopting AI models. A shared methodology is needed to evaluate AI models across parameters such as security, quality, operational risks, and openness.

Beyond transparency: Measures for a responsible AI future  

To ensure the responsible growth of AI, the industry must adopt controls that operate across several vectors:  

  • SaaS models: Safeguarding employee use of hosted models.
  • API integrations: Developers embedding third-party APIs like DeepSeek into applications, which, through tools like OpenAI integrations, can switch deployment with just two lines of code.
  • Open-source models: Developers leveraging community-built models or creating their own models from existing foundations maintained by companies like DeepSeek.

Sobrier warned of complacency in the face of rapid AI progress. “The community needs to build best practices to develop safe and open AI models,” he advised, “and a methodology to rate them along security, quality, operational risks, and openness.”  

As Stiefel succinctly summarised: “Think about security across multiple vectors and implement the appropriate controls for each.”

See also: AI in 2025: Purpose-driven models, human integration, and more

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Endor Labs: AI transparency vs ‘open-washing’ appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/endor-labs-ai-transparency-vs-open-washing/feed/ 0
DeepSeek to open-source AGI research amid privacy concerns https://www.artificialintelligence-news.com/news/deepseek-open-source-agi-research-amid-privacy-concerns/ https://www.artificialintelligence-news.com/news/deepseek-open-source-agi-research-amid-privacy-concerns/#respond Fri, 21 Feb 2025 13:56:59 +0000 https://www.artificialintelligence-news.com/?p=104592 DeepSeek, a Chinese AI startup aiming for artificial general intelligence (AGI), announced plans to open-source five repositories starting next week as part of its commitment to transparency and community-driven innovation. However, this development comes against the backdrop of mounting controversies that have drawn parallels to the TikTok saga. Today, DeepSeek shared its intentions in a […]

The post DeepSeek to open-source AGI research amid privacy concerns appeared first on AI News.

]]>
DeepSeek, a Chinese AI startup aiming for artificial general intelligence (AGI), announced plans to open-source five repositories starting next week as part of its commitment to transparency and community-driven innovation.

However, this development comes against the backdrop of mounting controversies that have drawn parallels to the TikTok saga.

Today, DeepSeek shared its intentions in a tweet that outlined its vision of open collaboration: “We’re a tiny team at DeepSeek exploring AGI. Starting next week, we’ll be open-sourcing five repos, sharing our small but sincere progress with full transparency.”

The repositories – which the company describes as “documented, deployed, and battle-tested in production” – include fundamental building blocks of DeepSeek’s online service.

By open-sourcing its tools, DeepSeek hopes to contribute to the broader AI research community.

“As part of the open-source community, we believe that every line shared becomes collective momentum that accelerates the journey. No ivory towers – just pure garage-energy and community-driven innovation,” the company said.

This philosophy has drawn praise for fostering collaboration in a field that often suffers from secrecy, but DeepSeek’s rapid rise has also raised eyebrows.

Despite being a small team with a mission rooted in transparency, the company has been under intense scrutiny amid allegations of data misuse and geopolitical entanglements.

Rising fast, under fire

Practically unknown until recently, DeepSeek burst onto the scene with a business model that stood in stark contrast to more established players like OpenAI and Google.

Offering its advanced AI capabilities for free, DeepSeek quickly gained global acclaim for its cutting-edge performance. However, its exponential rise has also sparked debates about the trade-offs between innovation and privacy.

US lawmakers are now pushing for a ban on DeepSeek after security researchers found the app transferring user data to a banned state-owned company.

A probe has also been launched by Microsoft and OpenAI over a breach of the latter’s systems by a group allegedly linked to DeepSeek.

Concerns about data collection and potential misuse have triggered comparisons to the controversies surrounding TikTok, another Chinese tech success story grappling with regulatory pushback in the West.

DeepSeek continues AGI innovation amid controversy

DeepSeek’s commitment to open-source its technology appears timed to deflect criticism and reassure sceptics about its intentions.

Open-sourcing has long been heralded as a way to democratise technology and increase transparency, and DeepSeek’s “daily unlocks,” that are set to begin soon, could offer the community reassuring insight into its operations.

Nevertheless, questions remain over how much of the technology will be open for scrutiny and whether the move is an attempt to shift the narrative amid growing political and regulatory pressure.

It’s unclear whether this balancing act will be enough to satisfy lawmakers or deter critics, but one thing is certain: DeepSeek’s open-source leap marks another turn in its dramatic rise.

While the company’s motto of “garage-energy and community-driven innovation” resonates with developers eager for open collaboration, its future may rest as much on its ability to address security concerns as on its technical prowess.

(Photo by Solen Feyissa)

See also: DeepSeek’s AI dominance expands from EVs to e-scooters in China

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including IoT Tech Expo, Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post DeepSeek to open-source AGI research amid privacy concerns appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/deepseek-open-source-agi-research-amid-privacy-concerns/feed/ 0
AI in 2025: Purpose-driven models, human integration, and more https://www.artificialintelligence-news.com/news/ai-in-2025-purpose-driven-models-human-integration-and-more/ https://www.artificialintelligence-news.com/news/ai-in-2025-purpose-driven-models-human-integration-and-more/#respond Fri, 14 Feb 2025 17:16:32 +0000 https://www.artificialintelligence-news.com/?p=104468 As AI becomes increasingly embedded in our daily lives, industry leaders and experts are forecasting a transformative 2025. From groundbreaking developments to existential challenges, AI’s evolution will continue to shape industries, change workflows, and spark deeper conversations about its implications. For this article, AI News caught up with some of the world’s leading minds to […]

The post AI in 2025: Purpose-driven models, human integration, and more appeared first on AI News.

]]>
As AI becomes increasingly embedded in our daily lives, industry leaders and experts are forecasting a transformative 2025.

From groundbreaking developments to existential challenges, AI’s evolution will continue to shape industries, change workflows, and spark deeper conversations about its implications.

For this article, AI News caught up with some of the world’s leading minds to see what they envision for the year ahead.

Smaller, purpose-driven models

Grant Shipley, Senior Director of AI at Red Hat, predicts a shift away from valuing AI models by their sizeable parameter counts.

Grant Shipley, Senior Director of AI at Red Hat

“2025 will be the year when we stop using the number of parameters that models have as a metric to indicate the value of a model,” he said.  

Instead, AI will focus on specific applications. Developers will move towards chaining together smaller models in a manner akin to microservices in software development. This modular, task-based approach is likely to facilitate more efficient and bespoke applications suited to particular needs.

Open-source leading the way

Bill Higgins, VP of watsonx Platform Engineering and Open Innovation at IBM

Bill Higgins, VP of watsonx Platform Engineering and Open Innovation at IBM, expects open-source AI models will grow in popularity in 2025.

“Despite mounting pressure, many enterprises are still struggling to show measurable returns on their AI investments—and the high licensing fees of proprietary models is a major factor. In 2025, open-source AI solutions will emerge as a dominant force in closing this gap,” he explains.

Alongside the affordability of open-source AI models comes transparency and increased customisation potential, making them ideal for multi-cloud environments. With open-source models matching proprietary systems in power, they could offer a way for enterprises to move beyond experimentation and into scalability.

Nick Burling, SVP at Nasuni

This plays into a prediction from Nick Burling, SVP at Nasuni, who believes that 2025 will usher in a more measured approach to AI investments. 

“Enterprises will focus on using AI strategically, ensuring that every AI initiative is justified by clear, measurable returns,” said Burling.

Cost efficiency and edge data management will become crucial, helping organisations optimise operations while keeping budgets in check.  

Augmenting human expertise

Jonathan Siddharth, CEO of Turing

For Jonathan Siddharth, CEO of Turing, the standout feature of 2025 AI systems will be their ability to learn from human expertise at scale.

“The key advancement will come from teaching AI not just what to do, but how to approach problems with the logical reasoning that coding naturally cultivates,” he says.

Competitiveness, particularly in industries like finance and healthcare, will hinge on mastering this integration of human expertise with AI.  

Behavioural psychology will catch up

Understanding the interplay between human behaviour and AI systems is at the forefront of predictions for Niklas Mortensen, Chief Design Officer at Designit.

Niklas Mortensen, Chief Design Officer at Designit

“With so many examples of algorithmic bias leading to unwanted outputs – and humans being, well, humans – behavioural psychology will catch up to the AI train,” explained Mortensen.  

The solutions? Experimentation with ‘pause moments’ for human oversight and intentional balance between automation and human control in critical operations such as healthcare and transport.

Mortensen also believes personal AI assistants will finally prove their worth by meeting their long-touted potential in organising our lives efficiently and intuitively.

Bridge between physical and digital worlds

Andy Wilson, Senior Director at Dropbox

Andy Wilson, Senior Director at Dropbox, envisions AI becoming an indispensable part of our daily lives.

“AI will evolve from being a helpful tool to becoming an integral part of daily life and work – offering innovative ways to connect, create, and collaborate,” Wilson says.  

Mobile devices and wearables will be at the forefront of this transformation, delivering seamless AI-driven experiences.

However, Wilson warns of new questions on boundaries between personal and workplace data, spurred by such integrations.

Driving sustainability goals 

Kendra DeKeyrel, VP ESG & Asset Management at IBM

With 2030 sustainability targets looming over companies, Kendra DeKeyrel, VP ESG & Asset Management at IBM, highlights how AI can help fill the gap.

DeKeyrel calls on organisations to adopt AI-powered technologies for managing energy consumption, lifecycle performance, and data centre strain.

“These capabilities can ultimately help progress sustainability goals overall,” she explains.

Unlocking computational power and inference

James Ingram, VP Technology at Streetbees

James Ingram, VP Technology at Streetbees, foresees a shift in computational requirements as AI scales to handle increasingly complex problems.

“The focus will move from pre-training to inference compute,” he said, highlighting the importance of real-time reasoning capabilities.

Expanding context windows will also significantly enhance how AI retains and processes information, likely surpassing human efficiency in certain domains.

Rise of agentic AI and unified data foundations

Dominic Wellington, Enterprise Architect at SnapLogic

According to Dominic Wellington, Enterprise Architect at SnapLogic, “Agentic AI marks a more flexible and creative era for AI in 2025.”

However, such systems require robust data integration because siloed information risks undermining their reliability.

Wellington anticipates that 2025 will witness advanced solutions for improving data hygiene, integrity, and lineage—all vital for enabling agentic AI to thrive.  

From hype to reality

Jason Schern, Field CTO of Cognite

Jason Schern, Field CTO of Cognite, predicts that 2025 will be remembered as the year when truly transformative, validated generative AI solutions emerge.

“Through the fog of AI for AI’s sake noise, singular examples of truly transformative embedding of Gen AI into actual workflows will stand out,” predicts Schern.  

These domain-specific AI agents will revolutionise industrial workflows by offering tailored decision-making. Schern cited an example in which AI slashed time-consuming root cause analyses from months to mere minutes.

Deepfakes and crisis of trust

Siggi Stefnisson, CTO at Gen

Sophisticated generative AI threatens the authenticity of images, videos, and information, warns Siggi Stefnisson, Cyber Safety CTO at Gen.

“Even experts may not be able to tell what’s authentic,” warns Stefnisson.

Combating this crisis requires robust digital credentials for verifying authenticity and promoting trust in increasingly blurred digital realities.

2025: Foundational shifts in the AI landscape

As multiple predictions converge, it’s clear that foundational shifts are on the horizon.

The experts that contributed to this year’s industry predictions highlight smarter applications, stronger integration with human expertise, closer alignment with sustainability goals, and heightened security. However, many also foresee significant ethical challenges.

2025 represents a crucial year: a transition from the initial excitement of AI proliferation to mature and measured adoption that promises value and a more nuanced understanding of its impact.

See also: AI Action Summit: Leaders call for unity and equitable development

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post AI in 2025: Purpose-driven models, human integration, and more appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/ai-in-2025-purpose-driven-models-human-integration-and-more/feed/ 0
Ai2 OLMo 2: Raising the bar for open language models https://www.artificialintelligence-news.com/news/ai2-olmo-2-raising-bar-open-language-models/ https://www.artificialintelligence-news.com/news/ai2-olmo-2-raising-bar-open-language-models/#respond Wed, 27 Nov 2024 18:43:42 +0000 https://www.artificialintelligence-news.com/?p=16566 Ai2 is releasing OLMo 2, a family of open-source language models that advances the democratisation of AI and narrows the gap between open and proprietary solutions. The new models, available in 7B and 13B parameter versions, are trained on up to 5 trillion tokens and demonstrate performance levels that match or exceed comparable fully open […]

The post Ai2 OLMo 2: Raising the bar for open language models appeared first on AI News.

]]>
Ai2 is releasing OLMo 2, a family of open-source language models that advances the democratisation of AI and narrows the gap between open and proprietary solutions.

The new models, available in 7B and 13B parameter versions, are trained on up to 5 trillion tokens and demonstrate performance levels that match or exceed comparable fully open models whilst remaining competitive with open-weight models such as Llama 3.1 on English academic benchmarks.

“Since the release of the first OLMo in February 2024, we’ve seen rapid growth in the open language model ecosystem, and a narrowing of the performance gap between open and proprietary models,” explained Ai2.

The development team achieved these improvements through several innovations, including enhanced training stability measures, staged training approaches, and state-of-the-art post-training methodologies derived from their Tülu 3 framework. Notable technical improvements include the switch from nonparametric layer norm to RMSNorm and the implementation of rotary positional embedding.

OLMo 2 model training breakthrough

The training process employed a sophisticated two-stage approach. The initial stage utilised the OLMo-Mix-1124 dataset of approximately 3.9 trillion tokens, sourced from DCLM, Dolma, Starcoder, and Proof Pile II. The second stage incorporated a carefully curated mixture of high-quality web data and domain-specific content through the Dolmino-Mix-1124 dataset.

Particularly noteworthy is the OLMo 2-Instruct-13B variant, which is the most capable model in the series. The model demonstrates superior performance compared to Qwen 2.5 14B instruct, Tülu 3 8B, and Llama 3.1 8B instruct models across various benchmarks.

Benchmarks comparing the OLMo 2 open large language model to other models such as Mistral, Qwn, Llama, Gemma, and more.
(Credit: Ai2)

Commiting to open science

Reinforcing its commitment to open science, Ai2 has released comprehensive documentation including weights, data, code, recipes, intermediate checkpoints, and instruction-tuned models. This transparency allows for full inspection and reproduction of results by the wider AI community.

The release also introduces an evaluation framework called OLMES (Open Language Modeling Evaluation System), comprising 20 benchmarks designed to assess core capabilities such as knowledge recall, commonsense reasoning, and mathematical reasoning.

OLMo 2 raises the bar in open-source AI development, potentially accelerating the pace of innovation in the field whilst maintaining transparency and accessibility.

(Photo by Rick Barrett)

See also: OpenAI enhances AI safety with new red teaming methods

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Ai2 OLMo 2: Raising the bar for open language models appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/ai2-olmo-2-raising-bar-open-language-models/feed/ 0
Industry leaders back open-source AI definition https://www.artificialintelligence-news.com/news/industry-leaders-back-open-source-ai-definition/ https://www.artificialintelligence-news.com/news/industry-leaders-back-open-source-ai-definition/#respond Tue, 29 Oct 2024 14:36:15 +0000 https://www.artificialintelligence-news.com/?p=16411 The Open Source Initiative (OSI) has unveiled a definition framework to evaluate whether AI systems can be classified as open-source. The announcement of the first Open Source AI Definition (OSAID) was made at All Things Open and marks the culmination of a comprehensive global effort spanning multiple years of research, international workshops, and a year-long […]

The post Industry leaders back open-source AI definition appeared first on AI News.

]]>
The Open Source Initiative (OSI) has unveiled a definition framework to evaluate whether AI systems can be classified as open-source.

The announcement of the first Open Source AI Definition (OSAID) was made at All Things Open and marks the culmination of a comprehensive global effort spanning multiple years of research, international workshops, and a year-long community design process.

The OSI – widely recognised as the definitive authority on open-source definitions by individuals, organisations, and government bodies worldwide – developed the framework through extensive collaboration with industry stakeholders. This framework defines what open-source AI means, insisting that the same open-source requirements apply whether to a fully functional AI system, a model, weights and parameters, or other structural elements.

An open-source AI system must be made available under terms that grant four essential freedoms:

  • Use the system for any purpose and without having to ask for permission.
  • Study how the system works and inspect its components.
  • Modify the system for any purpose, including to change its output.
  • Share the system for others to use with or without modifications, for any purpose.

These freedoms apply both to a fully functional system and to discrete elements of a system. A precondition to exercising these freedoms is having access to the preferred form to make modifications to the system, which includes detailed data information, complete source code, and model parameters.

“The co-design process that led to version 1.0 of the Open Source AI Definition was well-developed, thorough, inclusive, and fair,” said Carlo Piana, OSI board chair. “The board is confident that the process has resulted in a definition that meets the standards of open-source as defined in the open-source definition and the four essential freedoms.”

One of the framework’s most significant requirements is the mandate for open-source models to provide sufficient information about their training data, ensuring that “a skilled person can recreate a substantially equivalent system using the same or similar data,” according to Ayah Bdeir, who leads AI strategy at Mozilla.

Bdeir acknowledged that whilst this approach might not be perfect, it represents a practical compromise between ideological purity and real-world implementation. She suggested that demanding an unrealistically high standard could prove counterproductive to the initiative’s goals.

The Digital Public Goods Alliance (DPGA) has expressed support for the OSI’s leadership in defining open-source AI. Liv Marte Nordhaug, CEO of the DPGA secretariat, confirmed that her organisation will incorporate this foundational work into updates to their Digital Public Goods Standard for AI applications.

EleutherAI Institute, known for its non-profit work in AI development, has also endorsed the definition.

“The Open Source AI Definition is a necessary step towards promoting the benefits of open-source principles in the field of AI,” stated Stella Biderman, Executive Director of the EleutherAI Institute. “We believe that this definition supports the needs of independent machine learning researchers and promotes greater transparency among the largest AI developers.”

The definition highlights the importance of including data information and code when sharing open-source models and weights. These requirements ensure transparency and the ability to modify the AI system.

OSI Executive Director Stefano Maffulli acknowledged the challenges faced during the development process, noting that despite occasional heated exchanges and differing opinions, the final result aligned with the project’s initial objectives.

“This is a starting point for a continued effort to engage with the communities to improve the definition over time,” he stated.

The OSAID does not require a specific legal mechanism for assuring that model parameters are freely available to all, though it may involve licences or legal instruments. This aspect is expected to become clearer over time as the legal system addresses these open-source AI systems.

See also: President Biden issues first National Security Memorandum on AI

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Industry leaders back open-source AI definition appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/industry-leaders-back-open-source-ai-definition/feed/ 0
Stability AI releases most powerful image generation models to date https://www.artificialintelligence-news.com/news/stability-ai-releases-most-powerful-image-generation-models-date/ https://www.artificialintelligence-news.com/news/stability-ai-releases-most-powerful-image-generation-models-date/#respond Tue, 22 Oct 2024 16:28:28 +0000 https://www.artificialintelligence-news.com/?p=16370 Stability AI has announced the release of Stable Diffusion 3.5, marking a leap forward in open-source AI image generation models. The latest models from Stability AI include multiple variants designed to cater to different user needs, from hobbyists to enterprise-level applications. The announcement follows June’s Stable Diffusion 3 Medium release, which the company acknowledges didn’t […]

The post Stability AI releases most powerful image generation models to date appeared first on AI News.

]]>
Stability AI has announced the release of Stable Diffusion 3.5, marking a leap forward in open-source AI image generation models.

The latest models from Stability AI include multiple variants designed to cater to different user needs, from hobbyists to enterprise-level applications.

The announcement follows June’s Stable Diffusion 3 Medium release, which the company acknowledges didn’t meet expectations.

“This release didn’t fully meet our standards or our communities’ expectations,” Stability AI stated.

Rather than rushing a quick fix, Stability AI says it invested time in developing a more robust solution.

The flagship model, Stable Diffusion 3.5 Large, boasts 8 billion parameters and operates at 1 megapixel resolution—making it the most powerful in the Stable Diffusion family. Alongside it, the Large Turbo variant offers comparable quality but generates images in just four steps, significantly reducing processing time.

A Medium version, scheduled for release on 29th October, will feature 2.5 billion parameters and support image generation between 0.25 and 2 megapixel resolution. This variant is specifically optimised for consumer hardware.

Benchmark comparing the performance of the new Stable Diffusion 3.5 image generation models from Stability AI.

The models incorporate Query-Key Normalisation in transformer blocks, enhancing training stability and simplifying fine-tuning processes. However, this flexibility comes with trade-offs, including greater variation in outputs from identical prompts with different seeds.

Stability AI has implemented a notably permissive community licence for the release. The models are free for non-commercial use and available to businesses with annual revenues under $1 million. Enterprises exceeding this threshold must secure separate licensing arrangements.

The company emphasised its commitment to responsible AI development, implementing safety measures from the early stages. Additional features, including ControlNets for advanced control features, are planned for release following the Medium model’s launch.

Stability AI’s latest image generation models are currently available via Hugging Face and GitHub, with additional access through platforms including the Stability AI API, Replicate, ComfyUI, and DeepInfra.

(Image Credit: Stability AI)

See also: Anthropic unveils new Claude AI models and ‘computer control’

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Stability AI releases most powerful image generation models to date appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/stability-ai-releases-most-powerful-image-generation-models-date/feed/ 0
Ivo Everts, Databricks: Enhancing open-source AI and improving data governance https://www.artificialintelligence-news.com/news/ivo-everts-databricks-open-source-ai-improving-data-governance/ https://www.artificialintelligence-news.com/news/ivo-everts-databricks-open-source-ai-improving-data-governance/#respond Fri, 27 Sep 2024 12:29:26 +0000 https://www.artificialintelligence-news.com/?p=16179 Ahead of AI & Big Data Expo Europe, AI News caught up with Ivo Everts, Senior Solutions Architect at Databricks, to discuss several key developments set to shape the future of open-source AI and data governance. One of Databricks’ notable achievements is the DBRX model, which set a new standard for open large language models […]

The post Ivo Everts, Databricks: Enhancing open-source AI and improving data governance appeared first on AI News.

]]>
Ahead of AI & Big Data Expo Europe, AI News caught up with Ivo Everts, Senior Solutions Architect at Databricks, to discuss several key developments set to shape the future of open-source AI and data governance.

One of Databricks’ notable achievements is the DBRX model, which set a new standard for open large language models (LLMs).

“Upon release, DBRX outperformed all other leading open models on standard benchmarks and has up to 2x faster inference than models like Llama2-70B,” Everts explains. “It was trained more efficiently due to a variety of technological advances.

“From a quality standpoint, we believe that DBRX is one of the best open-source models out there and when we refer to ‘best’ this means a wide range of industry benchmarks, including language understanding (MMLU), Programming (HumanEval), and Math (GSM8K).”

The open-source AI model aims to “democratise the training of custom LLMs beyond a small handful of model providers and show organisations that they can train world-class LLMs on their data in a cost-effective way.”

In line with their commitment to open ecosystems, Databricks has also open-sourced Unity Catalog.

“Open-sourcing Unity Catalog enhances its adoption across cloud platforms (e.g., AWS, Azure) and on-premise infrastructures,” Everts notes. “This flexibility allows organisations to uniformly apply data governance policies regardless of where the data is stored or processed.”

Unity Catalog addresses the challenges of data sprawl and inconsistent access controls through various features:

  1. Centralised data access management: “Unity Catalog centralises the governance of data assets, allowing organisations to manage access controls in a unified manner,” Everts states.
  2. Role-Based Access Control (RBAC): According to Everts, Unity Catalog “implements Role-Based Access Control (RBAC), allowing organisations to assign roles and permissions based on user profiles.”
  3. Data lineage and auditing: This feature “helps organisations monitor data usage and dependencies, making it easier to identify and eliminate redundant or outdated data,” Everts explains. He adds that it also “logs all data access and changes, providing a detailed audit trail to ensure compliance with data security policies.”
  4. Cross-cloud and hybrid support: Everts points out that Unity Catalog “is designed to manage data governance in multi-cloud and hybrid environments” and “ensures that data is governed uniformly, regardless of where it resides.”

The company has introduced Databricks AI/BI, a new business intelligence product that leverages generative AI to enhance data exploration and visualisation. Everts believes that “a truly intelligent BI solution needs to understand the unique semantics and nuances of a business to effectively answer questions for business users.”

The AI/BI system includes two key components:

  1. Dashboards: Everts describes this as “an AI-powered, low-code interface for creating and distributing fast, interactive dashboards.” These include “standard BI features like visualisations, cross-filtering, and periodic reports without needing additional management services.”
  2. Genie: Everts explains this as “a conversational interface for addressing ad-hoc and follow-up questions through natural language.” He adds that it “learns from underlying data to generate adaptive visualisations and suggestions in response to user queries, improving over time through feedback and offering tools for analysts to refine its outputs.”

Everts states that Databricks AI/BI is designed to provide “a deep understanding of your data’s semantics, enabling self-service data analysis for everyone in an organisation.” He notes it’s powered by “a compound AI system that continuously learns from usage across an organisation’s entire data stack, including ETL pipelines, lineage, and other queries.”

Databricks also unveiled Mosaic AI, which Everts describes as “a comprehensive platform for building, deploying, and managing machine learning and generative AI applications, integrating enterprise data for enhanced performance and governance.”

Mosaic AI offers several key components, which Everts outlines:

  1. Unified tooling: Provides “tools for building, deploying, evaluating, and governing AI and ML solutions, supporting predictive models and generative AI applications.”
  2. Generative AI patterns: “Supports prompt engineering, retrieval augmented generation (RAG), fine-tuning, and pre-training, offering flexibility as business needs evolve.”
  3. Centralised model management: “Model Serving allows for centralised deployment, governance, and querying of AI models, including custom ML models and foundation models.”
  4. Monitoring and governance: “Lakehouse Monitoring and Unity Catalog ensure comprehensive monitoring, governance, and lineage tracking across the AI lifecycle.”
  5. Cost-effective custom LLMs: “Enables training and serving custom large language models at significantly lower costs, tailored to specific organisational domains.”

Everts highlights that Mosaic AI’s approach to fine-tuning and customising foundation models includes unique features like “fast startup times” by “utilising in-cluster base model caching,” “live prompt evaluation” where users can “track how the model’s responses change throughout the training process,” and support for “custom pre-trained checkpoints.”

At the heart of these innovations lies the Data Intelligence Platform, which Everts says “transforms data management by using AI models to gain deep insights into the semantics of enterprise data.” The platform combines features of data lakes and data warehouses, utilises Delta Lake technology for real-time data processing, and incorporates Delta Sharing for secure data exchange across organisational boundaries.

Everts explains that the Data Intelligence Platform plays a crucial role in supporting new AI and data-sharing initiatives by providing:

  1. A unified data and AI platform that “combines the features of data lakes and data warehouses into a single architecture.”
  2. Delta Lake for real-time data processing, ensuring “reliable data governance, ACID transactions, and real-time data processing.”
  3. Collaboration and data sharing via Delta Sharing, enabling “secure and open data sharing across organisational boundaries.”
  4. Integrated support for machine learning and AI model development with popular libraries like MLflow, PyTorch, and TensorFlow.
  5. Scalability and performance through its cloud-native architecture and the Photon engine, “an optimised query execution engine.”

As a key sponsor of AI & Big Data Expo Europe, Databricks plans to showcase their open-source AI and data governance solutions during the event.

“At our stand, we will also showcase how to create and deploy – with Lakehouse apps – a custom GenAI app from scratch using open-source models from Hugging Face and data from Unity Catalog,” says Everts.

“With our GenAI app you can generate your own cartoon picture, all running on the Data Intelligence Platform.”

Databricks will be sharing more of their expertise at this year’s AI & Big Data Expo Europe. Swing by Databricks’ booth at stand #280 to hear more about open AI and improving data governance.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Ivo Everts, Databricks: Enhancing open-source AI and improving data governance appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/ivo-everts-databricks-open-source-ai-improving-data-governance/feed/ 0
Alibaba Cloud unleashes over 100 open-source AI models https://www.artificialintelligence-news.com/news/alibaba-cloud-unleashes-over-100-open-source-ai-models/ https://www.artificialintelligence-news.com/news/alibaba-cloud-unleashes-over-100-open-source-ai-models/#respond Fri, 20 Sep 2024 13:08:45 +0000 https://www.artificialintelligence-news.com/?p=16135 Alibaba Cloud has open-sourced more than 100 of its newly-launched AI models, collectively known as Qwen 2.5. The announcement was made during the company’s annual Apsara Conference. The cloud computing arm of Alibaba Group has also unveiled a revamped full-stack infrastructure designed to meet the surging demand for robust AI computing. This new infrastructure encompasses […]

The post Alibaba Cloud unleashes over 100 open-source AI models appeared first on AI News.

]]>
Alibaba Cloud has open-sourced more than 100 of its newly-launched AI models, collectively known as Qwen 2.5. The announcement was made during the company’s annual Apsara Conference.

The cloud computing arm of Alibaba Group has also unveiled a revamped full-stack infrastructure designed to meet the surging demand for robust AI computing. This new infrastructure encompasses innovative cloud products and services that enhance computing, networking, and data centre architecture, all aimed at supporting the development and wide-ranging applications of AI models.

Eddie Wu, Chairman and CEO of Alibaba Cloud Intelligence, said: “Alibaba Cloud is investing, with unprecedented intensity, in the research and development of AI technology and the building of its global infrastructure. We aim to establish an AI infrastructure of the future to serve our global customers and unlock their business potential.”

The newly-released Qwen 2.5 models range from 0.5 to 72 billion parameters in size and boast enhanced knowledge and stronger capabilities in maths and coding. Supporting over 29 languages, these models cater to a wide array of AI applications both at the edge and in the cloud across various sectors, from automotive and gaming to scientific research.

Alibaba Cloud’s open-source AI models gain traction

Since its debut in April 2023, the Qwen model series has garnered significant traction, surpassing 40 million downloads across platforms such as Hugging Face and ModelScope. These models have also inspired the creation of over 50,000 derivative models on Hugging Face alone.

Jingren Zhou, CTO of Alibaba Cloud Intelligence, commented: “This initiative is set to empower developers and corporations of all sizes, enhancing their ability to leverage AI technologies and further stimulating the growth of the open-source community.”

In addition to the open-source models, Alibaba Cloud announced an upgrade to its proprietary flagship model, Qwen-Max. The enhanced version reportedly demonstrates performance on par with other state-of-the-art models in areas such as language comprehension, reasoning, mathematics, and coding.

The company has also expanded its multimodal capabilities with a new text-to-video model as part of its Tongyi Wanxiang large model family. This model can generate high-quality videos in various visual styles, from realistic scenes to 3D animation, based on Chinese and English text instructions.

Furthermore, Alibaba Cloud introduced Qwen2-VL, an updated vision language model capable of comprehending videos lasting over 20 minutes and supporting video-based question-answering. The company also launched an AI Developer, a Qwen-powered AI assistant designed to support programmers in automating tasks such as requirement analysis, code programming, and bug identification and fixing.

To support these AI advancements, Alibaba Cloud has announced several infrastructure upgrades, including:

  • CUBE DC 5.0, a next-generation data centre architecture that increases energy and operational efficiency.
  • Alibaba Cloud Open Lake, a solution to maximise data utility for generative AI applications.
  • PAI AI Scheduler, a proprietary cloud-native scheduling engine for enhanced computing resource management.
  • DMS: OneMeta+OneOps, a platform for unified management of metadata across multiple cloud environments.
  • 9th Generation Enterprise Elastic Compute Service (ECS) instance, offering improved performance for various applications.

These updates from Alibaba Cloud – including the release of over 100 open-source models – aim to provide comprehensive support for customers and partners to maximise the benefits of the latest technology in building more efficient, sustainable, and inclusive AI applications.

(Image Source: www.alibabagroup.com)

See also: Tech industry giants urge EU to streamline AI regulations

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Alibaba Cloud unleashes over 100 open-source AI models appeared first on AI News.

]]>
https://www.artificialintelligence-news.com/news/alibaba-cloud-unleashes-over-100-open-source-ai-models/feed/ 0