Artificial Intelligence thread

escobar

Brigadier
From Beijing Academy of Artificial Intelligence. fundamental challenges to China's domestic large AI model development:
No self sufficiency in model architecture, over-reliance on LLaMA is a severe problem
Multiple domestic chip vendors each with own ecosystem, making high performance deployment difficult
GIpysfjbAAAlmoo.jpg
 

tphuang

Lieutenant General
Staff member
Super Moderator
VIP Professional
Registered Member
From Beijing Academy of Artificial Intelligence. fundamental challenges to China's domestic large AI model development:
No self sufficiency in model architecture, over-reliance on LLaMA is a severe problem
Multiple domestic chip vendors each with own ecosystem, making high performance deployment difficult
View attachment 126660
i mean this is just stupid.

100B parameter large models are by nature extremely unstable to run. In fact, Huawei/Ascend is the most stable platform. More so than Cuda for sure.

Also it say most of domestic large models is based on Llama, not all of them. And even if they are based on Llama, what's the problem? It's open source. If you can improve it, that's even better.
 

fatzergling

New Member
Registered Member
From Beijing Academy of Artificial Intelligence. fundamental challenges to China's domestic large AI model development:
No self sufficiency in model architecture, over-reliance on LLaMA is a severe problem
Multiple domestic chip vendors each with own ecosystem, making high performance deployment difficult
View attachment 126660
Model architectures are open-source at this point.
 

Bellum_Romanum

Brigadier
Registered Member
i mean this is just stupid.

100B parameter large models are by nature extremely unstable to run. In fact, Huawei/Ascend is the most stable platform. More so than Cuda for sure.

Also it say most of domestic large models is based on Llama, not all of them. And even if they are based on Llama, what's the problem? It's open source. If you can improve it, that's even better.
Not according to this expert:
 

broadsword

Brigadier

Please, Log in or Register to view URLs content!

PixArt-Σ can keep up with commercial alternatives such as Adobe Firefly 2, Google Imagen 2, OpenAI DALL-E 3 and Midjourney v6, the researchers claim.

I just tried Huawei's PixArt-Alpha 1024px by inputting 'chocolate cake with gooey topping', and the delicious result.



027c01fd-0ea7-4a6c-9b0e-94f35c71f294.png
 

gpt

Junior Member
Registered Member
Please, Log in or Register to view URLs content!

Nvidia unveils Blackwell (Perplexity AI summary)

NVIDIA has introduced the Blackwell platform, marking a significant advancement in computing technology. The Blackwell platform, featuring the new Blackwell GPU, is designed to power a new era of computing, enabling organizations to build and run AI models with up to trillion parameters. This innovation is expected to unlock breakthroughs in various fields, including data processing, engineering simulation, electronic design automation, computer-aided drug design, quantum computing, and generative AI.

The Blackwell GPU architecture incorporates six transformative technologies for accelerated computing. Among the notable advancements are new Tensor Cores and a TensorRT-LLM Compiler, which together reduce the operating cost and energy consumption of LLM inference by up to 25 times. This makes the Blackwell platform a highly efficient solution for running large-scale AI models.

The platform has garnered widespread support from major cloud providers, server makers, and leading AI companies, including Amazon Web Services, Dell Technologies, Google, Meta, Microsoft, OpenAI, Oracle, Tesla, and xAI. This broad adoption underscores the platform's potential to revolutionize computing across industries[1].

One of the key features of the Blackwell architecture is its ability to perform matrix math with floating point numbers just 4 bits wide, offering unprecedented precision and efficiency. The architecture also includes the fifth generation of NVIDIA's computer interconnect technology, NVLink, which delivers 1.8 terabytes per second bidirectionally between GPUs. This high-speed communication capability is crucial for building large-scale computers capable of processing trillion-parameter neural network models.

The DGX SuperPOD, powered by NVIDIA GB200 Grace Blackwell Superchips, represents the next generation of AI supercomputing. It is designed for generative AI supercomputing at a trillion-parameter scale, showcasing the immense computational power of the Blackwell platform.

NVIDIA's unveiling of the next-gen Blackwell GPUs highlights a significant reduction in costs and energy consumption for AI processing tasks. The GB200 Grace Blackwell Superchip, which consists of multiple chips in a single package, promises up to 30 times performance increase for LLM inference workloads compared to previous iterations. This advancement is set to herald a transformative era in computing, with potential applications extending beyond traditional fields to include gaming products in the future.
 
Top