Artificial Intelligence thread

Bellum_Romanum

Brigadier
Registered Member
Instead of focusing on how effective and useful of Baidu's AI chat, SCMP solely focuses on politics. It is just so typical of SCMP.
Please, Log in or Register to view URLs content!
Western journalism is not journalism but YELLOW, FAKE JOURNALISM. I lean on these folks regardless of ethnic backgrounds inkling to write some hit pieces that will buy them notice and notoriety with American elites so that in turn, they can rub elbows and hang out with the elite clubs to feel special and important. These so-called "JOURNALISTS" just want to gain status without actually doing legitimate work. The Asian Boba liberals are the worst of them all.
 

tokenanalyst

Brigadier
Registered Member

Baichuan Intelligent announced that Baichuan 2 is open source​



Baichuan Intelligent announced the open source and fine-tuned Baichuan 2-7B, Baichuan 2-13B, Baichuan 2-13B-Chat and their 4bit quantized versions, and they are all free and commercially available.
In addition, Baichuan Intelligence has also open sourced the Check Point model training, and announced that it will release a Baichuan 2 technical report, which will introduce the training details of Baichuan 2 in detail, and help large-scale model academic institutions, developers and enterprise users to better understand its training process, and more Well done to advance large-scale academic research and the technical development of the community.
According to reports, both Baichuan 2-7B-Base and Baichuan 2-13B-Base are trained on the basis of 2.6 trillion high-quality multilingual data, while retaining the good generation and creation capabilities of the previous generation of open source models, as well as the smooth multi-round dialogue ability And based on many features such as low deployment threshold, the two models have significantly improved their capabilities in mathematics, code, security, logical reasoning, and semantic understanding.
Among them, compared with the previous generation 13B model, Baichuan 2-13B-Base has a 49% improvement in mathematics ability, a 46% increase in code ability, a 37% increase in security ability, a 25% increase in logical reasoning ability, and a 15% increase in semantic understanding ability.

Please, Log in or Register to view URLs content!
 

european_guy

Junior Member
Registered Member
tencent announces it's own LLM with > 100B parameters & >2B tokens
Please, Log in or Register to view URLs content!
they are a little late to this imo

Well, to train a 100B model with 2T tokens takes a lot of time.

It took
Please, Log in or Register to view URLs content!
(from Jan to June) to train a 70B model with 2T tokens.

As a ball-park estimate, training time is proportional to model size * number of tokens / number of GPUs

Meta does not reveal on how many GPU it was trained LLama 2, but you can bet it is a huge number!

Currently 2T tokens is state-of-the-art, apart from GPT 4 that according to leaks has been trained on a whopping 13T tokens, all other big models are in the 2/3T tokens range or lower...also a lot lower than that actually. For instance the open source 176B parameters
Please, Log in or Register to view URLs content!
model (honestly, not among the top ones) has been trained on only 350B tokens, i.e. almost 7 times less than the new Tencent's one.

Just to give an idea of what it means to train on such big datasets, we can assume a token corresponds to a word in English (it is actually less than a word, a 70% of a word on average), and to a single character in Chinese.

Now, in English and other languages with Latin alphabet, 1 page is about 500 words, so a 200 page book is about 100K words, and 1B tokens corresponds to the equivalent of training on 10K books.

In our case, Tencent model has been trained on 2T tokens, i.e. on the equivalent of 20 million books of 200 page each.

 

luminary

Senior Member
Registered Member
I'm sorry
Not looking good, I work in finance, maybe I should start a career as a chef? Because most of the finance job in Poland is outsourcing so a monkey can do the job too, so it's ideal for automation

Please, Log in or Register to view URLs content!

Ant Group unveiled a finance-specific artificial intelligence (AI) model on Friday and started testing consumer and professional apps for the product.

Ant said its new model had begun closed tests of the two apps (Zhixiaobao 1.0 and 2.0) on the company's wealth management and insurance platforms. It will be available once it receives regulatory approval.

  • The Zhixiaobao 2.0 app, designed to give consumers financial tips, can match the average financial professional in market analysis and reasoning capability, Ant said.
  • Zhixiaozhu 1.0 can conduct investment analysis and information extraction, among other business tasks, for financial professionals, Ant said.
 

Jiang ZeminFanboy

Senior Member
Registered Member
I'm sorry


Please, Log in or Register to view URLs content!

Ant Group unveiled a finance-specific artificial intelligence (AI) model on Friday and started testing consumer and professional apps for the product.

Ant said its new model had begun closed tests of the two apps (Zhixiaobao 1.0 and 2.0) on the company's wealth management and insurance platforms. It will be available once it receives regulatory approval.

  • The Zhixiaobao 2.0 app, designed to give consumers financial tips, can match the average financial professional in market analysis and reasoning capability, Ant said.
  • Zhixiaozhu 1.0 can conduct investment analysis and information extraction, among other business tasks, for financial professionals, Ant said.
I got into the university for two majors I wanted to study, civil engineering and economics, and in the end, I chose economics, when you're young you make stupid choices.
 
Top