Competing hard on the AI front, China’s DeepSeek AI introduced a new LLM called DeepSeek Chat this week, which is more powerful than any other current LLM.
DeepSeek Chat has two variants of 7B and 67B parameters, which are trained on a dataset of 2 trillion tokens, says the maker. Open-sourcing the new LLM for public research, DeepSeek AI proved that their DeepSeek Chat is much better than Meta’s Llama 2-70B in various fields.
A Much Better Chinese LLM
A year after ChatGPT’s launch, the Generative AI race is filled with many LLMs from various firms, all trying to excel by offering the best productivity tools. The latest in this pursuit is DeepSeek Chat, from China’s DeepSeek AI.
The company launched two variants of it’s DeepSeek Chat this week: a 7B and 67B-parameter DeepSeek LLM, trained on a dataset of 2 trillion tokens in English and Chinese. People who tested the 67B-parameter assistant said the tool had outperformed Meta’s Llama 2-70B – the current best we have in the LLM market.
As per benchmarks, 7B and 67B DeepSeek Chat variants have recorded strong performance in coding, mathematics and Chinese comprehension. DeepSeek AI has open-sourced both these models, allowing businesses to leverage under specific terms.
While it’s praised for it’s technical capabilities, some noted the LLM has censorship issues! An X user shared that a query made regarding China was automatically redacted by the assistant, with a message saying the content was “withdrawn” for security reasons.
On the western part, Microsoft has recently released Orca 2 models that performed better than Llama-2Chat-70B, which is at least 5x bigger than Orca 2. We may see more such models outperform the current leaders, like GPT-4 or Claude 2, soon.
Other Trending News:- News