The fintech affiliate of Alibaba said its Ling-Plus-Base model can be ‘effectively trained on lower-performance devices’.
Alibaba-affiliate Ant Group is reportedly using a mix of U.S.- and Chinese-made semiconductors to enhance the efficiency of ...
Ant used domestic chips to train models using Mixture of Experts machine learning approach Jack Ma-backed Ant Group Co. used Chinese-made semiconductors to develop techniques for training AI ...
Chain-of-experts chains LLM experts in a sequence, outperforming mixture-of-experts (MoE) with lower memory and compute costs.
Chinese AI startup DeepSeek upgrades its V3 model with the V3‑0324 update, enhancing programming capabilities and shifting to ...
TikTok owner ByteDance said it has achieved a 1.71 times efficiency improvement in large language model (LLM) training, the latest Chinese tech company to achieve a breakthrough that could potentially ...
Aardvark Weather, by the University of Cambridge, uses AI to deliver fast, accurate forecasts in minutes with minimal ...
Alibaba has upgraded its Quark AI assistant with the Qwen reasoning model, enhancing its ability to process complex queries ...
TikTok owner ByteDance said it has achieved a 1.71 times efficiency improvement in large language model (LLM ... an optimised Mixture-of-Experts (MoE) system, according to a recent paper published ...