Yequan's Academic
Yequan's Academic
Home
Projects
Publications
Patents
Talks
Contact
Light
Dark
Automatic
English
English
中文 (简体)
Large Model
Not All Layers of LLMs Are Necessary During Inference
We propose AdaInfer, a lightweight algorithm that adaptively stops LLM inference early based on statistical cues, cutting up to 43% of computation with <1% performance loss and no model modification.
Siqi Fan
,
Xin Jiang
,
Xuying Meng
,
Peng Han
,
Shuo Shang
,
Aixin Sun
,
Yequan Wang
PDF
Cite
Project
Project
Few-Shot Learner Generalizes Across AI-Generated Image Detection
We propose Few-Shot Detector (FSD), a novel AI-generated image detector that learns a metric space to recognize unseen fake images with only a few samples, achieving +11.6% higher accuracy and strong generalization without retraining.
Shiyu Wu
,
Jing Liu
,
Jing Li
,
Yequan Wang
PDF
Cite
Code
Project
Project
52B to 1T: Lessons Learned via Tele-FLM Series
As scaling laws underscore the potential of increasing model sizes, the academic community has intensified its investigations into LLMs with capacities exceeding 50 billion parameters. This technical report builds on our prior work with Tele-FLM (also known as FLM-2), a publicly available 52-billion-parameter model.
Xiang Li
,
Yiqun Yao
,
Xin Jiang
,
Xuezhi Fang
,
China Telecom
,
Yequan Wang
,
Zhongjiang He
,
Zhongyuan Wang
,
Xuelong Li
,
Tiejun Huang
PDF
Cite
Project
Tele-FLM-1T
Tele-FLM
Masked Structural Growth for 2x Faster Language Model Pre-training
To lower the computional cost of training large model, we focus on speeding up pre-training by progressively growing from a small Transformer structure to a large one.
Yiqun Yao
,
Zheng Zhang
,
Jing Li
,
Yequan Wang
PDF
Cite
Code
Project
Project
FLM Family
The FLM series of large models are a set of large models completed by the Cofe-AI team, including FLM-2, FLM-101B, and FreeLM. The core technology mainly consists of growth techniques, loss prediction technologies, and the FreeLM framework, among others.
PDF
Cite
×