0 reviews
Open bilingual pre-trained model supporting Chinese and English
130B-parameter transformer-based large language model
Open repository with model/code licenses; permitted research and certain commercial uses per repo
Trained on 400B text tokens with the GLM training framework
Optimized large-scale training using the GLM library (parallelism and efficiency)
Strong reported performance across multiple NLP benchmarks
Inference and deployment scripts, including multi-GPU and mixed-precision
Downloadable checkpoints/weights for immediate use
Instruction-style and few-shot prompting capabilities
Associated ICLR 2023 paper describing model design and training
If you've used this product, share your thoughts with other customers
High-Performance Tensor Library for Machine Learning
NVIDIA Megatron-LM: Training Large-Scale Transformer Models Made Easy
Stability.AI's StableLM Zephyr-3B: Advanced and Versatile Language Model
Unleash the Power of Language with 10x LLM.
Transform Text Creation with GPT-3's Advanced Language Model
Transform Finance with BloombergGPT's AI-Powered Insights
Benchmarking bilingual (Chinese–English) NLU and generation tasks with a large, open model.
Building question answering and dialogue systems via prompt-based inference.
Rapid prototyping of domain prompts and few-shot workflows without fine-tuning.
Evaluating 130B-scale model performance for internal pilots under the repository license.
Drafting bilingual content and templates in Chinese or English for user-facing features.
Studying large-model inference on multi-GPU and mixed-precision setups using provided scripts.
Demonstrating state-of-the-art transformer behavior in classroom or lab settings.
Comparing strong baseline results across standard NLP leaderboards.
Exploring instruction-style prompting and few-shot examples for task completion.
Extending the GLM library or improving deployment tooling for large models.