Repository navigation
bert
- Website
- Wikipedia
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
《李宏毅深度学习教程》(李宏毅老师推荐👍,苹果书🍎),PDF下载地址:https://github.com/datawhalechina/leedl-tutorial/releases
Natural Language Processing Tutorial for Deep Learning Researchers
🏄 Scalable embedding, reasoning, ranking for images and sentences with CLIP
Easy-to-use and powerful LLM and SLM library with awesome model zoo.
This repository contains demos I made with the Transformers library by HuggingFace.
Pre-Training with Whole Word Masking for Chinese BERT(中文BERT-wwm系列模型)
大规模中文自然语言处理语料 Large Scale Chinese Corpus for NLP
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
vits2 backbone with multilingual-bert
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
Leveraging BERT and c-TF-IDF to create easily interpretable topics.
Google AI 2018 BERT pytorch implementation
Official implementations for various pre-training models of ERNIE-family, covering topics of Language Understanding & Generation, Multimodal Understanding & Generation, and beyond.
Transformer related optimization, including BERT, GPT
Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合
Tensorflow solution of NER task Using BiLSTM-CRF model with Google BERT Fine-tuning And private Server services
Tutorials on getting started with PyTorch and TorchText for sentiment analysis.
Must-read papers on prompt-based tuning for pre-trained language models.