Repository navigation

#

pretraining

Llama中文社区,实时汇总最新Llama学习资料,构建最好的中文Llama大模型开源生态,完全开源可商用

Python
14546
14 天前

General technology for enabling AI capabilities w/ LLMs and MLLMs

Python
3934
4 天前

Official repository of OFA (ICML 2022). Paper: OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework

Python
2495
1 年前

Papers about pretraining and self-supervised learning on Graph Neural Networks (GNN).

Python
1666
1 年前
keyu-tian/SparK

[ICLR'23 Spotlight🔥] The first successful BERT/MAE-style pretraining on any convolutional network; Pytorch impl. of "Designing BERT for Convolutional Networks: Sparse and Hierarchical Masked Modeling"

Python
1336
1 年前

Recent Advances in Vision and Language PreTrained Models (VL-PTMs)

1152
3 年前

X-modaler is a versatile and high-performance codebase for cross-modal analytics(e.g., image captioning, video captioning, vision-language pre-training, visual question answering, visual commonsense reasoning, and cross-modal retrieval).

Python
970
2 年前

Official Repository for the Uni-Mol Series Methods

Python
835
6 天前

【ICLR 2024🔥】 Extending Video-Language Pretraining to N-modality by Language-based Semantic Alignment

Python
802
1 年前

Official Pytorch Implementation of: "ImageNet-21K Pretraining for the Masses"(NeurIPS, 2021) paper

Python
760
2 年前

Pretraining code for a large-scale depth-recurrent language model

Python
743
7 天前

A curated list of 3D Vision papers relating to Robotics domain in the era of large models i.e. LLMs/VLMs, inspired by awesome-computer-vision, including papers, codes, and related websites

680
6 个月前

Official repository for "Craw4LLM: Efficient Web Crawling for LLM Pretraining"

Python
612
2 个月前

The official implementation of MARS: Unleashing the Power of Variance Reduction for Training Large Models

Python
609
2 个月前

PITI: Pretraining is All You Need for Image-to-Image Translation

Python
499
1 年前

飞桨大模型开发套件,提供大语言模型、跨模态大模型、生物计算大模型等领域的全流程开发工具链。

Python
465
1 年前

[ACL 2022] LinkBERT: A Knowledgeable Language Model 😎 Pretrained with Document Links

Python
437
3 年前

Personal Project: MPP-Qwen14B & MPP-Qwen-Next(Multimodal Pipeline Parallel based on Qwen-LM). Support [video/image/multi-image] {sft/conversations}. Don't let the poverty limit your imagination! Train your own 8B/14B LLaVA-training-like MLLM on RTX3090/4090 24GB.

Jupyter Notebook
437
1 个月前