-
South China University of Technology
- GuangZhou, China
Stars
ODM: A Text-Image Further Alignment Pre-training Approach for Scene Text Detection and Spotting
Fast and memory-efficient exact attention
Qwen2.5-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
Agent Laboratory is an end-to-end autonomous research workflow meant to assist you as the human researcher toward implementing your research ideas
🔥 Sa2VA: Marrying SAM2 with LLaVA for Dense Grounded Understanding of Images and Videos
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型
Official code for "F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching"
The official implementation of Autoregressive Image Generation using Residual Quantization (CVPR '22)
[NeurIPS 2024 Best Paper][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ult…
PixArt-α: Fast Training of Diffusion Transformer for Photorealistic Text-to-Image Synthesis
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
Tensorflow implementation for DilatedRNN
The paper "Learning Representations for Time Series Clustering"
This is an official implementation code for paper "A Survey on Time-Series Pre-Trained Models" (TKDE-24).
The official implementation of SPTS v2: Single-Point Text Spotting
Official implementation of SPTS: Single-Point Text Spotting (ACM MM 2022 Oral)
Implementation of Paper Generating Sequences With Recurrent Neural Networks
Magenta: Music and Art Generation with Machine Intelligence
This repository is a non-official implementation of Recurrent (Conditional) GAN (Esteban et al., 2017) using PyTorch.
A concise but complete full-attention transformer with a set of promising experimental features from various papers
[official] PyTorch implementation of TimeVQVAE from the paper ["Vector Quantized Time Series Generation with a Bidirectional Prior Model", AISTATS 2023]
About Code release for "PatchMixer: A Patch-Mixing Architecture for Long-Term Time Series Forecasting"