Data processing with ML and LLM
-
Updated
Sep 13, 2024 - Python
Data processing with ML and LLM
🧠💬 Articles I wrote about machine learning, archived from MachineCurve.com.
中文nlp解决方案(大模型、数据、模型、训练、推理)
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
Social networking platform with automated content moderation and context-based authentication system
Simple UI for LLM Model Finetuning
Label, clean and enrich text datasets with LLMs.
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
Pocket-Sized Multimodal AI for content understanding and generation across multilingual texts, images, and 🔜 video, up to 5x faster than OpenAI CLIP and LLaVA 🖼️ & 🖋️
Trained models & code to predict toxic comments on all 3 Jigsaw Toxic Comment Challenges. Built using ⚡ Pytorch Lightning and 🤗 Transformers. For access to our API, please email us at [email protected].
Transformer models from BERT to GPT-4, environments from Hugging Face to OpenAI. Fine-tuning, training, and prompt engineering examples. A bonus section with ChatGPT, GPT-3.5-turbo, GPT-4, and DALL-E including jump starting GPT-4, speech-to-text, text-to-speech, text to image generation with DALL-E, Google Cloud AI,HuggingGPT, and more
Official PyTorch Implementation of MambaVision: A Hybrid Mamba-Transformer Vision Backbone
Learn Cloud Applied Generative AI Engineering (GenEng) using OpenAI, Gemini, Streamlit, Containers, Serverless, Postgres, LangChain, Pinecone, and Next.js
Multimodal model for text and tabular data with HuggingFace transformers as building block for text data
Fast Inference Solutions for BLOOM
[EMNLP 2022] Unifying and multi-tasking structured knowledge grounding with language models
Exact structure out of any language model completion.
Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed
Low latency JSON generation using LLMs ⚡️
Add a description, image, and links to the huggingface-transformers topic page so that developers can more easily learn about it.
To associate your repository with the huggingface-transformers topic, visit your repo's landing page and select "manage topics."