🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
-
Updated
May 29, 2024 - Python
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
An animal can do training and inference every day of its existence until the day of its death. A forward pass is all you need.
Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper
ChatGPT CLI is an advanced command-line interface for ChatGPT models via OpenAI and Azure, offering streaming, query mode, and history tracking for seamless, context-aware conversations. Ideal for both users and developers, it provides advanced configuration and easy setup options to ensure a tailored conversational experience with the GPT model.
An AI agent for your Github project that can search and manipulate the code base, browse the internet and interact with Github issues and pull requests
Build AI-powered applications with React, Svelte, Vue, and Solid
たまに追加される論文メモ
Mitigating a language model's over-confidence with NLI predictions on Multi-NLI hypotheses with random word order using PAWS (paraphrase) and Winogrande (anaphora).
Universal LLM Deployment Engine with ML Compilation
LangSmith Client SDK Implementations
Official code for paper "UniIR: Training and Benchmarking Universal Multimodal Information Retrievers"
The official gpt4free repository | various collection of powerful language models
Language modeling and instruction tuning for Russian
An implementation of the GPT(generative pretrained transformer) model, from scratch, which produces Shakespearean text by training on the dialogues written by Shakespeare along with the GPT Encoder.
A framework for few-shot evaluation of language models.
Unify Efficient Fine-Tuning of 100+ LLMs
🔍 LLM orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
Code the ICML 2024 paper: "Variance-reduced Zeroth-Order Methods for Fine-Tuning Language Models"
Latency and Memory Analysis of Transformer Models for Training and Inference
Add a description, image, and links to the language-model topic page so that developers can more easily learn about it.
To associate your repository with the language-model topic, visit your repo's landing page and select "manage topics."