>_Skillful
Need help with advanced AI agent engineering?Contact FirmAdapt

Agents

2,229

Autonomous AI agents that perform tasks independently

Sort:
Showing packages with source repositories & descriptions only
Category:

Transformer Engine

A library for accelerating Transformer model training on NVIDIA GPUs.

AgentLLM Training Frameworks
1 dir

OpenRLHF

An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT).

AgentLLM Training Frameworks
1 dir

unslothai

A framework that specializes in efficient fine-tuning. On its GitHub page, you can find ready-to-use fine-tuning templates for various LLMs, allowing you to easily train your own data for free on the Google Colab cloud.

AgentLLM Training Frameworks
1 dir

Axolotl

Open-source framework for fine-tuning and evaluating LLMs. It simplifies the process of experimenting with different training configurations and makes it easy to reproduce and share results, supporting features like LoRA, QLoRA, DeepSpeed, PEFT, and multi-GPU setups.

AgentLLM Training Frameworks
1 dir

SGLang

SGLang is a fast serving framework for large language models and vision language models.

AgentLLM Inference
1 dir

TensorRT-LLM

Nvidia Framework for LLM Inference

AgentLLM Inference
1 dir

FasterTransformer

NVIDIA Framework for LLM Inference(Transitioned to TensorRT-LLM)

AgentLLM Inference
1 dir

MInference

To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention, which reduces inference latency by up to 10x for pre-filling on an A100 while maintaining accuracy.

AgentLLM Inference
1 dir

exllama

A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.

AgentLLM Inference
1 dir

mistral.rs

Blazingly fast LLM inference.

AgentLLM Inference
1 dir

SkyPilot

Run LLMs and batch jobs on any cloud. Get maximum cost savings, highest GPU availability, and managed execution -- all with a simple interface.

AgentLLM Inference
1 dir

DeepSpeed-Mii

MII makes low-latency and high-throughput inference, similar to vLLM powered by DeepSpeed.

AgentLLM Inference
1 dir

Text-Embeddings-Inference

Inference for text-embeddings in Rust, HFOIL Licence.

AgentLLM Inference
1 dir

Infinity

Inference for text-embeddings in Python

AgentLLM Inference
1 dir

LMDeploy

A high-throughput and low-latency inference and serving framework for LLMs and VLs

AgentLLM Inference
1 dir

Liger-Kernel

Efficient Triton Kernels for LLM Training.

AgentLLM Inference
1 dir

prima.cpp

A distributed implementation of llama.cpp that lets you run 70B-level LLMs on your everyday devices.

AgentLLM Inference
1 dir

deploy-llms-with-ansible

Easily deploy any LLM on a VM with minimal configuration, using Ansible.

AgentLLM Inference
1 dir

Swiss Army Llama

Comprehensive set of tools for working with local LLMs for various tasks.

AgentLLM Applications
1 dir

wechat-chatgpt

Use ChatGPT On Wechat via wechaty

AgentLLM Applications
1 dir

Serge

a chat interface crafted with llama.cpp for running Alpaca models. No API keys, entirely self-hosted!

AgentLLM Applications
1 dir

IntelliServer

simplifies the evaluation of LLMs by providing a unified microservice to access and test multiple AI models.

AgentLLM Applications
1 dir

Search with Lepton

Build your own conversational search engine using less than 500 lines of code by [LeptonAI](https://github.com/leptonai).

AgentLLM Applications
1 dir

Robocorp

Create, deploy and operate Actions using Python anywhere to enhance your AI agents and assistants. Batteries included with an extensive set of libraries, helpers and logging.

AgentLLM Applications
1 dir