A high-throughput and memory-efficient inference and serving engine for LLMs
-
Updated
Sep 1, 2025 - Python
A high-throughput and memory-efficient inference and serving engine for LLMs
Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.
Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, 17+ clouds, or on-prem).
Fast and flexible AutoML with learning guarantees.
Everything we actually know about the Apple Neural Engine (ANE)
GPT2 for Multiple Languages, including pretrained models. GPT2 多语言支持, 15亿参数中文预训练模型
Large-scale LLM inference engine
Elegant and Performant Deep Learning
Everything you want to know about Google Cloud TPU
Implementation of a Tensor Processing Unit for embedded systems and the IoT.
Neural network-based chess engine capable of natural language commentary
Differentiable Fluid Dynamics Package
JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
Dual Edge TPU Adapter to use it on a system with single PCIe port on m.2 A/B/E/M slot
DECIMER Image Transformer is a deep-learning-based tool designed for automated recognition of chemical structure images. Leveraging transformer architectures, the model converts chemical images into SMILES strings, enabling the digitization of chemical data from scanned documents, literature, and patents.
Free TPU for FPGA with compiler supporting Pytorch/Caffe/Darknet/NCNN. An AI processor for using Xilinx FPGA to solve image classification, detection, and segmentation problem.
Julia on TPUs
Small-scale Tensor Processing Unit built on an FPGA
Benchmarking suite to evaluate 🤖 robotics computing performance. Vendor-neutral. ⚪Grey-box and ⚫Black-box approaches.
Add a description, image, and links to the tpu topic page so that developers can more easily learn about it.
To associate your repository with the tpu topic, visit your repo's landing page and select "manage topics."