- Seoul, Republic of Korea
- sooftware.io
- sooftware95
- in/Soo-hwan
Highlights
- Pro
Block or Report
Block or report sooftware
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseLists (2)
Sort Name ascending (A-Z)
Stars
Language: C++
Sort by: Most stars
An Open Source Machine Learning Framework for Everyone
Carbon Language's main repository: documents, design, implementation, and related tools. (NOTE: Carbon Language is experimental; see README)
A library for efficient similarity search and clustering of dense vectors.
DeepSpeech is an open source embedded (offline, on-device) speech-to-text engine which can run in real time on devices ranging from a Raspberry Pi 4 to high power GPU servers.
Karabiner-Elements is a powerful utility for keyboard customization on macOS Sierra (10.12) or later.
Apache Arrow is a multi-language toolbox for accelerated data interchange and in-memory processing
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
Development repository for the Triton language and compiler
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
Unsupervised text tokenizer for Neural Network-based text generation.
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficie…
Facebook AI Research's Automatic Speech Recognition Toolkit
Transformer related optimization, including BERT, GPT
A C++ standalone library for machine learning
A GPU-accelerated library containing highly optimized building blocks and an execution engine for data processing to accelerate deep learning training and inference applications.
LightSeq: A High Performance Library for Sequence Processing and Generation
Optimized primitives for collective multi-GPU communication
A machine learning compiler for GPUs, CPUs, and ML accelerators
FlexFlow Serve: Low-Latency, High-Performance LLM Serving
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
Fast implementation of the edit distance(Levenshtein distance)
Modern spell checking library - accurate, fast, multi-language