🌐 WebThinker: Empowering Large Reasoning Models with Deep Research Capability
-
Updated
Jul 30, 2025 - Python
🌐 WebThinker: Empowering Large Reasoning Models with Deep Research Capability
Higher performance OpenAI LLM service than vLLM serve: A pure C++ high-performance OpenAI LLM service implemented with GPRS+TensorRT-LLM+Tokenizers.cpp, supporting chat and function call, AI agents, distributed multi-GPU inference, multimodal capabilities, and a Gradio chat interface.
Official PyTorch implementation for Hogwild! Inference: Parallel LLM Generation with a Concurrent Attention Cache
Deploy open-source LLMs on AWS in minutes — with OpenAI-compatible APIs and a powerful CLI/SDK toolkit.
To Think or Not to Think: Exploring the Unthinking Vulnerability in Large Reasoning Models
Breaking long thought processes of o1-like LLMs, such as DeepSeek-R1, QwQ
Add a description, image, and links to the qwq topic page so that developers can more easily learn about it.
To associate your repository with the qwq topic, visit your repo's landing page and select "manage topics."