mirror of
https://www.modelscope.cn/openai-mirror/gpt-oss-20b.git
synced 2026-04-02 18:12:56 +08:00
Update chat_template.jinja (#94)
- Update chat_template.jinja (6965815ae896176f6b820fd508d15ebbca89a903) Co-authored-by: Quentin Gallouédec <qgallouedec@users.noreply.huggingface.co>
This commit is contained in:
@ -22,7 +22,7 @@ tags:
|
|||||||
Welcome to the gpt-oss series, [OpenAI’s open-weight models](https://openai.com/open-models) designed for powerful reasoning, agentic tasks, and versatile developer use cases.
|
Welcome to the gpt-oss series, [OpenAI’s open-weight models](https://openai.com/open-models) designed for powerful reasoning, agentic tasks, and versatile developer use cases.
|
||||||
|
|
||||||
We’re releasing two flavors of these open models:
|
We’re releasing two flavors of these open models:
|
||||||
- `gpt-oss-120b` — for production, general purpose, high reasoning use cases that fit into a single H100 GPU (117B parameters with 5.1B active parameters)
|
- `gpt-oss-120b` — for production, general purpose, high reasoning use cases that fit into a single 80GB GPU (like NVIDIA H100 or AMD MI300X) (117B parameters with 5.1B active parameters)
|
||||||
- `gpt-oss-20b` — for lower latency, and local or specialized use cases (21B parameters with 3.6B active parameters)
|
- `gpt-oss-20b` — for lower latency, and local or specialized use cases (21B parameters with 3.6B active parameters)
|
||||||
|
|
||||||
Both models were trained on our [harmony response format](https://github.com/openai/harmony) and should only be used with the harmony format as it will not work correctly otherwise.
|
Both models were trained on our [harmony response format](https://github.com/openai/harmony) and should only be used with the harmony format as it will not work correctly otherwise.
|
||||||
@ -38,7 +38,7 @@ Both models were trained on our [harmony response format](https://github.com/ope
|
|||||||
* **Full chain-of-thought:** Gain complete access to the model’s reasoning process, facilitating easier debugging and increased trust in outputs. It’s not intended to be shown to end users.
|
* **Full chain-of-thought:** Gain complete access to the model’s reasoning process, facilitating easier debugging and increased trust in outputs. It’s not intended to be shown to end users.
|
||||||
* **Fine-tunable:** Fully customize models to your specific use case through parameter fine-tuning.
|
* **Fine-tunable:** Fully customize models to your specific use case through parameter fine-tuning.
|
||||||
* **Agentic capabilities:** Use the models’ native capabilities for function calling, [web browsing](https://github.com/openai/gpt-oss/tree/main?tab=readme-ov-file#browser), [Python code execution](https://github.com/openai/gpt-oss/tree/main?tab=readme-ov-file#python), and Structured Outputs.
|
* **Agentic capabilities:** Use the models’ native capabilities for function calling, [web browsing](https://github.com/openai/gpt-oss/tree/main?tab=readme-ov-file#browser), [Python code execution](https://github.com/openai/gpt-oss/tree/main?tab=readme-ov-file#python), and Structured Outputs.
|
||||||
* **Native MXFP4 quantization:** The models are trained with native MXFP4 precision for the MoE layer, making `gpt-oss-120b` run on a single H100 GPU and the `gpt-oss-20b` model run within 16GB of memory.
|
* **Native MXFP4 quantization:** The models are trained with native MXFP4 precision for the MoE layer, making `gpt-oss-120b` run on a single 80GB GPU (like NVIDIA H100 or AMD MI300X) and the `gpt-oss-20b` model run within 16GB of memory.
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
|
|||||||
@ -245,9 +245,9 @@
|
|||||||
{%- if developer_message %}
|
{%- if developer_message %}
|
||||||
{{- "# Instructions\n\n" }}
|
{{- "# Instructions\n\n" }}
|
||||||
{{- developer_message }}
|
{{- developer_message }}
|
||||||
|
{{- "\n\n" }}
|
||||||
{%- endif %}
|
{%- endif %}
|
||||||
{%- if tools -%}
|
{%- if tools -%}
|
||||||
{{- "\n\n" }}
|
|
||||||
{{- "# Tools\n\n" }}
|
{{- "# Tools\n\n" }}
|
||||||
{{- render_tool_namespace("functions", tools) }}
|
{{- render_tool_namespace("functions", tools) }}
|
||||||
{%- endif -%}
|
{%- endif -%}
|
||||||
|
|||||||
Reference in New Issue
Block a user