mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2024-10-01 01:06:10 -04:00
translations: fix typos in Traditional Chinese (#2852)
Signed-off-by: Shiranui <supersonic@livemail.tw> Signed-off-by: Jared Van Bortel <jared@nomic.ai> Co-authored-by: Jared Van Bortel <jared@nomic.ai>
This commit is contained in:
parent
32b56e819d
commit
ace79959d1
@ -12,6 +12,7 @@ The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/).
|
|||||||
### Fixed
|
### Fixed
|
||||||
- Bring back "Auto" option for Embeddings Device as "Application default," which went missing in v3.1.0 ([#2873](https://github.com/nomic-ai/gpt4all/pull/2873))
|
- Bring back "Auto" option for Embeddings Device as "Application default," which went missing in v3.1.0 ([#2873](https://github.com/nomic-ai/gpt4all/pull/2873))
|
||||||
- Correct a few strings in the Italian translation (by [@Harvester62](https://github.com/Harvester62) in [#2872](https://github.com/nomic-ai/gpt4all/pull/2872))
|
- Correct a few strings in the Italian translation (by [@Harvester62](https://github.com/Harvester62) in [#2872](https://github.com/nomic-ai/gpt4all/pull/2872))
|
||||||
|
- Correct typos in Traditional Chinese translation (by [@supersonictw](https://github.com/supersonictw) in [#2852](https://github.com/nomic-ai/gpt4all/pull/2852))
|
||||||
|
|
||||||
## [3.2.1] - 2024-08-13
|
## [3.2.1] - 2024-08-13
|
||||||
|
|
||||||
|
@ -612,7 +612,7 @@
|
|||||||
<location filename="../qml/ApplicationSettings.qml" line="224"/>
|
<location filename="../qml/ApplicationSettings.qml" line="224"/>
|
||||||
<location filename="../../build_gpt4all-chat_Desktop_Qt_6_7_2/gpt4all/qml/ApplicationSettings.qml" line="224"/>
|
<location filename="../../build_gpt4all-chat_Desktop_Qt_6_7_2/gpt4all/qml/ApplicationSettings.qml" line="224"/>
|
||||||
<source>The compute device used for text generation.</source>
|
<source>The compute device used for text generation.</source>
|
||||||
<translation>用於生成文字的計算設備。</translation>
|
<translation>用於生成文字的計算裝置。</translation>
|
||||||
</message>
|
</message>
|
||||||
<message>
|
<message>
|
||||||
<location filename="../qml/ApplicationSettings.qml" line="242"/>
|
<location filename="../qml/ApplicationSettings.qml" line="242"/>
|
||||||
@ -1785,7 +1785,7 @@ model to get started</source>
|
|||||||
<message>
|
<message>
|
||||||
<location filename="../modellist.cpp" line="1300"/>
|
<location filename="../modellist.cpp" line="1300"/>
|
||||||
<source><strong>OpenAI-Compatible API Model</strong><br><ul><li>API Key: %1</li><li>Base URL: %2</li><li>Model Name: %3</li></ul></source>
|
<source><strong>OpenAI-Compatible API Model</strong><br><ul><li>API Key: %1</li><li>Base URL: %2</li><li>Model Name: %3</li></ul></source>
|
||||||
<translation><strong>OpenAI 相容 API 模型</strong><br><ul><li>API 金鑰:%1</li><li>Base URL: %2</li><li>模型名稱: %3</li></ul></translation>
|
<translation><strong>OpenAI API 相容模型</strong><br><ul><li>API 金鑰:%1</li><li>基底 URL:%2</li><li>模型名稱:%3</li></ul></translation>
|
||||||
</message>
|
</message>
|
||||||
<message>
|
<message>
|
||||||
<location filename="../modellist.cpp" line="1559"/>
|
<location filename="../modellist.cpp" line="1559"/>
|
||||||
@ -1830,17 +1830,17 @@ model to get started</source>
|
|||||||
<message>
|
<message>
|
||||||
<location filename="../modellist.cpp" line="1700"/>
|
<location filename="../modellist.cpp" line="1700"/>
|
||||||
<source><ul><li>Requires personal API key and the API base URL.</li><li>WARNING: Will send your chats to the OpenAI-compatible API Server you specified!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with the OpenAI-compatible API Server</li></source>
|
<source><ul><li>Requires personal API key and the API base URL.</li><li>WARNING: Will send your chats to the OpenAI-compatible API Server you specified!</li><li>Your API key will be stored on disk</li><li>Will only be used to communicate with the OpenAI-compatible API Server</li></source>
|
||||||
<translation><ul><li>需要個人的 API 金鑰和 API 的 base URL。</li><li>警告:這將會傳送您的交談紀錄到您所指定的 OpenAI 相容性 API 伺服器</li><li>您的 API 金鑰將被儲存在硬碟上</li><li>它只被用於與其 OpenAI 相容性 API 伺服器進行通訊</li></translation>
|
<translation><ul><li>需要個人的 API 金鑰和 API 的基底 URL(Base URL)。</li><li>警告:這將會傳送您的交談紀錄到您所指定的 OpenAI API 相容伺服器</li><li>您的 API 金鑰將被儲存在硬碟上</li><li>它只被用於與其 OpenAI API 相容伺服器進行通訊</li></translation>
|
||||||
</message>
|
</message>
|
||||||
<message>
|
<message>
|
||||||
<location filename="../modellist.cpp" line="1717"/>
|
<location filename="../modellist.cpp" line="1717"/>
|
||||||
<source><strong>Connect to OpenAI-compatible API server</strong><br> %1</source>
|
<source><strong>Connect to OpenAI-compatible API server</strong><br> %1</source>
|
||||||
<translation><strong>連線到 OpenAI API 相容性伺服器</strong><br> %1</translation>
|
<translation><strong>連線到 OpenAI API 相容伺服器</strong><br> %1</translation>
|
||||||
</message>
|
</message>
|
||||||
<message>
|
<message>
|
||||||
<location filename="../modellist.cpp" line="2131"/>
|
<location filename="../modellist.cpp" line="2131"/>
|
||||||
<source><strong>Created by %1.</strong><br><ul><li>Published on %2.<li>This model has %3 likes.<li>This model has %4 downloads.<li>More info can be found <a href="https://huggingface.co/%5">here.</a></ul></source>
|
<source><strong>Created by %1.</strong><br><ul><li>Published on %2.<li>This model has %3 likes.<li>This model has %4 downloads.<li>More info can be found <a href="https://huggingface.co/%5">here.</a></ul></source>
|
||||||
<translation><strong>建立者:%1。</strong><br><ul><li>發行於:%2。<li>這個模型有 %3 個讚。<li>這個模型有 %4 次下載次數。<li>更多資訊請查閱<a href="https://huggingface.co/%5">此處</a>。</ul></translation>
|
<translation><strong>模型作者:%1</strong><br><ul><li>發佈日期:%2<li>累積讚數:%3 個讚<li>下載次數:%4 次<li>更多資訊請查閱<a href="https://huggingface.co/%5">此處</a>。</ul></translation>
|
||||||
</message>
|
</message>
|
||||||
</context>
|
</context>
|
||||||
<context>
|
<context>
|
||||||
@ -2803,7 +2803,7 @@ Nomic AI 將保留附加在您的資料上的所有署名訊息,並且您將
|
|||||||
<location filename="../main.qml" line="111"/>
|
<location filename="../main.qml" line="111"/>
|
||||||
<location filename="../../build_gpt4all-chat_Desktop_Qt_6_7_2/gpt4all/main.qml" line="111"/>
|
<location filename="../../build_gpt4all-chat_Desktop_Qt_6_7_2/gpt4all/main.qml" line="111"/>
|
||||||
<source><h3>Encountered an error starting up:</h3><br><i>"Incompatible hardware detected."</i><br><br>Unfortunately, your CPU does not meet the minimal requirements to run this program. In particular, it does not support AVX intrinsics which this program requires to successfully run a modern large language model. The only solution at this time is to upgrade your hardware to a more modern CPU.<br><br>See here for more information: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a></source>
|
<source><h3>Encountered an error starting up:</h3><br><i>"Incompatible hardware detected."</i><br><br>Unfortunately, your CPU does not meet the minimal requirements to run this program. In particular, it does not support AVX intrinsics which this program requires to successfully run a modern large language model. The only solution at this time is to upgrade your hardware to a more modern CPU.<br><br>See here for more information: <a href="https://en.wikipedia.org/wiki/Advanced_Vector_Extensions">https://en.wikipedia.org/wiki/Advanced_Vector_Extensions</a></source>
|
||||||
<translation><h3>啟動時發生錯誤:</h3><br><i>「偵測到不相容的硬體。」</i><br><br>糟糕!您的中央處理器不符合運行所需的最低需求。尤其,它不支援本程式運行現代大型語言模型所需的 AVX 指令集。目前唯一的解決方案,只有更新您的中央處理器及其相關硬體設備。<br><br>更多資訊請查閱:<a href="https://zh.wikipedia.org/wiki/AVX指令集">AVX 指令集 - 維基百科</a></translation>
|
<translation><h3>啟動時發生錯誤:</h3><br><i>「偵測到不相容的硬體。」</i><br><br>糟糕!您的中央處理器不符合運行所需的最低需求。尤其,它不支援本程式運行現代大型語言模型所需的 AVX 指令集。目前唯一的解決方案,只有更新您的中央處理器及其相關硬體裝置。<br><br>更多資訊請查閱:<a href="https://zh.wikipedia.org/wiki/AVX指令集">AVX 指令集 - 維基百科</a></translation>
|
||||||
</message>
|
</message>
|
||||||
<message>
|
<message>
|
||||||
<location filename="../main.qml" line="127"/>
|
<location filename="../main.qml" line="127"/>
|
||||||
|
Loading…
Reference in New Issue
Block a user