Sourced from shiki's releases.
v3.13.0
🚀 Features
- transformers: Render indent guides - by
@KazariEX
and@antfu
in shikijs/shiki#1060 (aecd1)View changes on GitHub
v3.12.3
🐞 Bug Fixes
@shikijs/twoslash
version specifier - by@9romise
in shikijs/shiki#1078 (a1cde)View changes on GitHub
v3.12.2
🐞 Bug Fixes
- twoslash: Fix
onTwoslashError
return value handling - by@Karibash
in shikijs/shiki#1070 (e86b0)View changes on GitHub
v3.12.1
No significant changes
View changes on GitHub
v3.12.0
🚀 Features
- vitepress-twoslash:
- Improve UX for option customization - by
@9romise
in shikijs/shiki#1066 (e3cfd)- Twoslash inline type cache for markdown - by
@serkodev
and@antfu
in shikijs/shiki#1063 (dc7fb)🐞 Bug Fixes
- remove-notation-escape: Correct escape sequence - by
@sor4chi
in shikijs/shiki#1065 (22d0c)View changes on GitHub
v3.11.0
🚀 Features
- core: Add
enforce
options toShikiTransformer
- by@serkodev
and@antfu
in shikijs/shiki#1062 (8ad05)View changes on GitHub
v3.10.0
🚀 Features
- Add funding links to playground - by
@jtbandes
in shikijs/shiki#1054 (e36eb)
... (truncated)
fd7326a
chore: release v3.13.05cbb052
chore: release v3.12.3e462618
chore: release v3.12.2793d71e
chore: release v3.12.19260f3f
chore: release v3.12.0d05f39b
chore: release v3.11.0bda1a76
chore: release v3.10.009921f1
chore: release v3.9.2854eddf
chore: release v3.9.1950ede5
chore: release v3.9.0This version was pushed to npm by [GitHub Actions](https://www.npmjs.com/~GitHub Actions), a new releaser for shiki since your current version.
Sourced from tw-animate-css's releases.
v1.4.0
Changelog
902e37a019ffd165ba078e0b3c02634526c54bf0: fix: remove support for prefix, add new export for prefixed version. Closes #58. fab2a5bf817605be1976e159976718a83489fc1c: chore: bump version to 1.4.0 and update dependencies c20dc32e2b532a8e74546879b4ce7d9ce89ba710: fix(build): make transform.ts accept two arguments
⚠️ BREAKING CHANGE ⚠️
Support for Tailwind CSS's prefix option was moved to
tw-animate-css/prefix
because it was breaking the--spacing
function. Users requiring prefixes should replace their import:- import "tw-animate-css"; + import "tw-animate-css/prefix";
I do not plan to introduce breaking changes like this to non-major releases in the future. But because more people use spacing rather than prefixes, reverting the previous version's (obviously breaking) change seems reasonable.
v1.3.8
Changelog
- b5ff23a: fix: add support for global CSS variable prefix. Closes #48
- 03e5f12: feat: add support for ng-primitives height variables #56 (thanks
@immohammadjaved
)- b076cfb: docs: fix various issues in accordion and collapsible docs
- 9485e33: chore: bump version to 1.3.8 and update dependencies
⚠️ BREAKING CHANGE ⚠️
Adding support for prefixes broke custom spacing. It is recommended that you skip this version if you do not use Tailwind CSS's prefix option, and use v1.4.0 instead. If you are actually using prefixes, you can use a special version supporting prefixes:
- import "tw-animate-css"; /* Version with spacing support */ + import "tw-animate-css/prefix"; /* Version with prefix support */
I do not plan to fix the incompatibility between the spacing and prefix versions due to time constraints. Feel free to investigate and open a pull request if you manage to fix it.
v1.3.7
Changelog
- 80dbfcc: feat: add utilities for blur transitions #54 (thanks
@coffeeispower
)- dc294f9: docs: add upcoming changes warning
- c640bb8: chore: update dependencies and package manager version
- 9e63e34: chore: bump version to 1.3.7
v1.3.6
Changelog
- 58f3396: fix: allow changing animation parameters for ready-to-use animations
- 8313476: chore: update dependencies nd package manager version
- f81346c: chore: bump version to 1.3.6
... (truncated)
c20dc32
fix(build): make transform.ts accept two argumentsfab2a5b
chore: bump version to 1.4.0 and update dependencies902e37a
fix: remove support for prefix, add new export for prefixed version9485e33
chore: bump version to 1.3.8 and update dependenciesb076cfb
docs: fix various issues in accordion and collapsible docs03e5f12
feat: add support for ng-primitives height variables (#56)b5ff23a
fix: add support for global CSS variable prefix. Closes #489e63e34
chore: bump version to 1.3.7c640bb8
chore: update dependencies and package manager versiondc294f9
docs: add upcoming changes warningSourced from actions/cache's releases.
v4.3.0
What's Changed
- Add note on runner versions by
@GhadimiR
in actions/cache#1642- Prepare
v4.3.0
release by@Link
- in actions/cache#1655New Contributors
@GhadimiR
made their first contribution in actions/cache#1642Full Changelog: https://github.com/actions/cache/compare/v4...v4.3.0
Sourced from actions/cache's changelog.
Releases
4.3.0
- Bump
@actions/cache
to v4.1.04.2.4
- Bump
@actions/cache
to v4.0.54.2.3
- Bump
@actions/cache
to v4.0.3 (obfuscates SAS token in debug logs for cache entries)4.2.2
- Bump
@actions/cache
to v4.0.24.2.1
- Bump
@actions/cache
to v4.0.14.2.0
TLDR; The cache backend service has been rewritten from the ground up for improved performance and reliability. actions/cache now integrates with the new cache service (v2) APIs.
The new service will gradually roll out as of February 1st, 2025. The legacy service will also be sunset on the same date. Changes in these release are fully backward compatible.
We are deprecating some versions of this action. We recommend upgrading to version
v4
orv3
as soon as possible before February 1st, 2025. (Upgrade instructions below).If you are using pinned SHAs, please use the SHAs of versions
v4.2.0
orv3.4.0
If you do not upgrade, all workflow runs using any of the deprecated actions/cache will fail.
Upgrading to the recommended versions will not break your workflows.
4.1.2
- Add GitHub Enterprise Cloud instances hostname filters to inform API endpoint choices - #1474
- Security fix: Bump braces from 3.0.2 to 3.0.3 - #1475
4.1.1
- Restore original behavior of
cache-hit
output - #14674.1.0
... (truncated)
client.inference.chat_completion(\*\*params) -> InferenceChatCompletionResponse
-- client.inference.completion(\*\*params) -> InferenceCompletionResponse
- client.inference.embeddings(\*\*params) -> EmbeddingsResponse
## VectorIo
diff --git a/docs/getting_started.ipynb b/docs/getting_started.ipynb
index 449bd2be1..56aef2b7d 100644
--- a/docs/getting_started.ipynb
+++ b/docs/getting_started.ipynb
@@ -824,16 +824,10 @@
"\n",
"\n",
"user_input = \"Michael Jordan was born in 1963. He played basketball for the Chicago Bulls. He retired in 2003. Extract this information into JSON for me. \"\n",
- "response = client.inference.completion(\n",
- " model_id=\"meta-llama/Llama-3.1-8B-Instruct\",\n",
- " content=user_input,\n",
- " stream=False,\n",
- " sampling_params={\n",
- " \"strategy\": {\n",
- " \"type\": \"greedy\",\n",
- " },\n",
- " \"max_tokens\": 50,\n",
- " },\n",
+ "response = client.chat.completions.create(\n",
+ " model=\"meta-llama/Llama-3.1-8B-Instruct\",\n",
+ " messages=[{\"role\": \"user\", \"content\": user_input}],\n",
+ " max_tokens=50,\n",
" response_format={\n",
" \"type\": \"json_schema\",\n",
" \"json_schema\": Output.model_json_schema(),\n",
diff --git a/docs/notebooks/nvidia/beginner_e2e/Llama_Stack_NVIDIA_E2E_Flow.ipynb b/docs/notebooks/nvidia/beginner_e2e/Llama_Stack_NVIDIA_E2E_Flow.ipynb
index d8f29d999..601276526 100644
--- a/docs/notebooks/nvidia/beginner_e2e/Llama_Stack_NVIDIA_E2E_Flow.ipynb
+++ b/docs/notebooks/nvidia/beginner_e2e/Llama_Stack_NVIDIA_E2E_Flow.ipynb
@@ -706,20 +706,15 @@
" provider_id=\"nvidia\",\n",
")\n",
"\n",
- "response = client.inference.completion(\n",
- " content=\"Complete the sentence using one word: Roses are red, violets are \",\n",
+ "response = client.completions.create(\n",
+ " prompt=\"Complete the sentence using one word: Roses are red, violets are \",\n",
" stream=False,\n",
- " model_id=CUSTOMIZED_MODEL_DIR,\n",
- " sampling_params={\n",
- " \"strategy\": {\n",
- " \"type\": \"top_p\",\n",
- " \"temperature\": 0.7,\n",
- " \"top_p\": 0.9\n",
- " },\n",
- " \"max_tokens\": 20,\n",
- " },\n",
+ " model=CUSTOMIZED_MODEL_DIR,\n",
+ " temperature=0.7,\n",
+ " top_p=0.9,\n",
+ " max_tokens=20,\n",
")\n",
- "print(f\"Inference response: {response.content}\")"
+ "print(f\"Inference response: {response.choices[0].text}\")"
]
},
{
@@ -1233,20 +1228,15 @@
" provider_id=\"nvidia\",\n",
")\n",
"\n",
- "response = client.inference.completion(\n",
- " content=\"Complete the sentence using one word: Roses are red, violets are \",\n",
+ "response = client.completions.create(\n",
+ " prompt=\"Complete the sentence using one word: Roses are red, violets are \",\n",
" stream=False,\n",
- " model_id=customized_chat_model_dir,\n",
- " sampling_params={\n",
- " \"strategy\": {\n",
- " \"type\": \"top_p\",\n",
- " \"temperature\": 0.7,\n",
- " \"top_p\": 0.9\n",
- " },\n",
- " \"max_tokens\": 20,\n",
- " },\n",
+ " model=customized_chat_model_dir,\n",
+ " temperature=0.7,\n",
+ " top_p=0.9,\n",
+ " max_tokens=20,\n",
")\n",
- "print(f\"Inference response: {response.content}\")"
+ "print(f\"Inference response: {response.choices[0].text}\")"
]
},
{
diff --git a/llama_stack/providers/remote/inference/nvidia/NVIDIA.md b/llama_stack/providers/remote/inference/nvidia/NVIDIA.md
index d9c18533a..4cb2dc394 100644
--- a/llama_stack/providers/remote/inference/nvidia/NVIDIA.md
+++ b/llama_stack/providers/remote/inference/nvidia/NVIDIA.md
@@ -39,25 +39,6 @@ client = LlamaStackAsLibraryClient("nvidia")
client.initialize()
```
-### Create Completion
-
-The following example shows how to create a completion for an NVIDIA NIM.
-
-> [!NOTE]
-> The hosted NVIDIA Llama NIMs (for example ```meta-llama/Llama-3.1-8B-Instruct```) that have ```NVIDIA_BASE_URL="https://integrate.api.nvidia.com"``` do not support the ```completion``` method, while locally deployed NIMs do.
-
-```python
-response = client.inference.completion(
- model_id="meta-llama/Llama-3.1-8B-Instruct",
- content="Complete the sentence using one word: Roses are red, violets are :",
- stream=False,
- sampling_params={
- "max_tokens": 50,
- },
-)
-print(f"Response: {response.content}")
-```
-
### Create Chat Completion
The following example shows how to create a chat completion for an NVIDIA NIM.
diff --git a/llama_stack/providers/remote/post_training/nvidia/README.md b/llama_stack/providers/remote/post_training/nvidia/README.md
index 6647316df..9b088a615 100644
--- a/llama_stack/providers/remote/post_training/nvidia/README.md
+++ b/llama_stack/providers/remote/post_training/nvidia/README.md
@@ -140,13 +140,11 @@ client.models.register(
#### 2. Inference with the fine-tuned model
```python
-response = client.inference.completion(
- content="Complete the sentence using one word: Roses are red, violets are ",
+response = client.completions.create(
+ prompt="Complete the sentence using one word: Roses are red, violets are ",
stream=False,
- model_id="test-example-model@v1",
- sampling_params={
- "max_tokens": 50,
- },
+ model="test-example-model@v1",
+ max_tokens=50,
)
-print(response.content)
+print(response.choices[0].text)
```
diff --git a/tests/integration/README.md b/tests/integration/README.md
index 467f97e02..b68526410 100644
--- a/tests/integration/README.md
+++ b/tests/integration/README.md
@@ -178,10 +178,10 @@ Note that when re-recording tests, you must use a Stack pointing to a server (i.
### Basic Test Pattern
```python
-def test_basic_completion(llama_stack_client, text_model_id):
- response = llama_stack_client.inference.completion(
+def test_basic_chat_completion(llama_stack_client, text_model_id):
+ response = llama_stack_client.inference.chat_completion(
model_id=text_model_id,
- content=CompletionMessage(role="user", content="Hello"),
+ messages=[{"role": "user", "content": "Hello"}],
)
# Test structure, not AI output quality