|
31 | 31 | "execution_count": 1,
|
32 | 32 | "metadata": {
|
33 | 33 | "execution": {
|
34 |
| - "iopub.execute_input": "2025-10-06T03:00:32.007007Z", |
35 |
| - "iopub.status.busy": "2025-10-06T03:00:32.006876Z", |
36 |
| - "iopub.status.idle": "2025-10-06T03:01:13.019849Z", |
37 |
| - "shell.execute_reply": "2025-10-06T03:01:13.019316Z" |
| 34 | + "iopub.execute_input": "2025-10-06T03:37:49.994905Z", |
| 35 | + "iopub.status.busy": "2025-10-06T03:37:49.994771Z", |
| 36 | + "iopub.status.idle": "2025-10-06T03:38:40.631780Z", |
| 37 | + "shell.execute_reply": "2025-10-06T03:38:40.631244Z" |
38 | 38 | }
|
39 | 39 | },
|
40 | 40 | "outputs": [
|
|
76 | 76 | "output_type": "stream",
|
77 | 77 | "text": [
|
78 | 78 | "/usr/local/lib/python3.10/dist-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.\n",
|
79 |
| - " import pynvml # type: ignore[import]\n" |
80 |
| - ] |
81 |
| - }, |
82 |
| - { |
83 |
| - "name": "stdout", |
84 |
| - "output_type": "stream", |
85 |
| - "text": [ |
| 79 | + " import pynvml # type: ignore[import]\n", |
86 | 80 | "/usr/local/lib/python3.10/dist-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.\n",
|
87 | 81 | " import pynvml # type: ignore[import]\n"
|
88 | 82 | ]
|
|
92 | 86 | "output_type": "stream",
|
93 | 87 | "text": [
|
94 | 88 | "`torch_dtype` is deprecated! Use `dtype` instead!\n",
|
95 |
| - "[2025-10-06 03:01:01] `torch_dtype` is deprecated! Use `dtype` instead!\n" |
| 89 | + "[2025-10-06 03:38:25] `torch_dtype` is deprecated! Use `dtype` instead!\n" |
96 | 90 | ]
|
97 | 91 | },
|
98 | 92 | {
|
|
102 | 96 | "[Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0\n",
|
103 | 97 | "[Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0\n",
|
104 | 98 | "[Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0\n",
|
105 |
| - "[Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0\n" |
106 |
| - ] |
107 |
| - }, |
108 |
| - { |
109 |
| - "name": "stdout", |
110 |
| - "output_type": "stream", |
111 |
| - "text": [ |
112 |
| - "[2025-10-06 03:01:02] MOE_RUNNER_BACKEND is not initialized, using triton backend\n" |
| 99 | + "[Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0\n", |
| 100 | + "[2025-10-06 03:38:29] MOE_RUNNER_BACKEND is not initialized, using triton backend\n" |
113 | 101 | ]
|
114 | 102 | },
|
115 | 103 | {
|
116 | 104 | "name": "stdout",
|
117 | 105 | "output_type": "stream",
|
118 | 106 | "text": [
|
119 | 107 | "\r",
|
120 |
| - "Loading safetensors checkpoint shards: 0% Completed | 0/1 [00:00<?, ?it/s]\n" |
| 108 | + "Loading safetensors checkpoint shards: 0% Completed | 0/1 [00:00<?, ?it/s]\n", |
| 109 | + "\r", |
| 110 | + "Loading safetensors checkpoint shards: 100% Completed | 1/1 [00:00<00:00, 6.19it/s]\n", |
| 111 | + "\r", |
| 112 | + "Loading safetensors checkpoint shards: 100% Completed | 1/1 [00:00<00:00, 6.18it/s]\n", |
| 113 | + "\n" |
121 | 114 | ]
|
122 | 115 | },
|
123 | 116 | {
|
124 | 117 | "name": "stdout",
|
125 | 118 | "output_type": "stream",
|
126 | 119 | "text": [
|
127 | 120 | "\r",
|
128 |
| - "Loading safetensors checkpoint shards: 100% Completed | 1/1 [00:00<00:00, 4.20it/s]\n", |
129 |
| - "\r", |
130 |
| - "Loading safetensors checkpoint shards: 100% Completed | 1/1 [00:00<00:00, 4.20it/s]\n", |
131 |
| - "\n" |
| 121 | + " 0%| | 0/3 [00:00<?, ?it/s]\r", |
| 122 | + "Capturing batches (bs=4 avail_mem=71.53 GB): 0%| | 0/3 [00:00<?, ?it/s]" |
132 | 123 | ]
|
133 | 124 | },
|
134 | 125 | {
|
135 | 126 | "name": "stdout",
|
136 | 127 | "output_type": "stream",
|
137 | 128 | "text": [
|
138 | 129 | "\r",
|
139 |
| - " 0%| | 0/3 [00:00<?, ?it/s]\r", |
140 |
| - "Capturing batches (bs=4 avail_mem=77.03 GB): 0%| | 0/3 [00:00<?, ?it/s]" |
| 130 | + "Capturing batches (bs=4 avail_mem=71.53 GB): 33%|███▎ | 1/3 [00:00<00:00, 2.84it/s]\r", |
| 131 | + "Capturing batches (bs=2 avail_mem=71.46 GB): 33%|███▎ | 1/3 [00:00<00:00, 2.84it/s]" |
141 | 132 | ]
|
142 | 133 | },
|
143 | 134 | {
|
144 | 135 | "name": "stdout",
|
145 | 136 | "output_type": "stream",
|
146 | 137 | "text": [
|
147 | 138 | "\r",
|
148 |
| - "Capturing batches (bs=4 avail_mem=77.03 GB): 33%|███▎ | 1/3 [00:00<00:01, 1.53it/s]\r", |
149 |
| - "Capturing batches (bs=2 avail_mem=76.97 GB): 33%|███▎ | 1/3 [00:00<00:01, 1.53it/s]\r", |
150 |
| - "Capturing batches (bs=1 avail_mem=76.96 GB): 33%|███▎ | 1/3 [00:00<00:01, 1.53it/s]\r", |
151 |
| - "Capturing batches (bs=1 avail_mem=76.96 GB): 100%|██████████| 3/3 [00:00<00:00, 4.17it/s]\n" |
| 139 | + "Capturing batches (bs=2 avail_mem=71.46 GB): 67%|██████▋ | 2/3 [00:00<00:00, 3.80it/s]\r", |
| 140 | + "Capturing batches (bs=1 avail_mem=71.46 GB): 67%|██████▋ | 2/3 [00:00<00:00, 3.80it/s]\r", |
| 141 | + "Capturing batches (bs=1 avail_mem=71.46 GB): 100%|██████████| 3/3 [00:00<00:00, 5.01it/s]\n" |
152 | 142 | ]
|
153 | 143 | },
|
154 | 144 | {
|
|
167 | 157 | "name": "stdout",
|
168 | 158 | "output_type": "stream",
|
169 | 159 | "text": [
|
170 |
| - "Server started on http://localhost:31747\n" |
| 160 | + "Server started on http://localhost:30884\n" |
171 | 161 | ]
|
172 | 162 | }
|
173 | 163 | ],
|
|
201 | 191 | "execution_count": 2,
|
202 | 192 | "metadata": {
|
203 | 193 | "execution": {
|
204 |
| - "iopub.execute_input": "2025-10-06T03:01:13.024173Z", |
205 |
| - "iopub.status.busy": "2025-10-06T03:01:13.023033Z", |
206 |
| - "iopub.status.idle": "2025-10-06T03:01:13.307561Z", |
207 |
| - "shell.execute_reply": "2025-10-06T03:01:13.307079Z" |
| 194 | + "iopub.execute_input": "2025-10-06T03:38:40.633900Z", |
| 195 | + "iopub.status.busy": "2025-10-06T03:38:40.633563Z", |
| 196 | + "iopub.status.idle": "2025-10-06T03:38:40.897940Z", |
| 197 | + "shell.execute_reply": "2025-10-06T03:38:40.897465Z" |
208 | 198 | }
|
209 | 199 | },
|
210 | 200 | "outputs": [
|
211 | 201 | {
|
212 | 202 | "data": {
|
213 | 203 | "text/html": [
|
214 |
| - "<strong style='color: #00008B;'>Response: ChatCompletion(id='22b3d25954cb404ca70949629f098753', choices=[Choice(finish_reason='stop', index=0, logprobs=None, message=ChatCompletionMessage(content='Sure, here are three countries and their respective capitals:\\n\\n1. **United States** - Washington, D.C.\\n2. **Canada** - Ottawa\\n3. **Australia** - Canberra', refusal=None, role='assistant', annotations=None, audio=None, function_call=None, tool_calls=None, reasoning_content=None), matched_stop=151645)], created=1759719673, model='qwen/qwen2.5-0.5b-instruct', object='chat.completion', service_tier=None, system_fingerprint=None, usage=CompletionUsage(completion_tokens=39, prompt_tokens=37, total_tokens=76, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
| 204 | + "<strong style='color: #00008B;'>Response: ChatCompletion(id='1b89b2c731db480b8715ea001e218847', choices=[Choice(finish_reason='stop', index=0, logprobs=None, message=ChatCompletionMessage(content='Sure, here are three countries and their respective capitals:\\n\\n1. **United States** - Washington, D.C.\\n2. **Canada** - Ottawa\\n3. **Australia** - Canberra', refusal=None, role='assistant', annotations=None, audio=None, function_call=None, tool_calls=None, reasoning_content=None), matched_stop=151645)], created=1759721920, model='qwen/qwen2.5-0.5b-instruct', object='chat.completion', service_tier=None, system_fingerprint=None, usage=CompletionUsage(completion_tokens=39, prompt_tokens=37, total_tokens=76, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
215 | 205 | ],
|
216 | 206 | "text/plain": [
|
217 | 207 | "<IPython.core.display.HTML object>"
|
|
401 | 391 | "execution_count": 3,
|
402 | 392 | "metadata": {
|
403 | 393 | "execution": {
|
404 |
| - "iopub.execute_input": "2025-10-06T03:01:13.309570Z", |
405 |
| - "iopub.status.busy": "2025-10-06T03:01:13.309423Z", |
406 |
| - "iopub.status.idle": "2025-10-06T03:01:13.543919Z", |
407 |
| - "shell.execute_reply": "2025-10-06T03:01:13.543474Z" |
| 394 | + "iopub.execute_input": "2025-10-06T03:38:40.899554Z", |
| 395 | + "iopub.status.busy": "2025-10-06T03:38:40.899407Z", |
| 396 | + "iopub.status.idle": "2025-10-06T03:38:41.150350Z", |
| 397 | + "shell.execute_reply": "2025-10-06T03:38:41.149891Z" |
408 | 398 | }
|
409 | 399 | },
|
410 | 400 | "outputs": [
|
411 | 401 | {
|
412 | 402 | "data": {
|
413 | 403 | "text/html": [
|
414 |
| - "<strong style='color: #00008B;'>The ancient Romans made significant contributions to various fields, including law, philosophy, science, and literature. They were known for their engineering skills, particularly in the construction of monumental architecture like the Colosseum and the Pantheon. The Roman Empire was one of the largest and most powerful empires in history, spanning across Europe, Asia Minor, and North Africa. Their cultural achievements included the development of Roman law and the spread of Christianity. Additionally, they were skilled in agriculture and had a sophisticated system of governance.</strong>" |
| 404 | + "<strong style='color: #00008B;'>Ancient Rome was a significant civilization that made significant contributions to human history. Some of their major achievements include:<br><br>1. The construction of the Colosseum, one of the largest amphitheaters in the world, which served as a venue for gladiatorial games and other public spectacles.<br><br>2. The development of Roman law, which established the principles of justice and legal procedures that are still used today.<br><br>3. The invention of the horse-drawn carriage, which allowed for faster transportation and trade.<br><br>4. The construction of aqueducts that provided water to cities and towns, improving living conditions and economic activity.<br><br>5. The</strong>" |
415 | 405 | ],
|
416 | 406 | "text/plain": [
|
417 | 407 | "<IPython.core.display.HTML object>"
|
|
460 | 450 | "execution_count": 4,
|
461 | 451 | "metadata": {
|
462 | 452 | "execution": {
|
463 |
| - "iopub.execute_input": "2025-10-06T03:01:13.545307Z", |
464 |
| - "iopub.status.busy": "2025-10-06T03:01:13.545169Z", |
465 |
| - "iopub.status.idle": "2025-10-06T03:01:13.655547Z", |
466 |
| - "shell.execute_reply": "2025-10-06T03:01:13.654789Z" |
| 453 | + "iopub.execute_input": "2025-10-06T03:38:41.151833Z", |
| 454 | + "iopub.status.busy": "2025-10-06T03:38:41.151685Z", |
| 455 | + "iopub.status.idle": "2025-10-06T03:38:41.252737Z", |
| 456 | + "shell.execute_reply": "2025-10-06T03:38:41.252288Z" |
467 | 457 | }
|
468 | 458 | },
|
469 | 459 | "outputs": [
|
470 | 460 | {
|
471 | 461 | "name": "stdout",
|
472 | 462 | "output_type": "stream",
|
473 | 463 | "text": [
|
474 |
| - "Yes, \"test\" is a code word for a survey or a sample in various contexts. It's common to use \"test\" to refer to a small group of people or a particular population whose opinions or behaviors will be surveyed in a testing or evaluation activity." |
| 464 | + "Yes, I am Qwen, a language model created by Alibaba Cloud. My purpose is to assist you with any questions or tasks you may have. If you have any questions or would like to talk about a specific topic, feel free to ask!" |
475 | 465 | ]
|
476 | 466 | }
|
477 | 467 | ],
|
|
501 | 491 | "execution_count": 5,
|
502 | 492 | "metadata": {
|
503 | 493 | "execution": {
|
504 |
| - "iopub.execute_input": "2025-10-06T03:01:13.657280Z", |
505 |
| - "iopub.status.busy": "2025-10-06T03:01:13.656894Z", |
506 |
| - "iopub.status.idle": "2025-10-06T03:01:13.822803Z", |
507 |
| - "shell.execute_reply": "2025-10-06T03:01:13.822237Z" |
| 494 | + "iopub.execute_input": "2025-10-06T03:38:41.254144Z", |
| 495 | + "iopub.status.busy": "2025-10-06T03:38:41.254004Z", |
| 496 | + "iopub.status.idle": "2025-10-06T03:38:41.404226Z", |
| 497 | + "shell.execute_reply": "2025-10-06T03:38:41.403853Z" |
508 | 498 | }
|
509 | 499 | },
|
510 | 500 | "outputs": [
|
511 | 501 | {
|
512 | 502 | "data": {
|
513 | 503 | "text/html": [
|
514 |
| - "<strong style='color: #00008B;'>Response: Completion(id='0a9bef63fc134f77b60a8501ba5d912d', choices=[CompletionChoice(finish_reason='length', index=0, logprobs=None, text=' 1. United States - Washington D.C.\\n2. Canada - Ottawa\\n3. France - Paris\\n4. Germany - Berlin\\n5. Japan - Tokyo\\n6. Italy - Rome\\n7. Spain - Madrid\\n8. United Kingdom - London\\n9. Australia - Canberra\\n10. New Zealand', matched_stop=None)], created=1759719673, model='qwen/qwen2.5-0.5b-instruct', object='text_completion', system_fingerprint=None, usage=CompletionUsage(completion_tokens=64, prompt_tokens=8, total_tokens=72, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
| 504 | + "<strong style='color: #00008B;'>Response: Completion(id='3be3295d77d44db9a1843f154ee197a7', choices=[CompletionChoice(finish_reason='length', index=0, logprobs=None, text=' 1. United States - Washington D.C.\\n2. Canada - Ottawa\\n3. France - Paris\\n4. Germany - Berlin\\n5. Japan - Tokyo\\n6. Italy - Rome\\n7. Spain - Madrid\\n8. United Kingdom - London\\n9. Australia - Canberra\\n10. New Zealand', matched_stop=None)], created=1759721921, model='qwen/qwen2.5-0.5b-instruct', object='text_completion', system_fingerprint=None, usage=CompletionUsage(completion_tokens=64, prompt_tokens=8, total_tokens=72, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
515 | 505 | ],
|
516 | 506 | "text/plain": [
|
517 | 507 | "<IPython.core.display.HTML object>"
|
|
550 | 540 | "execution_count": 6,
|
551 | 541 | "metadata": {
|
552 | 542 | "execution": {
|
553 |
| - "iopub.execute_input": "2025-10-06T03:01:13.824422Z", |
554 |
| - "iopub.status.busy": "2025-10-06T03:01:13.824267Z", |
555 |
| - "iopub.status.idle": "2025-10-06T03:01:13.902031Z", |
556 |
| - "shell.execute_reply": "2025-10-06T03:01:13.901479Z" |
| 543 | + "iopub.execute_input": "2025-10-06T03:38:41.405600Z", |
| 544 | + "iopub.status.busy": "2025-10-06T03:38:41.405462Z", |
| 545 | + "iopub.status.idle": "2025-10-06T03:38:41.558349Z", |
| 546 | + "shell.execute_reply": "2025-10-06T03:38:41.557974Z" |
557 | 547 | }
|
558 | 548 | },
|
559 | 549 | "outputs": [
|
560 | 550 | {
|
561 | 551 | "data": {
|
562 | 552 | "text/html": [
|
563 |
| - "<strong style='color: #00008B;'>Response: Completion(id='b96d15fb328c43c8afb3db6bf8ab7023', choices=[CompletionChoice(finish_reason='stop', index=0, logprobs=None, text=' As an AI language model, I do not have personal experiences or emotions, but I can generate a fictional story based on common themes and events that people might encounter in space travel.', matched_stop='\\n\\n')], created=1759719673, model='qwen/qwen2.5-0.5b-instruct', object='text_completion', system_fingerprint=None, usage=CompletionUsage(completion_tokens=36, prompt_tokens=9, total_tokens=45, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
| 553 | + "<strong style='color: #00008B;'>Response: Completion(id='8061143f030d483183cae76728e120c0', choices=[CompletionChoice(finish_reason='stop', index=0, logprobs=None, text=' Once upon a time, there was a space explorer named Dr. Amelia Hart. She had been on many missions to explore the stars and planets, but never before had she been in space with anyone.', matched_stop='\\n\\n')], created=1759721921, model='qwen/qwen2.5-0.5b-instruct', object='text_completion', system_fingerprint=None, usage=CompletionUsage(completion_tokens=40, prompt_tokens=9, total_tokens=49, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
564 | 554 | ],
|
565 | 555 | "text/plain": [
|
566 | 556 | "<IPython.core.display.HTML object>"
|
|
601 | 591 | "execution_count": 7,
|
602 | 592 | "metadata": {
|
603 | 593 | "execution": {
|
604 |
| - "iopub.execute_input": "2025-10-06T03:01:13.903616Z", |
605 |
| - "iopub.status.busy": "2025-10-06T03:01:13.903457Z", |
606 |
| - "iopub.status.idle": "2025-10-06T03:01:13.943927Z", |
607 |
| - "shell.execute_reply": "2025-10-06T03:01:13.943243Z" |
| 594 | + "iopub.execute_input": "2025-10-06T03:38:41.559683Z", |
| 595 | + "iopub.status.busy": "2025-10-06T03:38:41.559549Z", |
| 596 | + "iopub.status.idle": "2025-10-06T03:38:41.618507Z", |
| 597 | + "shell.execute_reply": "2025-10-06T03:38:41.617987Z" |
608 | 598 | }
|
609 | 599 | },
|
610 | 600 | "outputs": [],
|
|
0 commit comments