|
31 | 31 | "execution_count": 1,
|
32 | 32 | "metadata": {
|
33 | 33 | "execution": {
|
34 |
| - "iopub.execute_input": "2025-10-06T02:09:58.606964Z", |
35 |
| - "iopub.status.busy": "2025-10-06T02:09:58.606839Z", |
36 |
| - "iopub.status.idle": "2025-10-06T02:10:37.396079Z", |
37 |
| - "shell.execute_reply": "2025-10-06T02:10:37.395176Z" |
| 34 | + "iopub.execute_input": "2025-10-06T03:00:32.007007Z", |
| 35 | + "iopub.status.busy": "2025-10-06T03:00:32.006876Z", |
| 36 | + "iopub.status.idle": "2025-10-06T03:01:13.019849Z", |
| 37 | + "shell.execute_reply": "2025-10-06T03:01:13.019316Z" |
38 | 38 | }
|
39 | 39 | },
|
40 | 40 | "outputs": [
|
|
76 | 76 | "output_type": "stream",
|
77 | 77 | "text": [
|
78 | 78 | "/usr/local/lib/python3.10/dist-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.\n",
|
79 |
| - " import pynvml # type: ignore[import]\n", |
| 79 | + " import pynvml # type: ignore[import]\n" |
| 80 | + ] |
| 81 | + }, |
| 82 | + { |
| 83 | + "name": "stdout", |
| 84 | + "output_type": "stream", |
| 85 | + "text": [ |
80 | 86 | "/usr/local/lib/python3.10/dist-packages/torch/cuda/__init__.py:63: FutureWarning: The pynvml package is deprecated. Please install nvidia-ml-py instead. If you did not install pynvml directly, please report this to the maintainers of the package that installed pynvml for you.\n",
|
81 | 87 | " import pynvml # type: ignore[import]\n"
|
82 | 88 | ]
|
|
86 | 92 | "output_type": "stream",
|
87 | 93 | "text": [
|
88 | 94 | "`torch_dtype` is deprecated! Use `dtype` instead!\n",
|
89 |
| - "[2025-10-06 02:10:25] `torch_dtype` is deprecated! Use `dtype` instead!\n" |
| 95 | + "[2025-10-06 03:01:01] `torch_dtype` is deprecated! Use `dtype` instead!\n" |
90 | 96 | ]
|
91 | 97 | },
|
92 | 98 | {
|
|
96 | 102 | "[Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0\n",
|
97 | 103 | "[Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0\n",
|
98 | 104 | "[Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0\n",
|
99 |
| - "[Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0\n", |
100 |
| - "[2025-10-06 02:10:27] MOE_RUNNER_BACKEND is not initialized, using triton backend\n" |
| 105 | + "[Gloo] Rank 0 is connected to 0 peer ranks. Expected number of connected peer ranks is : 0\n" |
| 106 | + ] |
| 107 | + }, |
| 108 | + { |
| 109 | + "name": "stdout", |
| 110 | + "output_type": "stream", |
| 111 | + "text": [ |
| 112 | + "[2025-10-06 03:01:02] MOE_RUNNER_BACKEND is not initialized, using triton backend\n" |
101 | 113 | ]
|
102 | 114 | },
|
103 | 115 | {
|
104 | 116 | "name": "stdout",
|
105 | 117 | "output_type": "stream",
|
106 | 118 | "text": [
|
107 | 119 | "\r",
|
108 |
| - "Loading safetensors checkpoint shards: 0% Completed | 0/1 [00:00<?, ?it/s]\n", |
| 120 | + "Loading safetensors checkpoint shards: 0% Completed | 0/1 [00:00<?, ?it/s]\n" |
| 121 | + ] |
| 122 | + }, |
| 123 | + { |
| 124 | + "name": "stdout", |
| 125 | + "output_type": "stream", |
| 126 | + "text": [ |
109 | 127 | "\r",
|
110 |
| - "Loading safetensors checkpoint shards: 100% Completed | 1/1 [00:00<00:00, 5.47it/s]\n", |
| 128 | + "Loading safetensors checkpoint shards: 100% Completed | 1/1 [00:00<00:00, 4.20it/s]\n", |
111 | 129 | "\r",
|
112 |
| - "Loading safetensors checkpoint shards: 100% Completed | 1/1 [00:00<00:00, 5.46it/s]\n", |
| 130 | + "Loading safetensors checkpoint shards: 100% Completed | 1/1 [00:00<00:00, 4.20it/s]\n", |
113 | 131 | "\n"
|
114 | 132 | ]
|
115 | 133 | },
|
|
119 | 137 | "text": [
|
120 | 138 | "\r",
|
121 | 139 | " 0%| | 0/3 [00:00<?, ?it/s]\r",
|
122 |
| - "Capturing batches (bs=4 avail_mem=67.93 GB): 0%| | 0/3 [00:00<?, ?it/s]" |
| 140 | + "Capturing batches (bs=4 avail_mem=77.03 GB): 0%| | 0/3 [00:00<?, ?it/s]" |
123 | 141 | ]
|
124 | 142 | },
|
125 | 143 | {
|
126 | 144 | "name": "stdout",
|
127 | 145 | "output_type": "stream",
|
128 | 146 | "text": [
|
129 | 147 | "\r",
|
130 |
| - "Capturing batches (bs=4 avail_mem=67.93 GB): 33%|███▎ | 1/3 [00:00<00:00, 3.74it/s]\r", |
131 |
| - "Capturing batches (bs=2 avail_mem=67.87 GB): 33%|███▎ | 1/3 [00:00<00:00, 3.74it/s]\r", |
132 |
| - "Capturing batches (bs=1 avail_mem=67.86 GB): 33%|███▎ | 1/3 [00:00<00:00, 3.74it/s]\r", |
133 |
| - "Capturing batches (bs=1 avail_mem=67.86 GB): 100%|██████████| 3/3 [00:00<00:00, 8.74it/s]\n" |
| 148 | + "Capturing batches (bs=4 avail_mem=77.03 GB): 33%|███▎ | 1/3 [00:00<00:01, 1.53it/s]\r", |
| 149 | + "Capturing batches (bs=2 avail_mem=76.97 GB): 33%|███▎ | 1/3 [00:00<00:01, 1.53it/s]\r", |
| 150 | + "Capturing batches (bs=1 avail_mem=76.96 GB): 33%|███▎ | 1/3 [00:00<00:01, 1.53it/s]\r", |
| 151 | + "Capturing batches (bs=1 avail_mem=76.96 GB): 100%|██████████| 3/3 [00:00<00:00, 4.17it/s]\n" |
134 | 152 | ]
|
135 | 153 | },
|
136 | 154 | {
|
|
149 | 167 | "name": "stdout",
|
150 | 168 | "output_type": "stream",
|
151 | 169 | "text": [
|
152 |
| - "Server started on http://localhost:39399\n" |
| 170 | + "Server started on http://localhost:31747\n" |
153 | 171 | ]
|
154 | 172 | }
|
155 | 173 | ],
|
|
183 | 201 | "execution_count": 2,
|
184 | 202 | "metadata": {
|
185 | 203 | "execution": {
|
186 |
| - "iopub.execute_input": "2025-10-06T02:10:37.398656Z", |
187 |
| - "iopub.status.busy": "2025-10-06T02:10:37.398258Z", |
188 |
| - "iopub.status.idle": "2025-10-06T02:10:37.661630Z", |
189 |
| - "shell.execute_reply": "2025-10-06T02:10:37.661051Z" |
| 204 | + "iopub.execute_input": "2025-10-06T03:01:13.024173Z", |
| 205 | + "iopub.status.busy": "2025-10-06T03:01:13.023033Z", |
| 206 | + "iopub.status.idle": "2025-10-06T03:01:13.307561Z", |
| 207 | + "shell.execute_reply": "2025-10-06T03:01:13.307079Z" |
190 | 208 | }
|
191 | 209 | },
|
192 | 210 | "outputs": [
|
193 | 211 | {
|
194 | 212 | "data": {
|
195 | 213 | "text/html": [
|
196 |
| - "<strong style='color: #00008B;'>Response: ChatCompletion(id='375f25746b1e45be92cdf750444e14b7', choices=[Choice(finish_reason='stop', index=0, logprobs=None, message=ChatCompletionMessage(content='Sure, here are three countries and their respective capitals:\\n\\n1. **United States** - Washington, D.C.\\n2. **Canada** - Ottawa\\n3. **Australia** - Canberra', refusal=None, role='assistant', annotations=None, audio=None, function_call=None, tool_calls=None, reasoning_content=None), matched_stop=151645)], created=1759716637, model='qwen/qwen2.5-0.5b-instruct', object='chat.completion', service_tier=None, system_fingerprint=None, usage=CompletionUsage(completion_tokens=39, prompt_tokens=37, total_tokens=76, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
| 214 | + "<strong style='color: #00008B;'>Response: ChatCompletion(id='22b3d25954cb404ca70949629f098753', choices=[Choice(finish_reason='stop', index=0, logprobs=None, message=ChatCompletionMessage(content='Sure, here are three countries and their respective capitals:\\n\\n1. **United States** - Washington, D.C.\\n2. **Canada** - Ottawa\\n3. **Australia** - Canberra', refusal=None, role='assistant', annotations=None, audio=None, function_call=None, tool_calls=None, reasoning_content=None), matched_stop=151645)], created=1759719673, model='qwen/qwen2.5-0.5b-instruct', object='chat.completion', service_tier=None, system_fingerprint=None, usage=CompletionUsage(completion_tokens=39, prompt_tokens=37, total_tokens=76, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
197 | 215 | ],
|
198 | 216 | "text/plain": [
|
199 | 217 | "<IPython.core.display.HTML object>"
|
|
383 | 401 | "execution_count": 3,
|
384 | 402 | "metadata": {
|
385 | 403 | "execution": {
|
386 |
| - "iopub.execute_input": "2025-10-06T02:10:37.663358Z", |
387 |
| - "iopub.status.busy": "2025-10-06T02:10:37.663203Z", |
388 |
| - "iopub.status.idle": "2025-10-06T02:10:37.885552Z", |
389 |
| - "shell.execute_reply": "2025-10-06T02:10:37.884986Z" |
| 404 | + "iopub.execute_input": "2025-10-06T03:01:13.309570Z", |
| 405 | + "iopub.status.busy": "2025-10-06T03:01:13.309423Z", |
| 406 | + "iopub.status.idle": "2025-10-06T03:01:13.543919Z", |
| 407 | + "shell.execute_reply": "2025-10-06T03:01:13.543474Z" |
390 | 408 | }
|
391 | 409 | },
|
392 | 410 | "outputs": [
|
393 | 411 | {
|
394 | 412 | "data": {
|
395 | 413 | "text/html": [
|
396 |
| - "<strong style='color: #00008B;'>The major achievements of ancient Rome include the construction of the Colosseum, the Pantheon, and the aqueducts that brought water to the city. They also developed a complex system of governance, including the Roman Republic and the Roman Empire. The Romans were known for their military prowess, particularly in their campaigns against the Carthaginians and the Gauls. They are also famous for their art, architecture, and philosophy, which influenced later cultures.</strong>" |
| 414 | + "<strong style='color: #00008B;'>The ancient Romans made significant contributions to various fields, including law, philosophy, science, and literature. They were known for their engineering skills, particularly in the construction of monumental architecture like the Colosseum and the Pantheon. The Roman Empire was one of the largest and most powerful empires in history, spanning across Europe, Asia Minor, and North Africa. Their cultural achievements included the development of Roman law and the spread of Christianity. Additionally, they were skilled in agriculture and had a sophisticated system of governance.</strong>" |
397 | 415 | ],
|
398 | 416 | "text/plain": [
|
399 | 417 | "<IPython.core.display.HTML object>"
|
|
442 | 460 | "execution_count": 4,
|
443 | 461 | "metadata": {
|
444 | 462 | "execution": {
|
445 |
| - "iopub.execute_input": "2025-10-06T02:10:37.886984Z", |
446 |
| - "iopub.status.busy": "2025-10-06T02:10:37.886831Z", |
447 |
| - "iopub.status.idle": "2025-10-06T02:10:37.997297Z", |
448 |
| - "shell.execute_reply": "2025-10-06T02:10:37.996755Z" |
| 463 | + "iopub.execute_input": "2025-10-06T03:01:13.545307Z", |
| 464 | + "iopub.status.busy": "2025-10-06T03:01:13.545169Z", |
| 465 | + "iopub.status.idle": "2025-10-06T03:01:13.655547Z", |
| 466 | + "shell.execute_reply": "2025-10-06T03:01:13.654789Z" |
449 | 467 | }
|
450 | 468 | },
|
451 | 469 | "outputs": [
|
452 | 470 | {
|
453 | 471 | "name": "stdout",
|
454 | 472 | "output_type": "stream",
|
455 | 473 | "text": [
|
456 |
| - "I apologize, but I need more context to understand what you mean by \"a test.\" Could you please provide more details or clarify your question? Whether it's a specific technical question, a question about a specific language, or something else, I could help you with the information you need." |
| 474 | + "Yes, \"test\" is a code word for a survey or a sample in various contexts. It's common to use \"test\" to refer to a small group of people or a particular population whose opinions or behaviors will be surveyed in a testing or evaluation activity." |
457 | 475 | ]
|
458 | 476 | }
|
459 | 477 | ],
|
|
483 | 501 | "execution_count": 5,
|
484 | 502 | "metadata": {
|
485 | 503 | "execution": {
|
486 |
| - "iopub.execute_input": "2025-10-06T02:10:37.998642Z", |
487 |
| - "iopub.status.busy": "2025-10-06T02:10:37.998497Z", |
488 |
| - "iopub.status.idle": "2025-10-06T02:10:38.148945Z", |
489 |
| - "shell.execute_reply": "2025-10-06T02:10:38.148459Z" |
| 504 | + "iopub.execute_input": "2025-10-06T03:01:13.657280Z", |
| 505 | + "iopub.status.busy": "2025-10-06T03:01:13.656894Z", |
| 506 | + "iopub.status.idle": "2025-10-06T03:01:13.822803Z", |
| 507 | + "shell.execute_reply": "2025-10-06T03:01:13.822237Z" |
490 | 508 | }
|
491 | 509 | },
|
492 | 510 | "outputs": [
|
493 | 511 | {
|
494 | 512 | "data": {
|
495 | 513 | "text/html": [
|
496 |
| - "<strong style='color: #00008B;'>Response: Completion(id='b15c911e50ec48bb915fc8a641dd052c', choices=[CompletionChoice(finish_reason='length', index=0, logprobs=None, text=' 1. United States - Washington D.C.\\n2. Canada - Ottawa\\n3. France - Paris\\n4. Germany - Berlin\\n5. Japan - Tokyo\\n6. Italy - Rome\\n7. Spain - Madrid\\n8. United Kingdom - London\\n9. Australia - Canberra\\n10. New Zealand', matched_stop=None)], created=1759716638, model='qwen/qwen2.5-0.5b-instruct', object='text_completion', system_fingerprint=None, usage=CompletionUsage(completion_tokens=64, prompt_tokens=8, total_tokens=72, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
| 514 | + "<strong style='color: #00008B;'>Response: Completion(id='0a9bef63fc134f77b60a8501ba5d912d', choices=[CompletionChoice(finish_reason='length', index=0, logprobs=None, text=' 1. United States - Washington D.C.\\n2. Canada - Ottawa\\n3. France - Paris\\n4. Germany - Berlin\\n5. Japan - Tokyo\\n6. Italy - Rome\\n7. Spain - Madrid\\n8. United Kingdom - London\\n9. Australia - Canberra\\n10. New Zealand', matched_stop=None)], created=1759719673, model='qwen/qwen2.5-0.5b-instruct', object='text_completion', system_fingerprint=None, usage=CompletionUsage(completion_tokens=64, prompt_tokens=8, total_tokens=72, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
497 | 515 | ],
|
498 | 516 | "text/plain": [
|
499 | 517 | "<IPython.core.display.HTML object>"
|
|
532 | 550 | "execution_count": 6,
|
533 | 551 | "metadata": {
|
534 | 552 | "execution": {
|
535 |
| - "iopub.execute_input": "2025-10-06T02:10:38.150455Z", |
536 |
| - "iopub.status.busy": "2025-10-06T02:10:38.150314Z", |
537 |
| - "iopub.status.idle": "2025-10-06T02:10:38.246197Z", |
538 |
| - "shell.execute_reply": "2025-10-06T02:10:38.245724Z" |
| 553 | + "iopub.execute_input": "2025-10-06T03:01:13.824422Z", |
| 554 | + "iopub.status.busy": "2025-10-06T03:01:13.824267Z", |
| 555 | + "iopub.status.idle": "2025-10-06T03:01:13.902031Z", |
| 556 | + "shell.execute_reply": "2025-10-06T03:01:13.901479Z" |
539 | 557 | }
|
540 | 558 | },
|
541 | 559 | "outputs": [
|
542 | 560 | {
|
543 | 561 | "data": {
|
544 | 562 | "text/html": [
|
545 |
| - "<strong style='color: #00008B;'>Response: Completion(id='addc2fde8786458da18724887320bc9c', choices=[CompletionChoice(finish_reason='stop', index=0, logprobs=None, text=' Once upon a time, there was a space explorer named Captain Alex. He had always dreamed of exploring the stars and uncovering the mysteries of the universe. One day, he received an invitation to join a team of scientists on a mission to explore a new planet.', matched_stop='\\n\\n')], created=1759716638, model='qwen/qwen2.5-0.5b-instruct', object='text_completion', system_fingerprint=None, usage=CompletionUsage(completion_tokens=53, prompt_tokens=9, total_tokens=62, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
| 563 | + "<strong style='color: #00008B;'>Response: Completion(id='b96d15fb328c43c8afb3db6bf8ab7023', choices=[CompletionChoice(finish_reason='stop', index=0, logprobs=None, text=' As an AI language model, I do not have personal experiences or emotions, but I can generate a fictional story based on common themes and events that people might encounter in space travel.', matched_stop='\\n\\n')], created=1759719673, model='qwen/qwen2.5-0.5b-instruct', object='text_completion', system_fingerprint=None, usage=CompletionUsage(completion_tokens=36, prompt_tokens=9, total_tokens=45, completion_tokens_details=None, prompt_tokens_details=None, reasoning_tokens=0), metadata={'weight_version': 'default'})</strong>" |
546 | 564 | ],
|
547 | 565 | "text/plain": [
|
548 | 566 | "<IPython.core.display.HTML object>"
|
|
583 | 601 | "execution_count": 7,
|
584 | 602 | "metadata": {
|
585 | 603 | "execution": {
|
586 |
| - "iopub.execute_input": "2025-10-06T02:10:38.247645Z", |
587 |
| - "iopub.status.busy": "2025-10-06T02:10:38.247506Z", |
588 |
| - "iopub.status.idle": "2025-10-06T02:10:38.296314Z", |
589 |
| - "shell.execute_reply": "2025-10-06T02:10:38.294678Z" |
| 604 | + "iopub.execute_input": "2025-10-06T03:01:13.903616Z", |
| 605 | + "iopub.status.busy": "2025-10-06T03:01:13.903457Z", |
| 606 | + "iopub.status.idle": "2025-10-06T03:01:13.943927Z", |
| 607 | + "shell.execute_reply": "2025-10-06T03:01:13.943243Z" |
590 | 608 | }
|
591 | 609 | },
|
592 | 610 | "outputs": [],
|
|
0 commit comments