diff --git a/lumigator/jobs/inference/requirements.txt b/lumigator/jobs/inference/requirements.txt index 1a7bf2088..a21b91d32 100644 --- a/lumigator/jobs/inference/requirements.txt +++ b/lumigator/jobs/inference/requirements.txt @@ -1,12 +1,13 @@ -accelerate==1.1.1 -datasets==2.20.0 +accelerate==1.5.2 +datasets==2.19.1 langcodes==3.5.0 -litellm==1.60.6 -loguru==0.7.2 -pydantic>=2.10.0 -python-box==7.2.0 -requests-mock==1.12.1 -s3fs==2024.5.0 +litellm==1.63.12 +loguru==0.7.3 +numpy==1.26.3 +pandas==2.2.3 +pydantic==2.10.6 +python-box==7.3.2 +s3fs==2024.2.0 sentencepiece==0.2.0 -torch==2.5.1 -transformers==4.46.3 +torch==2.6.0 +transformers==4.49.0 diff --git a/lumigator/jobs/inference/requirements_cpu.txt b/lumigator/jobs/inference/requirements_cpu.txt index 99fd43a1f..0481a610d 100644 --- a/lumigator/jobs/inference/requirements_cpu.txt +++ b/lumigator/jobs/inference/requirements_cpu.txt @@ -1,12 +1,14 @@ --extra-index-url https://download.pytorch.org/whl/cpu -accelerate==1.1.1 -datasets==2.20.0 +accelerate==1.5.2 +datasets==2.19.1 langcodes==3.5.0 -litellm==1.60.4 -loguru==0.7.2 -pydantic>=2.10.0 -python-box==7.2.0 -s3fs +litellm==1.63.12 +loguru==0.7.3 +numpy==1.26.3 +pandas==2.2.3 +pydantic==2.10.6 +python-box==7.3.2 +s3fs==2024.2.0 sentencepiece==0.2.0 -torch==2.5.1 -transformers==4.46.3 +torch==2.6.0 +transformers==4.49.0 diff --git a/notebooks/walkthrough.ipynb b/notebooks/walkthrough.ipynb index a132231d1..c851d5f11 100644 --- a/notebooks/walkthrough.ipynb +++ b/notebooks/walkthrough.ipynb @@ -605,6 +605,8 @@ "metadata": {}, "outputs": [], "source": [ + "import time\n", + "\n", "# set this value to limit the evaluation to the first max_samples items (0=all)\n", "max_samples = 10\n", "# team_name is a way to group jobs together under the same namespace, feel free to customize it\n", @@ -626,9 +628,15 @@ " job_config=infer_job_config,\n", " )\n", " job_infer_creation_result = lm_client.jobs.create_job(infer_job_create)\n", - " lm_client.jobs.wait_for_job(job_infer_creation_result.id)\n", - "\n", - " infer_dataset = lm_client.jobs.get_job_dataset(str(job_infer_creation_result.id))\n", + " try:\n", + " lm_client.jobs.wait_for_job(job_infer_creation_result.id)\n", + " except Exception as e:\n", + " print(f\"Job {job_infer_creation_result.id} error: {e}\")\n", + " continue\n", + "\n", + " # Allow a few seconds for the new dataset to be added now that the job has succeeded\n", + " time.sleep(10)\n", + " infer_dataset = lm_client.jobs.get_job_dataset(job_infer_creation_result.id)\n", "\n", " eval_job_config = JobEvalConfig(\n", " metrics=[\"rouge\", \"meteor\", \"bertscore\"],\n",